Skip to content

Commit

Permalink
finish example readme
Browse files Browse the repository at this point in the history
  • Loading branch information
TheEimer committed May 30, 2024
1 parent 3115b4a commit 9d393e6
Show file tree
Hide file tree
Showing 2 changed files with 9 additions and 3 deletions.
2 changes: 1 addition & 1 deletion docs/contributing.rst
Original file line number Diff line number Diff line change
@@ -1 +1 @@
.. include:: ../CONTRIBUTING.rst
.. mdinclude:: ../CONTRIBUTING.md
10 changes: 8 additions & 2 deletions examples/Readme.md
Original file line number Diff line number Diff line change
Expand Up @@ -159,7 +159,7 @@ As you can see, most of the defaults are decided by the environment and algorith
In the current config, we do a single training interval consisting of the total number of environment steps suggested for our target domain. If we want to instead run a schedule of length 10 with each schedule segment taking 10e4 steps, we can change the configuration like this:
```bash
python run_heuristic_schedule.py autorl.n_steps=10 autorl.total_timesteps=10000
python run_heuristic_schedule.py autorl.n_steps=10 autorl.n_total_timesteps=10000
```

## 3. Reactive Schedules
Expand All @@ -170,4 +170,10 @@ Lastly, we can also adjust the hyperparameters based on algorithm statistics. In
python run_reactive_schedule.py
```

To actually configure to w
To actually configure to what information ARLBench returns about the RL algorithm's internal state, we can use the 'state' features key - in this case, we want to add the gradient norm and variance like this:

```bash
python run_reactive_schedule.py "autorl.state_features=['grad_info']"
```

Now we can build a schedule that takes the gradient information into account.

0 comments on commit 9d393e6

Please sign in to comment.