Skip to content

Commit a479985

Browse files
committed
reorder
1 parent 62cbec3 commit a479985

File tree

1 file changed

+41
-24
lines changed

1 file changed

+41
-24
lines changed

docs/pages/example_workflows/reinforcement_learning/step_2_policy_training.rst

Lines changed: 41 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -61,6 +61,47 @@ For example, to train with relu activation and a higher learning rate:
6161
agent.algorithm.learning_rate=0.001
6262
6363
64+
Resuming from a Checkpoint
65+
^^^^^^^^^^^^^^^^^^^^^^^^^^
66+
67+
To resume training from a previously saved checkpoint, use the ``--resume`` flag
68+
together with ``--load_run`` (run folder name) and ``--checkpoint`` (model filename).
69+
Both arguments are optional — when omitted, the most recent run and latest checkpoint
70+
are used automatically.
71+
72+
.. code-block:: bash
73+
74+
python submodules/IsaacLab/scripts/reinforcement_learning/rsl_rl/train.py \
75+
--external_callback isaaclab_arena.environments.isaaclab_interop.environment_registration_callback \
76+
--task lift_object \
77+
--rl_training_mode \
78+
--num_envs 4096 \
79+
--max_iterations 4000 \
80+
--resume \
81+
--load_run <timestamp> \
82+
--checkpoint model_1999.pt
83+
84+
Replace ``<timestamp>`` with the run folder name under ``logs/rsl_rl/generic_experiment/``.
85+
If ``--load_run`` is omitted, the latest run is selected. If ``--checkpoint`` is omitted,
86+
the latest checkpoint in that run is loaded.
87+
88+
.. tip::
89+
90+
You can also combine resume with Hydra overrides to change hyperparameters mid-training,
91+
e.g. lowering the learning rate for fine-tuning:
92+
93+
.. code-block:: bash
94+
95+
python submodules/IsaacLab/scripts/reinforcement_learning/rsl_rl/train.py \
96+
--external_callback isaaclab_arena.environments.isaaclab_interop.environment_registration_callback \
97+
--task lift_object \
98+
--rl_training_mode \
99+
--num_envs 4096 \
100+
--max_iterations 4000 \
101+
--resume \
102+
agent.algorithm.learning_rate=0.00005
103+
104+
64105
Monitoring Training
65106
^^^^^^^^^^^^^^^^^^^
66107

@@ -101,30 +142,6 @@ During training, each iteration prints a summary to the console:
101142
ETA: 00:00:49
102143
103144
104-
Resuming from a Checkpoint
105-
^^^^^^^^^^^^^^^^^^^^^^^^^^
106-
107-
To resume training from a previously saved checkpoint, use the ``--resume`` flag
108-
together with ``--load_run`` (run folder name) and ``--checkpoint`` (model filename).
109-
Both arguments are optional — when omitted, the most recent run and latest checkpoint
110-
are used automatically.
111-
112-
.. code-block:: bash
113-
114-
python submodules/IsaacLab/scripts/reinforcement_learning/rsl_rl/train.py \
115-
--external_callback isaaclab_arena.environments.isaaclab_interop.environment_registration_callback \
116-
--task lift_object \
117-
--rl_training_mode \
118-
--num_envs 4096 \
119-
--max_iterations 4000 \
120-
--resume \
121-
--load_run <timestamp> \
122-
--checkpoint model_1999.pt
123-
124-
Replace ``<timestamp>`` with the run folder name under ``logs/rsl_rl/generic_experiment/``.
125-
If ``--load_run`` is omitted, the latest run is selected. If ``--checkpoint`` is omitted,
126-
the latest checkpoint in that run is loaded.
127-
128145
129146
Multi-GPU Training
130147
^^^^^^^^^^^^^^^^^^

0 commit comments

Comments
 (0)