@@ -61,6 +61,47 @@ For example, to train with relu activation and a higher learning rate:
6161 agent.algorithm.learning_rate=0.001
6262
6363
64+ Resuming from a Checkpoint
65+ ^^^^^^^^^^^^^^^^^^^^^^^^^^
66+
67+ To resume training from a previously saved checkpoint, use the ``--resume `` flag
68+ together with ``--load_run `` (run folder name) and ``--checkpoint `` (model filename).
69+ Both arguments are optional — when omitted, the most recent run and latest checkpoint
70+ are used automatically.
71+
72+ .. code-block :: bash
73+
74+ python submodules/IsaacLab/scripts/reinforcement_learning/rsl_rl/train.py \
75+ --external_callback isaaclab_arena.environments.isaaclab_interop.environment_registration_callback \
76+ --task lift_object \
77+ --rl_training_mode \
78+ --num_envs 4096 \
79+ --max_iterations 4000 \
80+ --resume \
81+ --load_run < timestamp> \
82+ --checkpoint model_1999.pt
83+
84+ Replace ``<timestamp> `` with the run folder name under ``logs/rsl_rl/generic_experiment/ ``.
85+ If ``--load_run `` is omitted, the latest run is selected. If ``--checkpoint `` is omitted,
86+ the latest checkpoint in that run is loaded.
87+
88+ .. tip ::
89+
90+ You can also combine resume with Hydra overrides to change hyperparameters mid-training,
91+ e.g. lowering the learning rate for fine-tuning:
92+
93+ .. code-block :: bash
94+
95+ python submodules/IsaacLab/scripts/reinforcement_learning/rsl_rl/train.py \
96+ --external_callback isaaclab_arena.environments.isaaclab_interop.environment_registration_callback \
97+ --task lift_object \
98+ --rl_training_mode \
99+ --num_envs 4096 \
100+ --max_iterations 4000 \
101+ --resume \
102+ agent.algorithm.learning_rate=0.00005
103+
104+
64105 Monitoring Training
65106^^^^^^^^^^^^^^^^^^^
66107
@@ -101,30 +142,6 @@ During training, each iteration prints a summary to the console:
101142 ETA: 00:00:49
102143
103144
104- Resuming from a Checkpoint
105- ^^^^^^^^^^^^^^^^^^^^^^^^^^
106-
107- To resume training from a previously saved checkpoint, use the ``--resume `` flag
108- together with ``--load_run `` (run folder name) and ``--checkpoint `` (model filename).
109- Both arguments are optional — when omitted, the most recent run and latest checkpoint
110- are used automatically.
111-
112- .. code-block :: bash
113-
114- python submodules/IsaacLab/scripts/reinforcement_learning/rsl_rl/train.py \
115- --external_callback isaaclab_arena.environments.isaaclab_interop.environment_registration_callback \
116- --task lift_object \
117- --rl_training_mode \
118- --num_envs 4096 \
119- --max_iterations 4000 \
120- --resume \
121- --load_run < timestamp> \
122- --checkpoint model_1999.pt
123-
124- Replace ``<timestamp> `` with the run folder name under ``logs/rsl_rl/generic_experiment/ ``.
125- If ``--load_run `` is omitted, the latest run is selected. If ``--checkpoint `` is omitted,
126- the latest checkpoint in that run is loaded.
127-
128145
129146 Multi-GPU Training
130147^^^^^^^^^^^^^^^^^^
0 commit comments