Skip to content

Commit 5cb3325

Browse files
authored
Add RAM req in doc for post-training (#580)
## Summary Doc fix to https://nvbugspro.nvidia.com/bug/6062848, Readme updates. ## Detailed description - Policy training docs: Added a "Compute Requirements" section (GPU VRAM + system RAM guidance) to all three workflow tutorials (static_manipulation, sequential_static_manipulation, locomanipulation) and fixed the "an an" typo. - Arena-in-your-repo docs: Created an index.rst landing page for the section and updated docs/index.rst to use it instead of listing the three sub-pages individually. - README: Added a link to the "Installing IsaacLab-Arena in Your Repository" guide in the "Publishing Your Own Benchmark" section.
1 parent 71184d8 commit 5cb3325

6 files changed

Lines changed: 39 additions & 13 deletions

File tree

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -195,7 +195,7 @@ NIST Board 1, NVIDIA Isaac GR00T Industrial Benchmarks, NVIDIA DexBench, NVIDIA
195195

196196
We encourage the community to build and publish benchmarks on Isaac Lab-Arena. The recommended workflow:
197197

198-
1. **Maintain your benchmark in your own repository.** Create a branch or package that integrates with Isaac Lab-Arena (e.g. an `IsaacLab-Arena` branch). See [RoboTwin](https://github.com/RoboTwin-Platform/RoboTwin/tree/IsaacLab-Arena) for a reference example.
198+
1. **Maintain your benchmark in your own repository.** Create a branch or package that integrates with Isaac Lab-Arena (e.g. an `IsaacLab-Arena` branch). See [RoboTwin](https://github.com/RoboTwin-Platform/RoboTwin/tree/IsaacLab-Arena) for a reference example. For detailed setup instructions — including repository layout, Dockerfile setup, and how to register custom environments/robots/tasks — see the [Arena in Your Repository](https://isaac-sim.github.io/IsaacLab-Arena/main/pages/arena_in_your_repo/index.html) guide.
199199
2. **Reference your benchmark and Isaac Lab-Arena in publications.** When publishing on ArXiv or elsewhere, cite both your benchmark (by name, with a link to your repository) and Isaac Lab-Arena as the underlying evaluation framework.
200200
3. **List it here.** Open a PR to add your benchmark to the [Published Benchmarks](#published-benchmarks) list above. This README serves as the single source of truth for the Arena benchmark ecosystem so that community can discover and reuse.
201201

docs/index.rst

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -199,12 +199,10 @@ TABLE OF CONTENTS
199199
pages/quickstart/first_experiments/index
200200

201201
.. toctree::
202-
:maxdepth: 1
202+
:maxdepth: 2
203203
:caption: Arena in Your Repo
204204

205-
pages/arena_in_your_repo/external_installation
206-
pages/arena_in_your_repo/external_environments
207-
pages/arena_in_your_repo/external_tasks_and_embodiments
205+
pages/arena_in_your_repo/index
208206

209207
.. toctree::
210208
:maxdepth: 1
Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,13 @@
1+
Arena in Your Repository
2+
========================
3+
4+
This section explains how to integrate IsaacLab-Arena into your own repository — installing
5+
it as a submodule, defining custom environments, and registering your own tasks and
6+
embodiments.
7+
8+
.. toctree::
9+
:maxdepth: 1
10+
11+
external_installation
12+
external_environments
13+
external_tasks_and_embodiments

docs/pages/example_workflows/locomanipulation/step_4_policy_training.rst

Lines changed: 8 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -102,19 +102,24 @@ Step 2: Post-train Policy
102102

103103
We post-train the GR00T N1.6 policy on the task.
104104

105-
The GR00T N1.6 policy has 3 billion parameters so post-training is an an expensive operation.
106-
We provide one post-training option, 8 GPUs with 48GB memory, to achieve the best quality:
105+
The GR00T N1.6 policy has 3 billion parameters so post-training is an expensive operation.
106+
We provide one post-training option, 8 GPUs with 48GB memory, to achieve the best quality.
107107

108108
Training takes approximately 4-8 hours on 8x L40s GPUs.
109109

110+
Compute Requirements:
111+
112+
- **GPUs:** 8x with at least 48 GB VRAM each (e.g. L40s, GB200, etc.)
113+
- **System RAM:** 256 GB or more recommended — multi-GPU training with large batch sizes
114+
and multiple dataloader workers requires substantial host memory
115+
110116
Training Configuration:
111117

112118
- **Base Model:** GR00T-N1.6-3B (foundation model)
113119
- **Tuned Modules:** Visual backbone, projector, diffusion model
114120
- **Frozen Modules:** LLM (language model)
115121
- **Batch Size:** 96 (adjust based on GPU memory)
116122
- **Training Steps:** 20,000
117-
- **GPUs:** 8 (multi-GPU training)
118123

119124
To post-train the policy, run the following command
120125

docs/pages/example_workflows/sequential_static_manipulation/step_4_policy_training.rst

Lines changed: 8 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -110,8 +110,8 @@ Step 2: Post-train Policy
110110

111111
We post-train the GR00T N1.6 policy on the task.
112112

113-
The GR00T N1.6 policy has 3 billion parameters so post training is an an expensive operation.
114-
We provide three post-training options:
113+
The GR00T N1.6 policy has 3 billion parameters so post-training is an expensive operation.
114+
We provide two post-training options:
115115

116116
* Best Quality: 8 GPUs with 48GB memory
117117
* Low Hardware Requirements: 1 GPU with 24GB memory
@@ -123,14 +123,19 @@ We provide three post-training options:
123123

124124
Training takes approximately 4-8 hours on 8x L40s GPUs.
125125

126+
Compute Requirements:
127+
128+
- **GPUs:** 8x with at least 48 GB VRAM each (e.g. L40s, A6000, A100)
129+
- **System RAM:** 256 GB or more recommended — multi-GPU training with large batch sizes
130+
and multiple dataloader workers requires substantial host memory
131+
126132
Training Configuration:
127133

128134
- **Base Model:** GR00T-N1.6-3B (foundation model)
129135
- **Tuned Modules:** Visual backbone, projector, diffusion model
130136
- **Frozen Modules:** LLM (language model)
131137
- **Global Batch Size:** 96 (adjust based on GPU memory)
132138
- **Training Steps:** 20,000
133-
- **GPUs:** 8 (multi-GPU training)
134139

135140
To post-train the policy, run the following command
136141

docs/pages/example_workflows/static_manipulation/step_4_policy_training.rst

Lines changed: 7 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -103,7 +103,7 @@ Step 2: Post-train Policy
103103

104104
We post-train the GR00T N1.6 policy on the task.
105105

106-
The GR00T N1.6 policy has 3 billion parameters so post training is an an expensive operation.
106+
The GR00T N1.6 policy has 3 billion parameters so post-training is an expensive operation.
107107
We provide two post-training options:
108108

109109
* Best Quality: 8 GPUs with 48GB memory
@@ -116,14 +116,19 @@ We provide two post-training options:
116116

117117
Training takes approximately 4-8 hours on 8x L40s GPUs.
118118

119+
Compute Requirements:
120+
121+
- **GPUs:** 8x with at least 48 GB VRAM each (e.g. L40s, A6000, A100)
122+
- **System RAM:** 256 GB or more recommended — multi-GPU training with large batch sizes
123+
and multiple dataloader workers requires substantial host memory
124+
119125
Training Configuration:
120126

121127
- **Base Model:** GR00T-N1.6-3B (foundation model)
122128
- **Tuned Modules:** Visual backbone, projector, diffusion model
123129
- **Frozen Modules:** LLM (language model)
124130
- **Batch Size:** 24 (adjust based on GPU memory)
125131
- **Training Steps:** 20,000
126-
- **GPUs:** 8 (multi-GPU training)
127132

128133
To post-train the policy, run the following command
129134

0 commit comments

Comments
 (0)