We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent adf1d27 commit 59c1cf7Copy full SHA for 59c1cf7
1 file changed
examples/speculative_decoding/README.md
@@ -82,7 +82,7 @@ For small base models that fit in GPU memory, we can collocate them with draft m
82
83
All default training settings live in `eagle3.yaml`; override any field via OmegaConf dotlist arguments on the command line.
84
85
-To enable context parallelism for long-context training, add `training.cp_size=<N>` to the overrides.
+To enable context parallelism for long-context training, add `training.cp_size=<N>`.
86
The saved modelopt checkpoint is similar in architecture to HF models. It can be further optimized through **ModelOpt**, e.g., PTQ and QAT.
87
88
## Training Draft Model with Offline Base Model
0 commit comments