Commit d94e64e
committed
Refactor llm_qat: decouple dataset blend size from training runtime params
Dataset configs are now self-contained with blend_size and splits ratios.
train_samples/eval_samples in train configs become runtime caps that don't
invalidate the dataset cache. quantize.py only needs --dataset_config and
--recipe instead of a full train config. Convert int4_blockwise_weight_only
recipe to new list-of-dicts quant_cfg format.
Route simple_qat_train through the blend dataset:
- simple_qat_train.py: drop inline Daring-Anteater loader; load
configs/dataset/blend.yaml via build_blend_dataset and use the "eval"
split for calibration
- configs/dataset/blend.yaml: 100K -> 20K samples, 90/5/5 splits
- dataset_utils.py: widen _stream_samples try/except to skip sources that
fail (e.g. pyarrow schema errors); simplify chat-template fallback
detection
Address PR review:
- Fix recipe.ptq_cfg -> recipe.quantize in simple_qat_train.py, README,
and transformers_trainer.py (ModelOptPTQRecipe has no ptq_cfg attr)
- Add --config bounds check in ModelOptArgParser
- Narrow broad exception handlers in dataset_utils.py
- Fix <|im_end|> assistant mask to exclude boundary token
- Warn when --recipe is passed to train.py (ignored, use quantize.py)
- Hyphenate "right-padded" in arguments.py help text
- Move HF TrainingArguments default overrides from arguments.py to YAML
configs; drop redundant eval_accumulation_steps and learning_rate
- Rename .yml -> .yaml for normalize-yaml-ext hook
- Update ARGUMENTS.md help text to say "additional arguments"
Signed-off-by: realAsma <akuriparambi@nvidia.com>1 parent 4e33368 commit d94e64e
File tree
41 files changed
+2751
-1010
lines changed- examples
- llm_qad
- llm_qat
- accelerate_config
- configs
- accelerate
- dataset
- train
- llama_factory
- notebooks
- vllm_serve
- modelopt_recipes/general/ptq
- modelopt/torch
- distill/plugins
- opt/plugins
- quantization/plugins
- tests
- examples
- llm_qat
- unit/torch/opt/plugins
Some content is hidden
Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.
41 files changed
+2751
-1010
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
108 | 108 | | |
109 | 109 | | |
110 | 110 | | |
111 | | - | |
| 111 | + | |
112 | 112 | | |
113 | 113 | | |
114 | 114 | | |
| |||
136 | 136 | | |
137 | 137 | | |
138 | 138 | | |
| 139 | + | |
| 140 | + | |
| 141 | + | |
| 142 | + | |
| 143 | + | |
| 144 | + | |
| 145 | + | |
| 146 | + | |
| 147 | + | |
| 148 | + | |
| 149 | + | |
| 150 | + | |
| 151 | + | |
| 152 | + | |
| 153 | + | |
139 | 154 | | |
140 | 155 | | |
141 | 156 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
15 | 15 | | |
16 | 16 | | |
17 | 17 | | |
| 18 | + | |
| 19 | + | |
| 20 | + | |
18 | 21 | | |
19 | 22 | | |
20 | 23 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
2 | 2 | | |
3 | 3 | | |
4 | 4 | | |
| 5 | + | |
| 6 | + | |
5 | 7 | | |
6 | 8 | | |
7 | 9 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
| 1 | + | |
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
| 1 | + | |
| 2 | + | |
| 3 | + | |
| 4 | + | |
| 5 | + | |
| 6 | + | |
| 7 | + | |
| 8 | + | |
| 9 | + | |
| 10 | + | |
| 11 | + | |
| 12 | + | |
| 13 | + | |
| 14 | + | |
| 15 | + | |
| 16 | + | |
| 17 | + | |
| 18 | + | |
| 19 | + | |
| 20 | + | |
| 21 | + | |
| 22 | + | |
| 23 | + | |
| 24 | + | |
| 25 | + | |
| 26 | + | |
| 27 | + | |
| 28 | + | |
| 29 | + | |
| 30 | + | |
| 31 | + | |
| 32 | + | |
| 33 | + | |
| 34 | + | |
| 35 | + | |
| 36 | + | |
| 37 | + | |
| 38 | + | |
| 39 | + | |
| 40 | + | |
| 41 | + | |
| 42 | + | |
| 43 | + | |
| 44 | + | |
| 45 | + | |
| 46 | + | |
| 47 | + | |
| 48 | + | |
| 49 | + | |
| 50 | + | |
0 commit comments