|
71 | 71 | ) |
72 | 72 | from modelopt.torch.utils.dataset_utils import ( |
73 | 73 | create_forward_loop, |
74 | | - get_calib_and_holdout_dataloaders, |
75 | 74 | get_dataset_dataloader, |
76 | 75 | get_max_batch_size, |
77 | 76 | get_supported_datasets, |
@@ -204,10 +203,9 @@ def make_calib_dataloader( |
204 | 203 | tokenizer: PreTrainedTokenizerBase | None, |
205 | 204 | device: torch.device, |
206 | 205 | model_type: str | None, |
207 | | -) -> tuple[DataLoader | _DeviceDataLoader, str | None, Path | None]: |
| 206 | +) -> tuple[DataLoader | _DeviceDataLoader, str | None]: |
208 | 207 | calib_dataloader = None |
209 | 208 | first_text_speech_dataset = None |
210 | | - holdout_path = None |
211 | 209 | if args.specdec_offline_dataset is not None: |
212 | 210 | offline_data_path = Path(args.specdec_offline_dataset) |
213 | 211 | dumped_files = sorted(str(p) for p in offline_data_path.glob("*.pt")) |
@@ -286,28 +284,15 @@ def make_calib_dataloader( |
286 | 284 | args.auto_quantize_bits is not None and args.auto_quantize_method == "gradient" |
287 | 285 | ) |
288 | 286 |
|
289 | | - if args.holdout_size > 0: |
290 | | - calib_dataloader, holdout_path = get_calib_and_holdout_dataloaders( |
291 | | - dataset_name=args.dataset, |
292 | | - tokenizer=tokenizer, |
293 | | - batch_size=args.batch_size, |
294 | | - calib_size=args.calib_size, |
295 | | - holdout_size=args.holdout_size, |
296 | | - max_sample_length=args.calib_seq, |
297 | | - device=device, |
298 | | - include_labels=include_labels, |
299 | | - save_dir=args.calib_data_dir, |
300 | | - ) |
301 | | - else: |
302 | | - calib_dataloader = get_dataset_dataloader( |
303 | | - dataset_name=args.dataset, |
304 | | - tokenizer=tokenizer, |
305 | | - batch_size=args.batch_size, |
306 | | - num_samples=args.calib_size, |
307 | | - device=device, |
308 | | - include_labels=include_labels, |
309 | | - ) |
310 | | - return calib_dataloader, first_text_speech_dataset, holdout_path |
| 287 | + calib_dataloader = get_dataset_dataloader( |
| 288 | + dataset_name=args.dataset, |
| 289 | + tokenizer=tokenizer, |
| 290 | + batch_size=args.batch_size, |
| 291 | + num_samples=args.calib_size, |
| 292 | + device=device, |
| 293 | + include_labels=include_labels, |
| 294 | + ) |
| 295 | + return calib_dataloader, first_text_speech_dataset |
311 | 296 |
|
312 | 297 |
|
313 | 298 | def auto_quantize( |
@@ -1049,7 +1034,7 @@ def quantize_main( |
1049 | 1034 |
|
1050 | 1035 | print(f"Use calib batch_size {args.batch_size}") |
1051 | 1036 |
|
1052 | | - calib_dataloader, first_text_speech_dataset, holdout_path = make_calib_dataloader( |
| 1037 | + calib_dataloader, first_text_speech_dataset = make_calib_dataloader( |
1053 | 1038 | args, language_model, processor, tokenizer, device, model_type |
1054 | 1039 | ) |
1055 | 1040 |
|
@@ -1205,26 +1190,6 @@ def parse_args() -> argparse.Namespace: |
1205 | 1190 | type=str, |
1206 | 1191 | default="512", |
1207 | 1192 | ) |
1208 | | - parser.add_argument( |
1209 | | - "--holdout_size", |
1210 | | - help=( |
1211 | | - "Number of holdout samples to save as a .pt file for evaluation. " |
1212 | | - "Holdout samples are drawn from the same dataset immediately after " |
1213 | | - "the calibration samples so there is no overlap. 0 disables holdout." |
1214 | | - ), |
1215 | | - type=int, |
1216 | | - default=0, |
1217 | | - ) |
1218 | | - parser.add_argument( |
1219 | | - "--calib_data_dir", |
1220 | | - help=( |
1221 | | - "Directory to save/load calib.pt and holdout.pt. " |
1222 | | - "If both files exist, data is reloaded from disk instead of re-downloading. " |
1223 | | - "Defaults to --export_path if not specified." |
1224 | | - ), |
1225 | | - type=str, |
1226 | | - default=None, |
1227 | | - ) |
1228 | 1193 | parser.add_argument( |
1229 | 1194 | "--calib_seq", |
1230 | 1195 | help="Maximum sequence length for calibration.", |
|
0 commit comments