Skip to content

Commit 993a444

Browse files
Update README.md
1 parent a06c949 commit 993a444

1 file changed

Lines changed: 25 additions & 24 deletions

File tree

README.md

Lines changed: 25 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -241,9 +241,10 @@ For multi-node distributed training, configure your training script according to
241241

242242
### Attentive Probe Evaluation
243243

244-
#### Codec Evaluation
245244

246-
To evaluate the encoder with codec-style patch selection, first navigate to the evaluation directory:
245+
#### Sampling Evaluation
246+
247+
To evaluate the encoder with uniform frame sampling, first navigate to the evaluation directory:
247248

248249
```bash
249250
cd eval_encoder
@@ -252,33 +253,27 @@ cd eval_encoder
252253
Then run the following command:
253254

254255
```bash
255-
torchrun --nproc_per_node=8 --master_port=29512 attentive_prob_codec.py \
256+
torchrun --nproc_per_node=8 --master_port=29507 attentive_probe.py \
256257
--eval_freq 1 \
257258
--default_lr_list 0.0001 \
258-
--batch_size 4 \
259+
--batch_size 32 \
259260
--default_weight_decay 0 \
260261
--dali_py_num_workers 8 \
261-
--model_family llava_vit_codec \
262+
--model_family llava_vit_sampling \
262263
--dataset diving48 \
263-
--num_frames 64 \
264+
--num_frames 8 \
264265
--model_weight lmms-lab/onevision-encoder-large \
265266
--model_name hf_llava_vit_large_ln \
266267
--embedding_size 1024 \
267-
--default_epoch 30 \
268-
--data_root /path/to/your/data_attentive_probe/ \
269-
--cache_dir /path/to/your/cache_residuals/ \
270-
--K_keep 2048 \
271-
--mv_compensate median
268+
--frames_token_num 256
272269
```
273270

274-
**Codec-Specific Parameters:**
275-
- `cache_dir`: Directory for cached codec patches. This is where the codec-selected patches will be stored/loaded.
276-
- `K_keep`: Number of patches to keep. For example, 256 patches per frame × 8 frames = 2048 total patches. Adjust based on your frame count and desired compression ratio.
277-
- `mv_compensate`: Motion vector compensation method (e.g., `median`).
271+
**Sampling-Specific Parameters:**
272+
- `frames_token_num`: Number of tokens per frame (e.g., 256 tokens for standard sampling).
278273

279-
#### Sampling Evaluation
274+
#### Codec Evaluation
280275

281-
To evaluate the encoder with uniform frame sampling, first navigate to the evaluation directory:
276+
To evaluate the encoder with codec-style patch selection, first navigate to the evaluation directory:
282277

283278
```bash
284279
cd eval_encoder
@@ -287,23 +282,29 @@ cd eval_encoder
287282
Then run the following command:
288283

289284
```bash
290-
torchrun --nproc_per_node=8 --master_port=29507 attentive_probe.py \
285+
torchrun --nproc_per_node=8 --master_port=29512 attentive_prob_codec.py \
291286
--eval_freq 1 \
292287
--default_lr_list 0.0001 \
293-
--batch_size 32 \
288+
--batch_size 4 \
294289
--default_weight_decay 0 \
295290
--dali_py_num_workers 8 \
296-
--model_family llava_vit_sampling \
291+
--model_family llava_vit_codec \
297292
--dataset diving48 \
298-
--num_frames 8 \
293+
--num_frames 64 \
299294
--model_weight lmms-lab/onevision-encoder-large \
300295
--model_name hf_llava_vit_large_ln \
301296
--embedding_size 1024 \
302-
--frames_token_num 256
297+
--default_epoch 30 \
298+
--data_root /path/to/your/data_attentive_probe/ \
299+
--cache_dir /path/to/your/cache_residuals/ \
300+
--K_keep 2048 \
301+
--mv_compensate median
303302
```
304303

305-
**Sampling-Specific Parameters:**
306-
- `frames_token_num`: Number of tokens per frame (e.g., 256 tokens for standard sampling).
304+
**Codec-Specific Parameters:**
305+
- `cache_dir`: Directory for cached codec patches. This is where the codec-selected patches will be stored/loaded.
306+
- `K_keep`: Number of patches to keep. For example, 256 patches per frame × 8 frames = 2048 total patches. Adjust based on your frame count and desired compression ratio.
307+
- `mv_compensate`: Motion vector compensation method (e.g., `median`).
307308

308309
#### Shared Parameters
309310

0 commit comments

Comments
 (0)