Skip to content

Commit d867fda

Browse files
Copilotanxiangsir
andcommitted
Add theme-aware LMM Probe results images to README
Co-authored-by: anxiangsir <31175974+anxiangsir@users.noreply.github.com>
1 parent a0952e3 commit d867fda

1 file changed

Lines changed: 7 additions & 1 deletion

File tree

README.md

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -92,7 +92,13 @@ Performance comparison of different vision encoders using Attentive Probe evalua
9292

9393
Training on a mixed dataset of 740K samples from LLaVA-OneVision and 800K samples from LLaVA-Video SFT. The training pipeline proceeds directly to Stage 2 fine-tuning. We adopt a streamlined native-resolution strategy inspired by LLaVA-OneVision: when the input frame resolution matches the model's native input size, it is fed directly—without tiling or cropping—to evaluate the ViT's native resolution capability.
9494

95-
95+
<p align="center">
96+
<picture>
97+
<source media="(prefers-color-scheme: dark)" srcset="https://raw.githubusercontent.com/anxiangsir/asset/main/OneVision/probe_lmm_github_dark.png">
98+
<source media="(prefers-color-scheme: light)" srcset="https://raw.githubusercontent.com/anxiangsir/asset/main/OneVision/probe_lmm_github_light.png">
99+
<img alt="LMM Probe Results" src="https://raw.githubusercontent.com/anxiangsir/asset/main/OneVision/probe_lmm_github_light.png" width="800" style="max-width: 100%;">
100+
</picture>
101+
</p>
96102

97103
## 🔧 Setup
98104

0 commit comments

Comments
 (0)