Skip to content

Pull requests: Blaizzy/mlx-vlm

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Add --max-tokens CLI argument to server
#936 opened Apr 5, 2026 by nnorris7 Loading…
1 of 3 tasks
Fix Gemma 4 quantized per-layer projection loading
#935 opened Apr 5, 2026 by spicyneuron Loading…
test: add PaddleOCR-VL processor regression coverage
#933 opened Apr 5, 2026 by jimmyzhuu Loading…
docs: add Chinese documentation and restructure docs directory
#928 opened Apr 5, 2026 by Tsan1024 Loading…
3 tasks done
Fix Gemma 4 'No text generated' when chat template is missing
#924 opened Apr 4, 2026 by nnorris7 Loading…
3 of 4 tasks
Centralize server config and add CLI flags
#918 opened Apr 4, 2026 by spicyneuron Loading…
Fix batch generation and adopt mlx-lm batch improvements
#911 opened Apr 4, 2026 by Blaizzy Loading…
3 tasks done
Optimize TurboQuant: O(d log d) Walsh-Hadamard Transform
#860 opened Mar 26, 2026 by Trucker2827 Loading…
1 of 3 tasks
Add distributed infer for qwen3_vl_moe
#730 opened Feb 13, 2026 by Blaizzy Loading…
Distributed inference for Kimi K2.5
#689 opened Jan 27, 2026 by pcuenca Loading…
Implement Joycaption as a custom Llava model
#659 opened Jan 16, 2026 by nArn0 Loading…
[WIP] Token filtering + merging
#185 opened Jan 20, 2025 by Blaizzy Loading…
ProTip! no:milestone will show everything without a milestone.