feat: bump mlx-swift-lm for DeepSeek-V4 support#79
Merged
Conversation
Points mlx-swift-lm to feat/deepseek-v4 branch (SharpAI/mlx-swift-lm#33) which adds DeepseekV4.swift and registers the deepseek_v4 model type.
Contributor
There was a problem hiding this comment.
Pull request overview
This PR bumps the mlx-swift-lm git submodule to a newer revision that adds DeepSeek-V4 model support, including a new model implementation and factory registration (per PR description).
Changes:
- Update
mlx-swift-lmsubmodule reference to SharpAI/mlx-swift-lm#33 (feat/deepseek-v4). - Add DeepSeek-V4 architecture +
deepseek_v4registration inside the submodule (as described in the PR notes).
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
- README: add DeepSeek-V4-Flash (126GB Q3) benchmark table for M5 Pro 64GB SSD+TurboQuant delivers 4.16 tok/s at 40K context (13x vs plain SSD Stream) - profile_runner.py: track peak GPU InUse via background polling thread (0.5s) instead of single post-generation snapshot; rename gpu_in_use → gpu_in_use_peak throughout; add separate GPU_InUse peak visualization section - run_benchmark.sh: add Thump604/DeepSeek-V4-Flash-MLX-Q3-mixed-gs128-affine to Test 1 model list (option 11) - mlx-swift-lm: bump submodule to 8a8da29 (attn_sink dtype fix)
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Summary
Bumps the
mlx-swift-lmsubmodule to include DeepSeek-V4 model support.Changes
mlx-swift-lm→ feat: add DeepSeek-V4 model support mlx-swift-lm#33 (feat/deepseek-v4)DeepseekV4.swiftwith full V4 architecture (mHC, unified KV, sqrtsoftplus MoE)deepseek_v4model type inLLMModelFactoryNotes
deepseek-ai/DeepSeek-V4-Pro/inference/model.py)