Skip to content

Commit a197679

Browse files
Update to moe w/ hybrid attention
1 parent fa7769e commit a197679

1 file changed

Lines changed: 1 addition & 1 deletion

File tree

docs/reference/models/supported_models_and_architectures.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@ MaxText is an open-source, high-performance LLM framework written in Python/JAX.
4444

4545
### Qwen3
4646

47-
- **Variants**: Dense (0.6B–32B); MoE (30B-A3B, 235B-A22B, 480B Coder), MoE w/ Hybrid (Next-80B-a3b)
47+
- **Variants**: Dense (0.6B–32B); MoE (30B-A3B, 235B-A22B, 480B Coder), MoE w/ Hybrid Attention (Next-80B-a3b)
4848
- **Notes**: **QK-Norm**, GQA, SwiGLU, RMSNorm, RoPE, GatedDeltaNet.
4949

5050
### GPT-OSS

0 commit comments

Comments
 (0)