Skip to content

Commit 5a1b634

Browse files
committed
adapt both fla-attn and fused_moe
Signed-off-by: Ceng23333 <441651826@qq.com>
1 parent 5bf2e94 commit 5a1b634

1 file changed

Lines changed: 2 additions & 0 deletions

File tree

pyproject.toml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,8 @@ requires-python = ">=3.8"
1313
[project.optional-dependencies]
1414
# Same interpreter as InfiniCore built with ``--aten=y`` (see vllm_kernel_reuse_evaluation.md).
1515
vllm = ["vllm==0.19.0"]
16+
# MiniCPM5 fused stub + jiuge: vLLM fused MoE + helper deps (flash-attn install separately; see InfiniLM/examples/FLASH_ATTN_AND_VLLM_FUSED_MOE.md).
17+
vllm-fused-moe = ["vllm==0.19.0", "janus>=1.0.0", "xxhash>=3.0.0"]
1618

1719
classifiers = [
1820
"Programming Language :: Python :: 3",

0 commit comments

Comments
 (0)