Skip to content

Commit b1d1eac

Browse files
author
Yadan Wei
committed
enable all models
Signed-off-by: Yadan Wei <yadanwei@amazon.com>
1 parent cd46502 commit b1d1eac

1 file changed

Lines changed: 22 additions & 21 deletions

File tree

.github/config/vllm-omni-model-tests.yml

Lines changed: 22 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -11,22 +11,22 @@ s3_prefix: "s3://dlc-cicd-models/omni-models"
1111
smoke-test:
1212
codebuild-fleet:
1313
# --- TTS models (route: /v1/audio/speech) ---
14-
# - name: "qwen3-tts-1.7b-customvoice"
15-
# s3_model: "qwen3-tts-1.7b-customvoice.tar.gz"
16-
# fleet: "x86-g6xl-runner"
17-
# extra_args: ""
18-
# route: "/v1/audio/speech"
19-
# test_request: '{"input": "Hello, how are you?", "voice": "vivian", "language": "English"}'
20-
# validate: "binary_size_gt:1000"
14+
- name: "qwen3-tts-1.7b-customvoice"
15+
s3_model: "qwen3-tts-1.7b-customvoice.tar.gz"
16+
fleet: "x86-g6xl-runner"
17+
extra_args: ""
18+
route: "/v1/audio/speech"
19+
test_request: '{"input": "Hello, how are you?", "voice": "vivian", "language": "English"}'
20+
validate: "binary_size_gt:1000"
2121

2222
# --- Image generation models (route: /v1/images/generations) ---
23-
# - name: "flux2-klein-4b"
24-
# s3_model: "flux2-klein-4b.tar.gz"
25-
# fleet: "x86-g6xl-runner"
26-
# extra_args: ""
27-
# route: "/v1/images/generations"
28-
# test_request: '{"prompt": "a red apple on a white table", "size": "512x512", "n": 1}'
29-
# validate: "json_field:data[0].b64_json"
23+
- name: "flux2-klein-4b"
24+
s3_model: "flux2-klein-4b.tar.gz"
25+
fleet: "x86-g6xl-runner"
26+
extra_args: ""
27+
route: "/v1/images/generations"
28+
test_request: '{"prompt": "a red apple on a white table", "size": "512x512", "n": 1}'
29+
validate: "json_field:data[0].b64_json"
3030

3131
# --- Video generation models (route: /v1/videos) ---
3232
- name: "wan2.1-t2v-1.3b"
@@ -39,6 +39,7 @@ smoke-test:
3939
validate: "json_field:id"
4040

4141
# --- Omni chat models (route: /v1/chat/completions, fallthrough) ---
42+
# model is big, won't run for now
4243
# - name: "bagel-7b-mot"
4344
# s3_model: "bagel-7b-mot.tar.gz"
4445
# fleet: "x86-g6e4xl-runner"
@@ -47,10 +48,10 @@ smoke-test:
4748
# test_request: '{"messages": [{"role": "user", "content": [{"type": "text", "text": "<|im_start|>A cute cat<|im_end|>"}]}], "modalities": ["image"], "height": 512, "width": 512, "num_inference_steps": 4, "seed": 42}'
4849
# validate: "json_field:choices[0].message.content"
4950

50-
# - name: "qwen2.5-omni-3b"
51-
# s3_model: "qwen2.5-omni-3b.tar.gz"
52-
# fleet: "x86-g6e12xl-runner"
53-
# extra_args: ""
54-
# route: "/v1/chat/completions"
55-
# test_request: '{"messages": [{"role": "user", "content": "Say hello in one sentence."}], "max_tokens": 64}'
56-
# validate: "json_field:choices[0].message.content"
51+
- name: "qwen2.5-omni-3b"
52+
s3_model: "qwen2.5-omni-3b.tar.gz"
53+
fleet: "x86-g6e12xl-runner"
54+
extra_args: ""
55+
route: "/v1/chat/completions"
56+
test_request: '{"messages": [{"role": "user", "content": "Say hello in one sentence."}], "max_tokens": 64}'
57+
validate: "json_field:choices[0].message.content"

0 commit comments

Comments
 (0)