@@ -21,6 +21,7 @@ Arguments:
2121 - mistralai/Voxtral-Mini-3B-2507
2222 - openai/whisper series (whisper-{small, medium, large, large-v2, large-v3, large-v3-turbo})
2323 - google/gemma-3-4b-it
24+ - Qwen/Qwen3-0.6B
2425 - nvidia/parakeet-tdt
2526 - mistralai/Voxtral-Mini-4B-Realtime-2602
2627
@@ -151,6 +152,18 @@ case "$HF_MODEL" in
151152 AUDIO_FILE=" "
152153 IMAGE_PATH=" docs/source/_static/img/et-logo.png"
153154 ;;
155+ Qwen/Qwen3-0.6B)
156+ MODEL_NAME=" qwen3"
157+ RUNNER_TARGET=" llama_main"
158+ RUNNER_PATH=" llama"
159+ EXPECTED_OUTPUT=" Paris"
160+ PREPROCESSOR=" "
161+ TOKENIZER_URL=" https://huggingface.co/Qwen/Qwen3-0.6B/resolve/main" # @lint-ignore
162+ TOKENIZER_FILE=" "
163+ AUDIO_URL=" "
164+ AUDIO_FILE=" "
165+ IMAGE_PATH=" "
166+ ;;
154167 nvidia/parakeet-tdt)
155168 MODEL_NAME=" parakeet"
156169 RUNNER_TARGET=" parakeet_runner"
@@ -177,7 +190,7 @@ case "$HF_MODEL" in
177190 ;;
178191 * )
179192 echo " Error: Unsupported model '$HF_MODEL '"
180- echo " Supported models: mistralai/Voxtral-Mini-3B-2507, mistralai/Voxtral-Mini-4B-Realtime-2602, openai/whisper series (whisper-{small, medium, large, large-v2, large-v3, large-v3-turbo}), google/gemma-3-4b-it, nvidia/parakeet-tdt"
193+ echo " Supported models: mistralai/Voxtral-Mini-3B-2507, mistralai/Voxtral-Mini-4B-Realtime-2602, openai/whisper series (whisper-{small, medium, large, large-v2, large-v3, large-v3-turbo}), google/gemma-3-4b-it, Qwen/Qwen3-0.6B, nvidia/parakeet-tdt"
181194 exit 1
182195 ;;
183196esac
@@ -246,9 +259,14 @@ if [ "$(uname -s)" = "Darwin" ] && [ -f "$RUNNER_BIN" ]; then
246259 install_name_tool -change /opt/llvm-openmp/lib/libomp.dylib @rpath/libomp.dylib " $RUNNER_BIN "
247260 fi
248261fi
249- # For CUDA, add data_path argument (Metal embeds data in .pte)
262+ # For CUDA, add named data argument (Metal embeds data in .pte).
263+ # Llama runner uses --data_paths, other runners use --data_path.
250264if [ " $DEVICE " = " cuda" ]; then
251- RUNNER_ARGS=" $RUNNER_ARGS --data_path ${MODEL_DIR} /aoti_cuda_blob.ptd"
265+ if [ " $RUNNER_PATH " = " llama" ]; then
266+ RUNNER_ARGS=" $RUNNER_ARGS --data_paths ${MODEL_DIR} /aoti_cuda_blob.ptd"
267+ else
268+ RUNNER_ARGS=" $RUNNER_ARGS --data_path ${MODEL_DIR} /aoti_cuda_blob.ptd"
269+ fi
252270fi
253271
254272# Add model-specific arguments
@@ -262,6 +280,15 @@ case "$MODEL_NAME" in
262280 gemma3)
263281 RUNNER_ARGS=" $RUNNER_ARGS --tokenizer_path ${MODEL_DIR} / --image_path $IMAGE_PATH "
264282 ;;
283+ qwen3)
284+ PROMPT_FILE=" ${MODEL_DIR} /qwen3_prompt.txt"
285+ cat > " ${PROMPT_FILE} " << 'EOF '
286+ <|im_start|>user
287+ What is the capital of France?<|im_end|>
288+ <|im_start|>assistant
289+ EOF
290+ RUNNER_ARGS=" $RUNNER_ARGS --tokenizer_path ${MODEL_DIR} / --prompt_file ${PROMPT_FILE} "
291+ ;;
265292 parakeet)
266293 RUNNER_ARGS=" --model_path ${MODEL_DIR} /model.pte --audio_path ${MODEL_DIR} /$AUDIO_FILE --tokenizer_path ${MODEL_DIR} /$TOKENIZER_FILE "
267294 # For CUDA, add data_path argument (Metal embeds data in .pte)
0 commit comments