Skip to content

Commit e8e1564

Browse files
authored
Use vllm release image for AudioQnA and ChatQnA. (#2239)
Signed-off-by: ZePan110 <ze.pan@intel.com>
1 parent ed1cf28 commit e8e1564

35 files changed

Lines changed: 32 additions & 179 deletions

AudioQnA/docker_compose/amd/cpu/epyc/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ services:
2626
https_proxy: ${https_proxy}
2727
restart: unless-stopped
2828
vllm-service:
29-
image: ${REGISTRY:-opea}/vllm:${TAG:-latest}
29+
image: public.ecr.aws/q9t5s3a7/vllm-cpu-release-repo:v0.10.1
3030
container_name: vllm-service
3131
ports:
3232
- ${LLM_SERVER_PORT:-3006}:80

AudioQnA/docker_compose/intel/cpu/xeon/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ services:
2525
https_proxy: ${https_proxy}
2626
restart: unless-stopped
2727
vllm-service:
28-
image: ${REGISTRY:-opea}/vllm:${TAG:-latest}
28+
image: public.ecr.aws/q9t5s3a7/vllm-cpu-release-repo:v0.10.1
2929
container_name: vllm-service
3030
ports:
3131
- ${LLM_SERVER_PORT:-3006}:80

AudioQnA/docker_compose/intel/cpu/xeon/compose_multilang.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ services:
2929
# - ./pretrained_models/:/home/user/GPT-SoVITS/GPT_SoVITS/pretrained_models/
3030
restart: unless-stopped
3131
vllm-service:
32-
image: ${REGISTRY:-opea}/vllm:${TAG:-latest}
32+
image: public.ecr.aws/q9t5s3a7/vllm-cpu-release-repo:v0.10.1
3333
container_name: vllm-service
3434
ports:
3535
- ${LLM_SERVER_PORT:-3006}:80

AudioQnA/docker_image_build/build.yaml

Lines changed: 0 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -73,12 +73,6 @@ services:
7373
dockerfile: comps/third_parties/gpt-sovits/src/Dockerfile
7474
extends: audioqna
7575
image: ${REGISTRY:-opea}/gpt-sovits:${TAG:-latest}
76-
vllm:
77-
build:
78-
context: vllm
79-
dockerfile: docker/Dockerfile.cpu
80-
extends: audioqna
81-
image: ${REGISTRY:-opea}/vllm:${TAG:-latest}
8276
vllm-gaudi:
8377
build:
8478
context: vllm-fork

AudioQnA/tests/test_compose_multilang_on_xeon.sh

Lines changed: 1 addition & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -25,14 +25,8 @@ function build_docker_images() {
2525
docker build --no-cache -t ${REGISTRY}/comps-base:${TAG} --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
2626
popd && sleep 1s
2727

28-
git clone https://github.com/vllm-project/vllm.git
29-
cd ./vllm/
30-
VLLM_VER=v0.10.0
31-
echo "Check out vLLM tag ${VLLM_VER}"
32-
git checkout ${VLLM_VER} &> /dev/null && cd ../
33-
3428
echo "Build all the images with --no-cache, check docker_image_build.log for details..."
35-
service_list="audioqna-multilang audioqna-ui whisper gpt-sovits vllm"
29+
service_list="audioqna-multilang audioqna-ui whisper gpt-sovits"
3630
docker compose -f build.yaml build ${service_list} --no-cache > ${LOG_PATH}/docker_image_build.log
3731

3832
docker images && sleep 1s

AudioQnA/tests/test_compose_on_epyc.sh

Lines changed: 1 addition & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -27,20 +27,8 @@ function build_docker_images() {
2727
docker build --no-cache -t ${REGISTRY}/comps-base:${TAG} --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
2828
popd && sleep 1s
2929

30-
git clone https://github.com/vllm-project/vllm.git
31-
cd ./vllm/
32-
VLLM_VER=v0.10.0
33-
echo "Check out vLLM tag ${VLLM_VER}"
34-
git checkout ${VLLM_VER} &>/dev/null
35-
VLLM_REQ_FILE="requirements/cpu.txt"
36-
if ! grep -q "^transformers" "$VLLM_REQ_FILE"; then
37-
echo "Adding transformers<4.54.0 to $VLLM_REQ_FILE"
38-
echo "transformers<4.54.0" >>"$VLLM_REQ_FILE"
39-
fi
40-
cd ../
41-
4230
echo "Build all the images with --no-cache, check docker_image_build.log for details..."
43-
service_list="audioqna audioqna-ui whisper speecht5 vllm"
31+
service_list="audioqna audioqna-ui whisper speecht5"
4432
docker compose -f build.yaml build ${service_list} --no-cache >${LOG_PATH}/docker_image_build.log
4533

4634
docker images && sleep 1s

AudioQnA/tests/test_compose_on_xeon.sh

Lines changed: 1 addition & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -25,14 +25,8 @@ function build_docker_images() {
2525
docker build --no-cache -t ${REGISTRY}/comps-base:${TAG} --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f Dockerfile .
2626
popd && sleep 1s
2727

28-
git clone https://github.com/vllm-project/vllm.git
29-
cd ./vllm/
30-
VLLM_VER=v0.10.0
31-
echo "Check out vLLM tag ${VLLM_VER}"
32-
git checkout ${VLLM_VER} &> /dev/null && cd ../
33-
3428
echo "Build all the images with --no-cache, check docker_image_build.log for details..."
35-
service_list="audioqna audioqna-ui whisper speecht5 vllm"
29+
service_list="audioqna audioqna-ui whisper speecht5"
3630
docker compose -f build.yaml build ${service_list} --no-cache > ${LOG_PATH}/docker_image_build.log
3731

3832
docker images && sleep 1s

ChatQnA/docker_compose/amd/cpu/epyc/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -90,7 +90,7 @@ services:
9090
HF_HUB_ENABLE_HF_TRANSFER: 0
9191
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
9292
vllm-service:
93-
image: ${REGISTRY:-opea}/vllm:${TAG:-latest}
93+
image: public.ecr.aws/q9t5s3a7/vllm-cpu-release-repo:v0.10.1
9494
container_name: vllm-service
9595
ports:
9696
- "9009:80"

ChatQnA/docker_compose/amd/cpu/epyc/compose_faqgen.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -83,7 +83,7 @@ services:
8383
HF_HUB_ENABLE_HF_TRANSFER: 0
8484
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
8585
vllm-service:
86-
image: ${REGISTRY:-opea}/vllm:${TAG:-latest}
86+
image: public.ecr.aws/q9t5s3a7/vllm-cpu-release-repo:v0.10.1
8787
container_name: vllm-server
8888
ports:
8989
- ${LLM_ENDPOINT_PORT:-9009}:80

ChatQnA/docker_compose/amd/cpu/epyc/compose_milvus.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -147,7 +147,7 @@ services:
147147
command: --model-id ${RERANK_MODEL_ID} --auto-truncate
148148

149149
vllm-service:
150-
image: ${REGISTRY:-opea}/vllm:${TAG:-latest}
150+
image: public.ecr.aws/q9t5s3a7/vllm-cpu-release-repo:v0.10.1
151151
container_name: vllm-service
152152
ports:
153153
- "9009:80"

0 commit comments

Comments
 (0)