You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -158,7 +145,7 @@ function validate_megaservice() {
158
145
local DOCKER_NAME="docsum-backend-server"
159
146
local EXPECTED_RESULT="[DONE]"
160
147
local INPUT_DATA="messages=Text Embeddings Inference (TEI) is a toolkit for deploying and serving open source text embeddings and sequence classification models. TEI enables high-performance extraction for the most popular models, including FlagEmbedding, Ember, GTE and E5."
161
-
local URL="${host_ip}:${DOCSUM_BACKEND_SERVER_PORT}/v1/docsum"
148
+
local URL="${HOST_IP}:${DOCSUM_BACKEND_SERVER_PORT}/v1/docsum"
162
149
local DATA_TYPE="type=text"
163
150
164
151
local HTTP_STATUS=$(curl -s -o /dev/null -w "%{http_code}" -X POST -F "$DATA_TYPE" -F "$INPUT_DATA" -H 'Content-Type: multipart/form-data'"$URL")
@@ -188,23 +175,23 @@ function validate_megaservice_json() {
188
175
echo""
189
176
echo">>> Checking text data with Content-Type: application/json"
'{"type": "text", "messages": "Text Embeddings Inference (TEI) is a toolkit for deploying and serving open source text embeddings and sequence classification models. TEI enables high-performance extraction for the most popular models, including FlagEmbedding, Ember, GTE and E5."}'
0 commit comments