Skip to content

Commit 0bc2983

Browse files
committed
Add UI guide for Gaudi documentation.
Signed-off-by: ZePan110 <ze.pan@intel.com>
1 parent e089294 commit 0bc2983

1 file changed

Lines changed: 37 additions & 0 deletions

File tree

  • ChatQnA/docker_compose/intel/hpu/gaudi

ChatQnA/docker_compose/intel/hpu/gaudi/README.md

Lines changed: 37 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -294,6 +294,43 @@ The `tgi-guardrails-service` uses the `GUARDRAILS_MODEL_ID` parameter to select
294294

295295
The `vllm-guardrails-service` uses the `GUARDRAILS_MODEL_ID` parameter to select a [supported model](https://docs.vllm.ai/en/latest/models/supported_models.html) for the associated `opea/vllm-gaudi:latest` image. It uses the `NUM_CARDS` parameter.
296296

297+
## ChatQnA with Conversational UI (Optional)
298+
299+
To access the Conversational UI (react based) frontend, modify the UI service in the `compose` file used to deploy. Replace `chaqna-xeon-ui-server` service with the `chatqna-xeon-conversation-ui-server` service as per the config below:
300+
301+
```yaml
302+
chatqna-xeon-conversation-ui-server:
303+
image: opea/chatqna-conversation-ui:latest
304+
container_name: chatqna-xeon-conversation-ui-server
305+
environment:
306+
- APP_BACKEND_SERVICE_ENDPOINT=${BACKEND_SERVICE_ENDPOINT}
307+
- APP_DATA_PREP_SERVICE_URL=${DATAPREP_SERVICE_ENDPOINT}
308+
ports:
309+
- "5174:80"
310+
depends_on:
311+
- chaqna-xeon-backend-server
312+
ipc: host
313+
restart: always
314+
```
315+
316+
Once the services are up, open the following URL in the browser: http://{host_ip}:5174. By default, the UI runs on port 80 internally. If the developer prefers to use a different host port to access the frontend, it can be modified by port mapping in the `compose.yaml` file as shown below:
317+
318+
```yaml
319+
chaqna-gaudi-conversation-ui-server:
320+
image: opea/chatqna-conversation-ui:latest
321+
...
322+
ports:
323+
- "80:80"
324+
```
325+
326+
Here is an example of running ChatQnA (default UI):
327+
328+
![project-screenshot](../../../../assets/img/chat_ui_response.png)
329+
330+
Here is an example of running ChatQnA with Conversational UI (React):
331+
332+
![project-screenshot](../../../../assets/img/conversation_ui_response.png)
333+
297334
## Conclusion
298335

299336
In examining the various services and configurations across different deployments, developers should gain a comprehensive understanding of how each component contributes to the overall functionality and performance of a ChatQnA pipeline on an Intel® Gaudi® platform. Key services such as the `vllm-service`, `tei-embedding-service`, `tei-reranking-service`, `tgi-guardrails-service`and `vllm-guardrails-service` each consume Gaudi accelerators, leveraging specific models and hardware resources to optimize their respective tasks. The `LLM_MODEL_ID`, `EMBEDDING_MODEL_ID`, `RERANK_MODEL_ID`, and `GUARDRAILS_MODEL_ID` parameters specify the models used, directly impacting the quality and effectiveness of language processing, embedding, reranking, and safety operations.

0 commit comments

Comments
 (0)