The Translation example demonstrates the implementation of language translation using OPEA component-level microservices.
The architecture of the Translation Application is illustrated below:
The Translation example is implemented using the component-level microservices defined in GenAIComps. The flow chart below shows the information flow between different microservices for this example.
---
config:
flowchart:
nodeSpacing: 400
rankSpacing: 100
curve: linear
themeVariables:
fontSize: 50px
---
flowchart LR
%% Colors %%
classDef blue fill:#ADD8E6,stroke:#ADD8E6,stroke-width:2px,fill-opacity:0.5
classDef orange fill:#FBAA60,stroke:#ADD8E6,stroke-width:2px,fill-opacity:0.5
classDef orchid fill:#C26DBC,stroke:#ADD8E6,stroke-width:2px,fill-opacity:0.5
classDef invisible fill:transparent,stroke:transparent;
style Translation-MegaService stroke:#000000
%% Subgraphs %%
subgraph Translation-MegaService["Translation MegaService "]
direction LR
LLM([LLM MicroService]):::blue
end
subgraph UserInterface[" User Interface "]
direction LR
a([User Input Query]):::orchid
UI([UI server<br>]):::orchid
end
LLM_gen{{LLM Service <br>}}
GW([Translation GateWay<br>]):::orange
NG([Nginx MicroService]):::blue
%% Questions interaction
direction LR
a[User Input Query] --> UI
a[User Input Query] --> |Need Proxy Server|NG
NG --> UI
UI --> GW
GW <==> Translation-MegaService
%% Embedding service flow
direction LR
LLM <-.-> LLM_gen
This Translation use case performs Language Translation Inference across multiple platforms. Currently, we provide the example for Intel Gaudi2, Intel Xeon Scalable Processors and AMD EPYC™ Processors, and we invite contributions from other hardware vendors to expand OPEA ecosystem.
The table below lists the available deployment options and their implementation details for different hardware platforms.
| Platform | Deployment Method | Link |
|---|---|---|
| Intel Xeon | Docker compose | Deployment on Xeon |
| Intel Gaudi2 | Docker compose | Deployment on Gaudi |
| AMD EPYC | Docker compose | Deployment on AMD EPYC |
| AMD ROCm | Docker compose | Deployment on AMD Rocm |
| Deploy Method | LLM Engine | LLM Model | Hardware |
|---|---|---|---|
| Docker Compose | vLLM, TGI | haoranxu/ALMA-13B | Intel Gaudi |
| Docker Compose | vLLM, TGI | haoranxu/ALMA-13B | Intel Xeon |
| Docker Compose | vLLM, TGI | haoranxu/ALMA-13B | AMD Xeon |
| Docker Compose | vLLM, TGI | haoranxu/ALMA-13B | AMD EPYC |
| Helm Charts | vLLM, TGI | haoranxu/ALMA-13B | Intel Gaudi |
| Helm Charts | vLLM, TGI | haoranxu/ALMA-13B | Intel Xeon |
