Skip to content

Commit

Permalink
Add table to list port, endpoint, framework, model, serving, and hard…
Browse files Browse the repository at this point in the history
…ware for each microservice in ChatQnA (opea-project#697)

Signed-off-by: srinarayan-srikanthan <[email protected]>
Signed-off-by: Chun Tao <[email protected]>
Signed-off-by: letonghan <[email protected]>
Signed-off-by: Ye, Xinyu <[email protected]>
Signed-off-by: chensuyue <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Lianhao Lu <[email protected]>
Co-authored-by: srinarayan-srikanthan <[email protected]>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Co-authored-by: Letong Han <[email protected]>
Co-authored-by: XinyuYe-Intel <[email protected]>
Co-authored-by: chen, suyue <[email protected]>
Co-authored-by: Zhenzhong1 <[email protected]>
Co-authored-by: WenjiaoYue <[email protected]>
Co-authored-by: Lianhao Lu <[email protected]>
Co-authored-by: Ying Hu <[email protected]>
  • Loading branch information
10 people authored Sep 11, 2024
1 parent 5c67204 commit 1a934af
Showing 1 changed file with 15 additions and 0 deletions.
15 changes: 15 additions & 0 deletions ChatQnA/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -97,6 +97,21 @@ flowchart LR

This ChatQnA use case performs RAG using LangChain, Redis VectorDB and Text Generation Inference on Intel Gaudi2 or Intel XEON Scalable Processors. The Intel Gaudi2 accelerator supports both training and inference for deep learning models in particular for LLMs. Visit [Habana AI products](https://habana.ai/products) for more details.

In the below, we provide a table that describes for each microservice component in the ChatQnA architecture, the default configuration of the open source project, hardware, port, and endpoint.

<details>
<summary><b>Gaudi default compose.yaml</b></summary>

| MicroService | Open Source Project | HW | Port | Endpoint |
| ------------ | ------------------- | ----- | ---- | -------------------- |
| Embedding | Langchain | Xeon | 6000 | /v1/embaddings |
| Retriever | Langchain, Redis | Xeon | 7000 | /v1/retrieval |
| Reranking | Langchain, TEI | Gaudi | 8000 | /v1/reranking |
| LLM | Langchain, TGI | Gaudi | 9000 | /v1/chat/completions |
| Dataprep | Redis, Langchain | Xeon | 6007 | /v1/dataprep |

</details>

## Deploy ChatQnA Service

The ChatQnA service can be effortlessly deployed on either Intel Gaudi2 or Intel XEON Scalable Processors.
Expand Down

0 comments on commit 1a934af

Please sign in to comment.