Skip to content

Commit

Permalink
Add table to list port, endpoint, framework, model, serving, and hard…
Browse files Browse the repository at this point in the history
…ware for each microservice in ChatQnA (opea-project#697)

Signed-off-by: srinarayan-srikanthan <[email protected]>
Signed-off-by: Chun Tao <[email protected]>
Signed-off-by: letonghan <[email protected]>
Signed-off-by: Ye, Xinyu <[email protected]>
Signed-off-by: chensuyue <[email protected]>
Signed-off-by: Yue, Wenjiao <[email protected]>
Signed-off-by: Lianhao Lu <[email protected]>
Co-authored-by: srinarayan-srikanthan <[email protected]>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Co-authored-by: Letong Han <[email protected]>
Co-authored-by: XinyuYe-Intel <[email protected]>
Co-authored-by: chen, suyue <[email protected]>
Co-authored-by: Zhenzhong1 <[email protected]>
Co-authored-by: WenjiaoYue <[email protected]>
Co-authored-by: Lianhao Lu <[email protected]>
Co-authored-by: Ying Hu <[email protected]>
  • Loading branch information
10 people authored and dmsuehir committed Sep 11, 2024
1 parent b98e2b0 commit 13ba0c1
Showing 1 changed file with 15 additions and 0 deletions.
15 changes: 15 additions & 0 deletions ChatQnA/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -97,6 +97,21 @@ flowchart LR

This ChatQnA use case performs RAG using LangChain, Redis VectorDB and Text Generation Inference on Intel Gaudi2 or Intel XEON Scalable Processors. The Intel Gaudi2 accelerator supports both training and inference for deep learning models in particular for LLMs. Visit [Habana AI products](https://habana.ai/products) for more details.

In the below, we provide a table that describes for each microservice component in the ChatQnA architecture, the default configuration of the open source project, hardware, port, and endpoint.

<details>
<summary><b>Gaudi default compose.yaml</b></summary>

| MicroService | Open Source Project | HW | Port | Endpoint |
| ------------ | ------------------- | ----- | ---- | -------------------- |
| Embedding | Langchain | Xeon | 6000 | /v1/embaddings |
| Retriever | Langchain, Redis | Xeon | 7000 | /v1/retrieval |
| Reranking | Langchain, TEI | Gaudi | 8000 | /v1/reranking |
| LLM | Langchain, TGI | Gaudi | 9000 | /v1/chat/completions |
| Dataprep | Redis, Langchain | Xeon | 6007 | /v1/dataprep |

</details>

## Deploy ChatQnA Service

The ChatQnA service can be effortlessly deployed on either Intel Gaudi2 or Intel XEON Scalable Processors.
Expand Down

0 comments on commit 13ba0c1

Please sign in to comment.