diff --git a/ChatQnA/docker_compose/amd/gpu/rocm/README.md b/ChatQnA/docker_compose/amd/gpu/rocm/README.md index 9e18d0f61..9ef30d2a1 100644 --- a/ChatQnA/docker_compose/amd/gpu/rocm/README.md +++ b/ChatQnA/docker_compose/amd/gpu/rocm/README.md @@ -290,7 +290,7 @@ docker compose up -d Try the command below to check whether the TGI service is ready. ```bash - docker logs ${CONTAINER_ID} | grep Connected + docker logs chatqna-tgi-server | grep Connected ``` If the service is ready, you will get the response like below. diff --git a/ChatQnA/docker_compose/intel/hpu/gaudi/README.md b/ChatQnA/docker_compose/intel/hpu/gaudi/README.md index b083b3d40..9e2b5b545 100644 --- a/ChatQnA/docker_compose/intel/hpu/gaudi/README.md +++ b/ChatQnA/docker_compose/intel/hpu/gaudi/README.md @@ -314,7 +314,7 @@ For validation details, please refer to [how-to-validate_service](./how_to_valid Try the command below to check whether the LLM serving is ready. ```bash - docker logs tgi-service | grep Connected + docker logs tgi-gaudi-server | grep Connected ``` If the service is ready, you will get the response like below. @@ -327,7 +327,7 @@ For validation details, please refer to [how-to-validate_service](./how_to_valid ```bash # TGI service - curl http://${host_ip}:9009/v1/chat/completions \ + curl http://${host_ip}:8005/v1/chat/completions \ -X POST \ -d '{"model": ${LLM_MODEL_ID}, "messages": [{"role": "user", "content": "What is Deep Learning?"}], "max_tokens":17}' \ -H 'Content-Type: application/json' @@ -335,7 +335,7 @@ For validation details, please refer to [how-to-validate_service](./how_to_valid ```bash # vLLM Service - curl http://${host_ip}:9009/v1/chat/completions \ + curl http://${host_ip}:8007/v1/chat/completions \ -H "Content-Type: application/json" \ -d '{"model": ${LLM_MODEL_ID}, "messages": [{"role": "user", "content": "What is Deep Learning?"}]}' ``` diff --git a/ChatQnA/docker_compose/nvidia/gpu/README.md b/ChatQnA/docker_compose/nvidia/gpu/README.md index 686ead52d..92b7a26e7 100644 --- a/ChatQnA/docker_compose/nvidia/gpu/README.md +++ b/ChatQnA/docker_compose/nvidia/gpu/README.md @@ -273,7 +273,7 @@ docker compose up -d Try the command below to check whether the TGI service is ready. ```bash - docker logs ${CONTAINER_ID} | grep Connected + docker logs tgi-server | grep Connected ``` If the service is ready, you will get the response like below. @@ -285,7 +285,7 @@ docker compose up -d Then try the `cURL` command below to validate TGI. ```bash - curl http://${host_ip}:9009/v1/chat/completions \ + curl http://${host_ip}:8008/v1/chat/completions \ -X POST \ -d '{"model": "Intel/neural-chat-7b-v3-3", "messages": [{"role": "user", "content": "What is Deep Learning?"}], "max_tokens":17}' \ -H 'Content-Type: application/json'