Skip to content

Commit

Permalink
Update to latest version (#32)
Browse files Browse the repository at this point in the history
  • Loading branch information
oandreeva-nv authored Feb 12, 2024
1 parent 65821ed commit 6f0afff
Showing 1 changed file with 6 additions and 6 deletions.
12 changes: 6 additions & 6 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
<!--
# Copyright 2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
# Copyright 2023-2024, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions
Expand Down Expand Up @@ -30,8 +30,8 @@

**LATEST RELEASE: You are currently on the main branch which tracks
under-development progress towards the next release. The current release branch
is [r23.12](https://github.com/triton-inference-server/vllm_backend/tree/r23.12)
and which corresponds to the 23.12 container release on
is [r24.01](https://github.com/triton-inference-server/vllm_backend/tree/r24.01)
and which corresponds to the 24.01 container release on
[NVIDIA GPU Cloud (NGC)](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/tritonserver).**

# vLLM Backend
Expand Down Expand Up @@ -96,9 +96,9 @@ A sample command to build a Triton Server container with all options enabled is
--endpoint=grpc
--endpoint=sagemaker
--endpoint=vertex-ai
--upstream-container-version=23.12
--backend=python:r23.12
--backend=vllm:r23.12
--upstream-container-version=24.01
--backend=python:r24.01
--backend=vllm:r24.01
```

### Option 3. Add the vLLM Backend to the Default Triton Container
Expand Down

0 comments on commit 6f0afff

Please sign in to comment.