Skip to content

Will TensorRT-LLM be available within Triton or will it be a separate server ? #6290

Discussion options

You must be logged in to vote

Very happy to hear that, Matthieu! Thanks for sharing.

Triton will continue to work as the one solution for all of your AI model inferencing. We are working on a TensorRT-LLM backend that you can easily plug your models into.

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@MatthieuToulemont
Comment options

Answer selected by dyastremsky
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants