Using vLLM or TGI #7688
-
Hi, so I have a troubling question where I'm confused to use either vLLM or TGI (https://github.com/huggingface/text-generation-inference). Question:How can I distinguish that a model would work with vLLM or TGI before I am about to bring up the model itself? Is there a way that would tell me if the model would come up successfully? Can I make use of any parameters (which would be available for all models) present in config.json file for the models to say that yes, this model would work with vLLM and not TGI, and vice-versa. FYI: The models I'm looking to host are if HuggingFace format itself. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
The |
Beta Was this translation helpful? Give feedback.
The
architecture
field inconfig.json
should be what you want to cross check: https://docs.vllm.ai/en/latest/models/supported_models.html