How to run tabby on offline-environment with local model? #3517
-
Situation: Attempts:
# Chat model
[model.chat.http]
kind = "openai/chat"
model_name = "my_model"
api_endpoint = "http://localhost:8049/v1"
api_key = "my_token"
# Completion model
[model.completion.http]
kind = "vllm/completion"
model_name = "my_model"
api_endpoint = "http://localhost:8049/v1"
api_key = "my_token"
prompt_template = "<|fim_prefix|><|fim_suffix|>{suffix}<|fim_middle|>{prefix}" # SPM Then I ran with
I've come to realize that this release might need to be made available online. Consequently, I visited this blog at https://tabby.tabbyml.com/blog/2024/03/25/deploy-tabby-in-air-gapped-environment-with-docker/ and gave it a try. Question: |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments
-
Is there anyone can offer some help? |
Beta Was this translation helpful? Give feedback.
-
Hi @wapleeeeee, Tabby requires the use of an embedding model in addition to the other models. You need to add the HTTP embedding model to the configuration for Tabby to operate in an offline environment. for example: [model.embedding.http]
kind = "openai/embedding"
model_name = "text-embedding-3-small"
api_endpoint = "http://localhost:8099/v1"
api_key = "apikey" We have also created an example for vLLM at https://tabby.tabbyml.com/docs/references/models-http-api/vllm/. |
Beta Was this translation helpful? Give feedback.
Hi @wapleeeeee, Tabby requires the use of an embedding model in addition to the other models. You need to add the HTTP embedding model to the configuration for Tabby to operate in an offline environment.
for example:
We have also created an example for vLLM at https://tabby.tabbyml.com/docs/references/models-http-api/vllm/.