diff --git a/app.toml.template b/app.toml.template index 19f0a27f4..34dc86178 100644 --- a/app.toml.template +++ b/app.toml.template @@ -15,14 +15,12 @@ CPUReservation = "10" MemoryReservation = "118Gi" MemoryLimit = "118Gi" GPUCount = 1 -GPUType = "training-gpu" VolumeMount = "/home/llmstudio/mount" VolumeSize = "1Ti" ResourceVolumeSize = "1Ti" EnableSHM = true EnableOIDC = true RoutingMode = "BASE_URL" -CustomImage = "docker.io/h2oairelease/h2oai-llmstudio-app:v{{VERSION}}" [[Env]] Name = "H2O_LLM_STUDIO_WORKDIR" @@ -39,3 +37,7 @@ Value = "True" [[Env]] Name = "H2O_WAVE_PRIVATE_DIR" Value = "/download/@/home/llmstudio/mount/output/download" + +[[Env]] +Name = "HF_HUB_ENABLE_HF_TRANSFER" +Value = "0" diff --git a/llm_studio/app_utils/config.py b/llm_studio/app_utils/config.py index 2c1b023a1..88d055f6f 100644 --- a/llm_studio/app_utils/config.py +++ b/llm_studio/app_utils/config.py @@ -3,6 +3,7 @@ from types import SimpleNamespace import toml +from huggingface_hub.constants import _is_true toml_root_dir = os.path.abspath( os.path.join(os.path.dirname(os.path.abspath(__file__)), "../..") @@ -160,8 +161,8 @@ def get_size(x): "default_wandb_project": os.getenv("WANDB_PROJECT", ""), "default_wandb_entity": os.getenv("WANDB_ENTITY", ""), "default_huggingface_api_token": os.getenv("HF_TOKEN", ""), - "default_hf_hub_enable_hf_transfer": os.getenv( - "HF_HUB_ENABLE_HF_TRANSFER", True + "default_hf_hub_enable_hf_transfer": _is_true( + os.getenv("HF_HUB_ENABLE_HF_TRANSFER", "1") ), "default_openai_azure": os.getenv("OPENAI_API_TYPE", "open_ai") == "azure", "default_openai_api_token": os.getenv("OPENAI_API_KEY", ""),