Skip to content

Commit

Permalink
feat: apply new labels
Browse files Browse the repository at this point in the history
  • Loading branch information
bojiang committed Jul 12, 2024
1 parent 6424528 commit 6ac4f5d
Show file tree
Hide file tree
Showing 1,998 changed files with 3,241 additions and 3,288 deletions.
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# gemma:2b-instruct-fp16-26b3
# gemma:2b-instruct-fp16-0856

[![pypi_status](https://img.shields.io/badge/BentoML-1.2.19-informational)](https://pypi.org/project/BentoML)
[![documentation_status](https://readthedocs.org/projects/bentoml/badge/?version=latest)](https://docs.bentoml.com/)
Expand Down
Original file line number Diff line number Diff line change
@@ -1,12 +1,11 @@
service: service:VLLM
name: gemma
version: 2b-instruct-fp16-26b3
version: 2b-instruct-fp16-0856
bentoml_version: 1.2.19
creation_time: '2024-07-08T07:57:26.100524+00:00'
creation_time: '2024-07-12T13:40:09.050778+00:00'
labels:
model_name: google/gemma-2b-it
openllm_alias: 2b,2b-instruct
openllm_hf_model_id: google/gemma-2b-it
owner: bentoml-team
platforms: linux
source: https://github.com/bentoml/openllm-models-feed/tree/main/source/vllm-chat
models: []
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ inflection==0.5.1
interegular==0.3.3
jinja2==3.1.4
joblib==1.4.2
jsonschema==4.22.0
jsonschema==4.23.0
jsonschema-specifications==2023.12.1
lark==1.1.9
llvmlite==0.43.0
Expand Down Expand Up @@ -70,7 +70,7 @@ nvidia-ml-py==11.525.150
nvidia-nccl-cu12==2.20.5
nvidia-nvjitlink-cu12==12.5.82
nvidia-nvtx-cu12==12.1.105
openai==1.35.10
openai==1.35.13
opentelemetry-api==1.20.0
opentelemetry-instrumentation==0.41b0
opentelemetry-instrumentation-aiohttp-client==0.41b0
Expand Down Expand Up @@ -100,12 +100,12 @@ python-json-logger==2.0.7
python-multipart==0.0.9
pyyaml==6.0.1
pyzmq==26.0.3
ray==2.31.0
ray==2.32.0
referencing==0.35.1
regex==2024.5.15
requests==2.32.3
rich==13.7.1
rpds-py==0.18.1
rpds-py==0.19.0
safetensors==0.4.3
schema==0.7.7
scipy==1.13.1
Expand All @@ -115,7 +115,7 @@ simple-di==0.1.5
six==1.16.0
sniffio==1.3.1
starlette==0.37.2
sympy==1.12.1
sympy==1.13.0
tiktoken==0.7.0
tokenizers==0.19.1
tomli==2.0.1
Expand Down Expand Up @@ -144,4 +144,4 @@ zipp==3.19.2

# The following packages are considered to be unsafe in a requirements file:
pip==24.1.2
setuptools==70.2.0
setuptools==70.3.0
Original file line number Diff line number Diff line change
Expand Up @@ -5,8 +5,8 @@
max_model_len: 2048
model: google/gemma-2b-it
extra_labels:
model_name: google/gemma-2b-it
openllm_alias: 2b,2b-instruct
openllm_hf_model_id: google/gemma-2b-it
project: vllm-chat
service_config:
name: gemma
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,9 +28,8 @@ include:
- chat_templates/chat_templates/*.jinja
- chat_templates/generation_configs/*.json
labels:
model_name: google/gemma-2b-it
openllm_alias: 2b,2b-instruct
openllm_hf_model_id: google/gemma-2b-it
owner: bentoml-team
platforms: linux
source: https://github.com/bentoml/openllm-models-feed/tree/main/source/vllm-chat
models: []
Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# gemma:7b-instruct-awq-4bit-d670
# gemma:7b-instruct-awq-4bit-d11b

[![pypi_status](https://img.shields.io/badge/BentoML-1.2.19-informational)](https://pypi.org/project/BentoML)
[![documentation_status](https://readthedocs.org/projects/bentoml/badge/?version=latest)](https://docs.bentoml.com/)
Expand Down
Original file line number Diff line number Diff line change
@@ -1,12 +1,11 @@
service: service:VLLM
name: gemma
version: 7b-instruct-awq-4bit-d670
version: 7b-instruct-awq-4bit-d11b
bentoml_version: 1.2.19
creation_time: '2024-07-08T07:59:18.619602+00:00'
creation_time: '2024-07-12T13:42:23.707025+00:00'
labels:
model_name: casperhansen/gemma-7b-it-awq
openllm_alias: 7b-4bit,7b-instruct-4bit
openllm_hf_model_id: casperhansen/gemma-7b-it-awq
owner: bentoml-team
platforms: linux
source: https://github.com/bentoml/openllm-models-feed/tree/main/source/vllm-chat
models: []
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ inflection==0.5.1
interegular==0.3.3
jinja2==3.1.4
joblib==1.4.2
jsonschema==4.22.0
jsonschema==4.23.0
jsonschema-specifications==2023.12.1
lark==1.1.9
llvmlite==0.43.0
Expand Down Expand Up @@ -70,7 +70,7 @@ nvidia-ml-py==11.525.150
nvidia-nccl-cu12==2.20.5
nvidia-nvjitlink-cu12==12.5.82
nvidia-nvtx-cu12==12.1.105
openai==1.35.10
openai==1.35.13
opentelemetry-api==1.20.0
opentelemetry-instrumentation==0.41b0
opentelemetry-instrumentation-aiohttp-client==0.41b0
Expand Down Expand Up @@ -100,12 +100,12 @@ python-json-logger==2.0.7
python-multipart==0.0.9
pyyaml==6.0.1
pyzmq==26.0.3
ray==2.31.0
ray==2.32.0
referencing==0.35.1
regex==2024.5.15
requests==2.32.3
rich==13.7.1
rpds-py==0.18.1
rpds-py==0.19.0
safetensors==0.4.3
schema==0.7.7
scipy==1.13.1
Expand All @@ -115,7 +115,7 @@ simple-di==0.1.5
six==1.16.0
sniffio==1.3.1
starlette==0.37.2
sympy==1.12.1
sympy==1.13.0
tiktoken==0.7.0
tokenizers==0.19.1
tomli==2.0.1
Expand Down Expand Up @@ -144,4 +144,4 @@ zipp==3.19.2

# The following packages are considered to be unsafe in a requirements file:
pip==24.1.2
setuptools==70.2.0
setuptools==70.3.0
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,8 @@
model: casperhansen/gemma-7b-it-awq
quantization: awq
extra_labels:
model_name: casperhansen/gemma-7b-it-awq
openllm_alias: 7b-4bit,7b-instruct-4bit
openllm_hf_model_id: casperhansen/gemma-7b-it-awq
project: vllm-chat
service_config:
name: gemma
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,9 +28,8 @@ include:
- chat_templates/chat_templates/*.jinja
- chat_templates/generation_configs/*.json
labels:
model_name: casperhansen/gemma-7b-it-awq
openllm_alias: 7b-4bit,7b-instruct-4bit
openllm_hf_model_id: Qwen/Qwen2-7B-Instruct-AWQ
owner: bentoml-team
platforms: linux
source: https://github.com/bentoml/openllm-models-feed/tree/main/source/vllm-chat
models: []
Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# gemma:7b-instruct-fp16-ae3e
# gemma:7b-instruct-fp16-3e1c

[![pypi_status](https://img.shields.io/badge/BentoML-1.2.19-informational)](https://pypi.org/project/BentoML)
[![documentation_status](https://readthedocs.org/projects/bentoml/badge/?version=latest)](https://docs.bentoml.com/)
Expand Down
Original file line number Diff line number Diff line change
@@ -1,12 +1,11 @@
service: service:VLLM
name: gemma
version: 7b-instruct-fp16-ae3e
version: 7b-instruct-fp16-3e1c
bentoml_version: 1.2.19
creation_time: '2024-07-08T07:58:22.853090+00:00'
creation_time: '2024-07-12T13:41:08.692955+00:00'
labels:
model_name: google/gemma-7b-it
openllm_alias: 7b,7b-instruct
openllm_hf_model_id: google/gemma-7b-it
owner: bentoml-team
platforms: linux
source: https://github.com/bentoml/openllm-models-feed/tree/main/source/vllm-chat
models: []
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ inflection==0.5.1
interegular==0.3.3
jinja2==3.1.4
joblib==1.4.2
jsonschema==4.22.0
jsonschema==4.23.0
jsonschema-specifications==2023.12.1
lark==1.1.9
llvmlite==0.43.0
Expand Down Expand Up @@ -70,7 +70,7 @@ nvidia-ml-py==11.525.150
nvidia-nccl-cu12==2.20.5
nvidia-nvjitlink-cu12==12.5.82
nvidia-nvtx-cu12==12.1.105
openai==1.35.10
openai==1.35.13
opentelemetry-api==1.20.0
opentelemetry-instrumentation==0.41b0
opentelemetry-instrumentation-aiohttp-client==0.41b0
Expand Down Expand Up @@ -100,12 +100,12 @@ python-json-logger==2.0.7
python-multipart==0.0.9
pyyaml==6.0.1
pyzmq==26.0.3
ray==2.31.0
ray==2.32.0
referencing==0.35.1
regex==2024.5.15
requests==2.32.3
rich==13.7.1
rpds-py==0.18.1
rpds-py==0.19.0
safetensors==0.4.3
schema==0.7.7
scipy==1.13.1
Expand All @@ -115,7 +115,7 @@ simple-di==0.1.5
six==1.16.0
sniffio==1.3.1
starlette==0.37.2
sympy==1.12.1
sympy==1.13.0
tiktoken==0.7.0
tokenizers==0.19.1
tomli==2.0.1
Expand Down Expand Up @@ -144,4 +144,4 @@ zipp==3.19.2

# The following packages are considered to be unsafe in a requirements file:
pip==24.1.2
setuptools==70.2.0
setuptools==70.3.0
Original file line number Diff line number Diff line change
Expand Up @@ -5,8 +5,8 @@
max_model_len: 2048
model: google/gemma-7b-it
extra_labels:
model_name: google/gemma-7b-it
openllm_alias: 7b,7b-instruct
openllm_hf_model_id: google/gemma-7b-it
project: vllm-chat
service_config:
name: gemma
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,9 +28,8 @@ include:
- chat_templates/chat_templates/*.jinja
- chat_templates/generation_configs/*.json
labels:
model_name: google/gemma-7b-it
openllm_alias: 7b,7b-instruct
openllm_hf_model_id: google/gemma-7b-it
owner: bentoml-team
platforms: linux
source: https://github.com/bentoml/openllm-models-feed/tree/main/source/vllm-chat
models: []
Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# llama2:70b-chat-fp16-b0de
# llama2:13b-chat-fp16-921b

[![pypi_status](https://img.shields.io/badge/BentoML-1.2.19-informational)](https://pypi.org/project/BentoML)
[![documentation_status](https://readthedocs.org/projects/bentoml/badge/?version=latest)](https://docs.bentoml.com/)
Expand Down
Original file line number Diff line number Diff line change
@@ -1,12 +1,11 @@
service: service:VLLM
name: llama2
version: 13b-chat-fp16-4059
version: 13b-chat-fp16-921b
bentoml_version: 1.2.19
creation_time: '2024-07-08T07:49:27.527664+00:00'
creation_time: '2024-07-12T13:30:51.101256+00:00'
labels:
model_name: meta-llama/Llama-2-13b-chat-hf
openllm_alias: 13b,13b-chat
openllm_hf_model_id: meta-llama/Llama-2-13b-chat-hf
owner: bentoml-team
platforms: linux
source: https://github.com/bentoml/openllm-models-feed/tree/main/source/vllm-chat
models: []
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ inflection==0.5.1
interegular==0.3.3
jinja2==3.1.4
joblib==1.4.2
jsonschema==4.22.0
jsonschema==4.23.0
jsonschema-specifications==2023.12.1
lark==1.1.9
llvmlite==0.43.0
Expand Down Expand Up @@ -70,7 +70,7 @@ nvidia-ml-py==11.525.150
nvidia-nccl-cu12==2.20.5
nvidia-nvjitlink-cu12==12.5.82
nvidia-nvtx-cu12==12.1.105
openai==1.35.10
openai==1.35.13
opentelemetry-api==1.20.0
opentelemetry-instrumentation==0.41b0
opentelemetry-instrumentation-aiohttp-client==0.41b0
Expand Down Expand Up @@ -100,12 +100,12 @@ python-json-logger==2.0.7
python-multipart==0.0.9
pyyaml==6.0.1
pyzmq==26.0.3
ray==2.31.0
ray==2.32.0
referencing==0.35.1
regex==2024.5.15
requests==2.32.3
rich==13.7.1
rpds-py==0.18.1
rpds-py==0.19.0
safetensors==0.4.3
schema==0.7.7
scipy==1.13.1
Expand All @@ -115,7 +115,7 @@ simple-di==0.1.5
six==1.16.0
sniffio==1.3.1
starlette==0.37.2
sympy==1.12.1
sympy==1.13.0
tiktoken==0.7.0
tokenizers==0.19.1
tomli==2.0.1
Expand Down Expand Up @@ -144,4 +144,4 @@ zipp==3.19.2

# The following packages are considered to be unsafe in a requirements file:
pip==24.1.2
setuptools==70.2.0
setuptools==70.3.0
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,8 @@
max_model_len: 1024
model: meta-llama/Llama-2-13b-chat-hf
extra_labels:
model_name: meta-llama/Llama-2-13b-chat-hf
openllm_alias: 13b,13b-chat
openllm_hf_model_id: meta-llama/Llama-2-13b-chat-hf
project: vllm-chat
service_config:
name: llama2
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,9 +28,8 @@ include:
- chat_templates/chat_templates/*.jinja
- chat_templates/generation_configs/*.json
labels:
model_name: meta-llama/Llama-2-13b-chat-hf
openllm_alias: 13b,13b-chat
openllm_hf_model_id: meta-llama/Llama-2-13b-chat-hf
owner: bentoml-team
platforms: linux
source: https://github.com/bentoml/openllm-models-feed/tree/main/source/vllm-chat
models: []
Expand Down
Loading

0 comments on commit 6ac4f5d

Please sign in to comment.