forked from huggingface/chat-ui
-
Notifications
You must be signed in to change notification settings - Fork 1
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
Showing
75 changed files
with
1,385 additions
and
770 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,151 @@ | ||
# template used in production for HuggingChat. | ||
|
||
MODELS=`[ | ||
{ | ||
"name": "meta-llama/Llama-2-70b-chat-hf", | ||
"description": "The latest and biggest model from Meta, fine-tuned for chat.", | ||
"websiteUrl": "https://ai.meta.com/llama/", | ||
"userMessageToken": "", | ||
"userMessageEndToken": " [/INST] ", | ||
"assistantMessageToken": "", | ||
"assistantMessageEndToken": " </s><s>[INST] ", | ||
"preprompt": " ", | ||
"chatPromptTemplate" : "<s>[INST] <<SYS>>\n{{preprompt}}\n<</SYS>>\n\n{{#each messages}}{{#ifUser}}{{content}} [/INST] {{/ifUser}}{{#ifAssistant}}{{content}} </s><s>[INST] {{/ifAssistant}}{{/each}}", | ||
"promptExamples": [ | ||
{ | ||
"title": "Write an email from bullet list", | ||
"prompt": "As a restaurant owner, write a professional email to the supplier to get these products every week: \n\n- Wine (x10)\n- Eggs (x24)\n- Bread (x12)" | ||
}, { | ||
"title": "Code a snake game", | ||
"prompt": "Code a basic snake game in python, give explanations for each step." | ||
}, { | ||
"title": "Assist in a task", | ||
"prompt": "How do I make a delicious lemon cheesecake?" | ||
} | ||
], | ||
"parameters": { | ||
"temperature": 0.1, | ||
"top_p": 0.95, | ||
"repetition_penalty": 1.2, | ||
"top_k": 50, | ||
"truncate": 1000, | ||
"max_new_tokens": 1024 | ||
} | ||
}, | ||
{ | ||
"name": "codellama/CodeLlama-34b-Instruct-hf", | ||
"displayName": "codellama/CodeLlama-34b-Instruct-hf", | ||
"description": "Code Llama, a state of the art code model from Meta.", | ||
"websiteUrl": "https://about.fb.com/news/2023/08/code-llama-ai-for-coding/", | ||
"userMessageToken": "", | ||
"userMessageEndToken": " [/INST] ", | ||
"assistantMessageToken": "", | ||
"assistantMessageEndToken": " </s><s>[INST] ", | ||
"preprompt": " ", | ||
"chatPromptTemplate" : "<s>[INST] <<SYS>>\n{{preprompt}}\n<</SYS>>\n\n{{#each messages}}{{#ifUser}}{{content}} [/INST] {{/ifUser}}{{#ifAssistant}}{{content}} </s><s>[INST] {{/ifAssistant}}{{/each}}", | ||
"promptExamples": [ | ||
{ | ||
"title": "Fibonacci in Python", | ||
"prompt": "Write a python function to calculate the nth fibonacci number." | ||
}, { | ||
"title": "JavaScript promises", | ||
"prompt": "How can I wait for multiple JavaScript promises to fulfill before doing something with their values?" | ||
}, { | ||
"title": "Rust filesystem", | ||
"prompt": "How can I load a file from disk in Rust?" | ||
} | ||
], | ||
"parameters": { | ||
"temperature": 0.1, | ||
"top_p": 0.95, | ||
"repetition_penalty": 1.2, | ||
"top_k": 50, | ||
"truncate": 1000, | ||
"max_new_tokens": 2048 | ||
} | ||
}, | ||
{ | ||
"name": "tiiuae/falcon-180B-chat", | ||
"displayName": "tiiuae/falcon-180B-chat", | ||
"description": "Falcon-180B is a 180B parameters causal decoder-only model built by TII and trained on 3,500B tokens.", | ||
"websiteUrl": "https://www.tii.ae/news/technology-innovation-institute-introduces-worlds-most-powerful-open-llm-falcon-180b", | ||
"preprompt": " ", | ||
"chatPromptTemplate": "System: {{preprompt}}\nUser:{{#each messages}}{{#ifUser}}{{content}}\nFalcon:{{/ifUser}}{{#ifAssistant}}{{content}}\nUser:{{/ifAssistant}}{{/each}}", | ||
"parameters": { | ||
"temperature": 0.1, | ||
"top_p": 0.95, | ||
"repetition_penalty": 1.2, | ||
"top_k": 50, | ||
"truncate": 1000, | ||
"max_new_tokens": 1024, | ||
"stop": ["User:"] | ||
}, | ||
"promptExamples": [ | ||
{ | ||
"title": "Write an email from bullet list", | ||
"prompt": "As a restaurant owner, write a professional email to the supplier to get these products every week: \n\n- Wine (x10)\n- Eggs (x24)\n- Bread (x12)" | ||
}, { | ||
"title": "Code a snake game", | ||
"prompt": "Code a basic snake game in python, give explanations for each step." | ||
}, { | ||
"title": "Assist in a task", | ||
"prompt": "How do I make a delicious lemon cheesecake?" | ||
} | ||
] | ||
}, | ||
{ | ||
"name": "mistralai/Mistral-7B-Instruct-v0.1", | ||
"displayName": "mistralai/Mistral-7B-Instruct-v0.1", | ||
"description": "Mistral 7B is a new Apache 2.0 model, released by Mistral AI that outperforms Llama2 13B in benchmarks.", | ||
"websiteUrl": "https://mistral.ai/news/announcing-mistral-7b/", | ||
"preprompt": "", | ||
"chatPromptTemplate" : "<s>{{#each messages}}{{#ifUser}}[INST] {{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}}{{content}} [/INST]{{/ifUser}}{{#ifAssistant}}{{content}}</s>{{/ifAssistant}}{{/each}}", | ||
"parameters": { | ||
"temperature": 0.1, | ||
"top_p": 0.95, | ||
"repetition_penalty": 1.2, | ||
"top_k": 50, | ||
"truncate": 1000, | ||
"max_new_tokens": 2048, | ||
"stop": ["</s>"] | ||
}, | ||
"promptExamples": [ | ||
{ | ||
"title": "Write an email from bullet list", | ||
"prompt": "As a restaurant owner, write a professional email to the supplier to get these products every week: \n\n- Wine (x10)\n- Eggs (x24)\n- Bread (x12)" | ||
}, { | ||
"title": "Code a snake game", | ||
"prompt": "Code a basic snake game in python, give explanations for each step." | ||
}, { | ||
"title": "Assist in a task", | ||
"prompt": "How do I make a delicious lemon cheesecake?" | ||
} | ||
] | ||
} | ||
]` | ||
|
||
OLD_MODELS=`[{"name":"bigcode/starcoder"}, {"name":"OpenAssistant/oasst-sft-6-llama-30b-xor"}, {"name":"HuggingFaceH4/zephyr-7b-alpha"}]` | ||
|
||
TASK_MODEL='mistralai/Mistral-7B-Instruct-v0.1' | ||
|
||
|
||
APP_BASE="/chat" | ||
PUBLIC_ORIGIN=https://huggingface.co | ||
PUBLIC_SHARE_PREFIX=https://hf.co/chat | ||
PUBLIC_ANNOUNCEMENT_BANNERS=`[]` | ||
|
||
PUBLIC_APP_NAME=HuggingChat | ||
PUBLIC_APP_ASSETS=huggingchat | ||
PUBLIC_APP_COLOR=yellow | ||
PUBLIC_APP_DESCRIPTION="Making the community's best AI chat models available to everyone." | ||
PUBLIC_APP_DATA_SHARING=1 | ||
PUBLIC_APP_DISCLAIMER=1 | ||
|
||
RATE_LIMIT=16 | ||
MESSAGES_BEFORE_LOGIN=5# how many messages a user can send in a conversation before having to login. set to 0 to force login right away | ||
|
||
PUBLIC_GOOGLE_ANALYTICS_ID=G-8Q63TH4CSL | ||
|
||
# Not part of the .env but set as other variables in the space | ||
# ADDRESS_HEADER=X-Forwarded-For | ||
# XFF_DEPTH=2 |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,27 @@ | ||
# Prompt templates | ||
|
||
These are the templates used to format the conversation history for different models used in HuggingChat. Set them in your `.env.local` [like so](https://github.com/huggingface/chat-ui#chatprompttemplate). | ||
|
||
## Llama 2 | ||
|
||
```env | ||
<s>[INST] <<SYS>>\n{{preprompt}}\n<</SYS>>\n\n{{#each messages}}{{#ifUser}}{{content}} [/INST] {{/ifUser}}{{#ifAssistant}}{{content}} </s><s>[INST] {{/ifAssistant}}{{/each}} | ||
``` | ||
|
||
## CodeLlama | ||
|
||
```env | ||
<s>[INST] <<SYS>>\n{{preprompt}}\n<</SYS>>\n\n{{#each messages}}{{#ifUser}}{{content}} [/INST] {{/ifUser}}{{#ifAssistant}}{{content}} </s><s>[INST] {{/ifAssistant}}{{/each}} | ||
``` | ||
|
||
## Falcon | ||
|
||
```env | ||
System: {{preprompt}}\nUser:{{#each messages}}{{#ifUser}}{{content}}\nFalcon:{{/ifUser}}{{#ifAssistant}}{{content}}\nUser:{{/ifAssistant}}{{/each}} | ||
``` | ||
|
||
## Mistral | ||
|
||
```env | ||
<s>{{#each messages}}{{#ifUser}}[INST] {{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}} {{content}} [/INST]{{/ifUser}}{{#ifAssistant}}{{content}}</s> {{/ifAssistant}}{{/each}} | ||
``` |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.