Skip to content

Commit

Permalink
Merge pull request #5 from stackhpc/feat/ui-improvements
Browse files Browse the repository at this point in the history
Azimuth UI improvements
  • Loading branch information
sd109 authored Feb 1, 2024
2 parents 4ad7aa9 + cae797d commit 05cf9ba
Show file tree
Hide file tree
Showing 3 changed files with 20 additions and 11 deletions.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,7 @@ The following is a non-exhaustive list of models which have been tested with thi
- [Mistral 7B Instruct v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
<!-- - [AWQ Quantized Mixtral 8x7B Instruct v0.1](https://huggingface.co/TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ) (Not producing output properly) -->

Due to the combination of [components](##Components) used in this app, some HuggingFace models may not work as expected (usually due to the way in which LangChain formats the prompt messages). Any errors when using new model will appear in the pod logs for either the web-app deployment or the backend API deployment.
Due to the combination of [components](##Components) used in this app, some HuggingFace models may not work as expected (usually due to the way in which LangChain formats the prompt messages). Any errors when using new model will appear in the logs for either the web-app pod or the backend API pod. Please open an issue if you would like explicit support for a specific model which is not in the above list.


## Components
Expand Down
1 change: 1 addition & 0 deletions chart/azimuth-ui.schema.yaml
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
controls:
/huggingface/model:
type: TextControl
required: true
/huggingface/token:
type: TextControl
secret: true
Expand Down
28 changes: 18 additions & 10 deletions chart/values.schema.json
Original file line number Diff line number Diff line change
Expand Up @@ -8,12 +8,12 @@
"model": {
"type": "string",
"title": "Model",
"description": "The HuggingFace model to deploy (Hint: For a simple, lightweight demo try ise-uiuc/Magicoder-S-DS-6.7B)"
"description": "The [HuggingFace model](https://huggingface.co/models) to deploy (see [here](https://github.com/stackhpc/azimuth-llm?tab=readme-ov-file#tested-models) for a list of tested models)."
},
"token": {
"type": "string",
"title": "Access Token",
"description": "The HuggingFace access token to use for installing gated models.",
"description": "A HuggingFace [access token](https://huggingface.co/docs/hub/security-tokens). Only required for [gated models](https://huggingface.co/docs/hub/en/models-gated (e.g. Llama 2).",
"default": ""
}
},
Expand All @@ -28,42 +28,50 @@
"model_name": {
"type": "string",
"title": "Model Name",
"description": "Model name supplied to OpenAI client in frontend web app. Should match huggingface.model above."
"description": "Model name supplied to the OpenAI client in frontend web app. Should match huggingface.model above.",
"default": "mistralai/Mistral-7B-Instruct-v0.2"
},
"model_instruction": {
"type": "string",
"title": "Instruction",
"description": "The initial model prompt (i.e. the hidden instructions) to use when generating responses."
"description": "The initial model prompt (i.e. the hidden instructions) to use when generating responses.",
"default": "You are a helpful AI assistant. Please respond appropriately."
},
"page_title": {
"type": "string",
"title": "Page Title",
"description": "The title to use for the chat interface."
"description": "The title to use for the chat interface.",
"default": "Large Language Model"
},
"llm_max_tokens": {
"type": "number",
"title": "Max Tokens",
"description": "The maximum number of new [tokens](https://platform.openai.com/docs/api-reference/chat/create#chat-create-max_tokens) to generate for each LLM responses."
"description": "The maximum number of new [tokens](https://platform.openai.com/docs/api-reference/chat/create#chat-create-max_tokens) to generate for each LLM responses.",
"default": 1000
},
"llm_temperature": {
"type": "number",
"title": "LLM Temperature",
"description": "The '[temperature](https://platform.openai.com/docs/api-reference/chat/create#chat-create-temperature)' value to use when generating LLM responses."
"description": "The [temperature](https://platform.openai.com/docs/api-reference/chat/create#chat-create-temperature) value to use when generating LLM responses.",
"default": 1
},
"llm_top_p": {
"type": "number",
"title": "LLM Top P",
"description": "The [top p](https://platform.openai.com/docs/api-reference/chat/create#chat-create-top_p) value to use when generating LLM responses."
"description": "The [top p](https://platform.openai.com/docs/api-reference/chat/create#chat-create-top_p) value to use when generating LLM responses.",
"default": 1
},
"llm_presence_penalty": {
"type": "number",
"title": "LLM Presence Penalty",
"description": "The [presence penalty](https://platform.openai.com/docs/api-reference/chat/create#chat-create-presence_penalty) to use when generating LLM responses."
"description": "The [presence penalty](https://platform.openai.com/docs/api-reference/chat/create#chat-create-presence_penalty) to use when generating LLM responses.",
"default": 0
},
"llm_frequency_penalty": {
"type": "number",
"title": "LLM Frequency Penalty",
"description": "The [frequency_penalty](https://platform.openai.com/docs/api-reference/chat/create#chat-create-frequency_penalty) to use when generating LLM responses."
"description": "The [frequency_penalty](https://platform.openai.com/docs/api-reference/chat/create#chat-create-frequency_penalty) to use when generating LLM responses.",
"default": 0
}

},
Expand Down

0 comments on commit 05cf9ba

Please sign in to comment.