diff --git a/.env.template b/.env.template index 6d1a887d1f..1b34abf063 100644 --- a/.env.template +++ b/.env.template @@ -7,6 +7,7 @@ MODELS=`[ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png", "websiteUrl" : "https://mistral.ai/news/mixtral-of-experts/", "modelUrl": "https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1", + "tokenizer": "mistralai/Mixtral-8x7B-Instruct-v0.1", "preprompt" : "", "chatPromptTemplate": " {{#each messages}}{{#ifUser}}[INST]{{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}} {{content}} [/INST]{{/ifUser}}{{#ifAssistant}} {{content}} {{/ifAssistant}}{{/each}}", "parameters" : { @@ -63,7 +64,6 @@ MODELS=`[ "description": "The latest and biggest model from Meta, fine-tuned for chat.", "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/meta-logo.png", "websiteUrl": "https://ai.meta.com/llama/", - "modelUrl": "https://huggingface.co/meta-llama/Llama-2-70b-chat-hf", "preprompt": " ", "chatPromptTemplate" : "[INST] <>\n{{preprompt}}\n<>\n\n{{#each messages}}{{#ifUser}}{{content}} [/INST] {{/ifUser}}{{#ifAssistant}}{{content}} [INST] {{/ifAssistant}}{{/each}}", "promptExamples": [ @@ -94,6 +94,7 @@ MODELS=`[ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/nous-logo.png", "websiteUrl" : "https://nousresearch.com/", "modelUrl": "https://huggingface.co/NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO", + "tokenizer": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO", "chatPromptTemplate" : "{{#if @root.preprompt}}<|im_start|>system\n{{@root.preprompt}}<|im_end|>\n{{/if}}{{#each messages}}{{#ifUser}}<|im_start|>user\n{{content}}<|im_end|>\n<|im_start|>assistant\n{{/ifUser}}{{#ifAssistant}}{{content}}<|im_end|>\n{{/ifAssistant}}{{/each}}", "promptExamples": [ { @@ -155,6 +156,7 @@ MODELS=`[ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png", "websiteUrl": "https://mistral.ai/news/announcing-mistral-7b/", "modelUrl": "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1", + "tokenizer": "mistralai/Mistral-7B-Instruct-v0.1", "preprompt": "", "chatPromptTemplate" : "{{#each messages}}{{#ifUser}}[INST] {{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}}{{content}} [/INST]{{/ifUser}}{{#ifAssistant}}{{content}}{{/ifAssistant}}{{/each}}", "parameters": { @@ -187,6 +189,7 @@ MODELS=`[ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png", "websiteUrl": "https://mistral.ai/news/announcing-mistral-7b/", "modelUrl": "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2", + "tokenizer": "mistralai/Mistral-7B-Instruct-v0.2", "preprompt": "", "chatPromptTemplate" : "{{#each messages}}{{#ifUser}}[INST] {{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}}{{content}} [/INST]{{/ifUser}}{{#ifAssistant}}{{content}}{{/ifAssistant}}{{/each}}", "parameters": { @@ -218,6 +221,7 @@ MODELS=`[ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/openchat-logo.png", "websiteUrl": "https://huggingface.co/openchat/openchat-3.5-0106", "modelUrl": "https://huggingface.co/openchat/openchat-3.5-0106", + "tokenizer": "openchat/openchat-3.5-0106", "preprompt": "", "chatPromptTemplate" : "{{#each messages}}{{#ifUser}}GPT4 Correct User: {{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}}{{content}}<|end_of_turn|>GPT4 Correct Assistant:{{/ifUser}}{{#ifAssistant}}{{content}}<|end_of_turn|>{{/ifAssistant}}{{/each}}", "parameters": { diff --git a/src/lib/components/AssistantSettings.svelte b/src/lib/components/AssistantSettings.svelte index e0f22d98fb..bc5e872ada 100644 --- a/src/lib/components/AssistantSettings.svelte +++ b/src/lib/components/AssistantSettings.svelte @@ -12,6 +12,7 @@ import { useSettingsStore } from "$lib/stores/settings"; import { isHuggingChat } from "$lib/utils/isHuggingChat"; + import TokensCounter from "./TokensCounter.svelte"; type ActionData = { error: boolean; @@ -28,8 +29,10 @@ export let models: Model[] = []; let files: FileList | null = null; - const settings = useSettingsStore(); + let modelId = + assistant?.modelId ?? models.find((_model) => _model.id === $settings.activeModel)?.name; + let systemPrompt = assistant?.preprompt ?? ""; let compress: typeof readAndCompressImage | null = null; @@ -238,7 +241,11 @@