We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
I am using the following .env.local with llama-2-7b.Q4_K_S.gguf and llama prompt template
MODELS=`[ { "name": "llama-2-7b.Q4_K_S.gguf", "chatPromptTemplate": "<s>[INST] <<SYS>>\n{{preprompt}}\n<</SYS>>\n\n{{#each messages}}{{#ifUser}}{{content}} [/INST] {{/ifUser}}{{#ifAssistant}}{{content}} </s><s>[INST] {{/ifAssistant}}{{/each}}", "parameters": { "temperature": 0.1, "top_p": 0.95, "repetition_penalty": 1.2, "top_k": 50, "truncate": 1000, "max_new_tokens": 2048, "stop": ["</s>"] }, "endpoints": [ { "url": "http://127.0.0.1:8080", "type": "llamacpp" } ] } ]`
I am trying to get this work with chat-ui and it doesn't work and chat-ui is frozen. However server is receiving request from client.
The text was updated successfully, but these errors were encountered:
Quick question how did you start your llama.cpp server ? Did you specify -np 3 in the parameters ?
-np 3
Sorry, something went wrong.
@nsarrazin yes I have specified -np 2
likely resolved with my PR! #867 check out my branch and see if it helps ❤️
No branches or pull requests
I am using the following .env.local with llama-2-7b.Q4_K_S.gguf and llama prompt template
I am trying to get this work with chat-ui and it doesn't work and chat-ui is frozen. However server is receiving request from client.
The text was updated successfully, but these errors were encountered: