Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEATURE_REQUEST] Using the llama.cpp server save/load slot api for group chats #2044

Open
sasha0552 opened this issue Apr 8, 2024 · 0 comments
Labels
🦄 Feature Request [ISSUE] Suggestion for new feature, update or change

Comments

@sasha0552
Copy link
Contributor

sasha0552 commented Apr 8, 2024

Have you searched for similar requests?

Yes

Is your feature request related to a problem? If so, please describe.

No response

Describe the solution you'd like

Using the new llama.cpp server api to save/load slots for group chats (or maybe even regular chats and/or message histories "Start new chat"?). This is very useful to not have to reprocess the whole context on every message from different character.

Describe alternatives you've considered

Using group chats with context re-processing with each message from different character.

Additional context

I can create a PR that implements this feature.

Related: ggerganov/llama.cpp#6341

Priority

Low (Nice-to-have)

Are you willing to test this on staging/unstable branch if this is implemented?

Yes

@sasha0552 sasha0552 added the 🦄 Feature Request [ISSUE] Suggestion for new feature, update or change label Apr 8, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
🦄 Feature Request [ISSUE] Suggestion for new feature, update or change
Projects
None yet
Development

No branches or pull requests

1 participant