Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

limit? #198

Open
cusiman opened this issue Feb 1, 2024 · 4 comments
Open

limit? #198

cusiman opened this issue Feb 1, 2024 · 4 comments

Comments

@cusiman
Copy link

cusiman commented Feb 1, 2024

The bot works but when the conversation passes 4000 tokens the bot becomes unstable in its responses I already changed these parameters: truncation_length chat_prompt_size, but still with the same problem

@innightwolfsleep
Copy link
Owner

Which model do you use?
Except truncation_length chat_prompt_size there is no context length parameters. Perhaps it is model problem.

@cusiman
Copy link
Author

cusiman commented Feb 2, 2024

I tried with wizard vicuna 30b and now with LLaMA2-13B-Psyfighter2 and I got the same problem, which uncensored model would be the ideal?

@innightwolfsleep
Copy link
Owner

innightwolfsleep commented Feb 2, 2024

Hard to say... As I know, default LLaMA2 support 4096 token length, but Llama2 forks support up to 16k tokens, so I dont know about LLaMA2-13B-Psyfighter2. I'll try to test it later, perhaps will find something.

@cusiman
Copy link
Author

cusiman commented Feb 2, 2024

Thanks for your response, what model do you recommend?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants