-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Chatbot v0.2 #82
base: main
Are you sure you want to change the base?
Chatbot v0.2 #82
Conversation
need hf token
Waiting for HF_token authorization
Still have the Token authorization problem |
If you need assistance, feel free to open an issue and write down what the "Token authorization problem" is. ;-) |
…me testing. From Docker's log it looks like it is running on the CPU and needs some testing.
From Docker's log it looks like running on the CPU and needs some testing. |
It will always auto change to the CPU and show me it don't have enough memory on docker as "Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s] |
currently got "Error during response generation: Input length of input_ids is 200, but `max_length` is set to 200. This can lead to unexpected behavior."
The chat works after the update but is still on the CPU, making it slow. I'll try utilizing the laptop's NVIDIA RTX 500 Ada Generation GPU to improve performance.
The chat works(possible) after the update but is still on the CPU, making it slow. I'll try utilizing the laptop's NVIDIA RTX 500 Ada Generation GPU to improve performance. |
The CPU is still computing... |
Now, it got a good answer as: "curl -X POST http://localhost:5002/api/chat -H "Content-Type: application/json" -d "{"query": "Tell me about bioimage analysis"}" C:\Users\huang>" The current problem is that it still works on my CPU, so it is not efficient. |
No description provided.