-
Notifications
You must be signed in to change notification settings - Fork 204
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How is Llava quantized ? #621
Comments
@Abhranta Hi, there is AutoGPTQ :
|
Does this quantize only the LLM or the ViT too ? |
Hi @sailfish009, is there no native support for LLava based models. The solution you suggested seems very hacky:( I was also wondering if the quantization happens to the vision encoder too? |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
In autoawq, do we only quantize the LLM part of Llava or do we also quantize the ViT ? Can we add support for quantizing the vision models like ViT or SIGLIP?
The text was updated successfully, but these errors were encountered: