-
Notifications
You must be signed in to change notification settings - Fork 42
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Any plans on supporting Llama3.2 text and multimodal on Qualcomm AI 100? #152
Comments
BTW, your README link to models comming soon is broken (https://github.com/quic/efficient-transformers/blob/main/models-coming-soon) |
Hi Aless, The Llama 3.2 1B and 3B models work out of the box in the current repository, provided you use one of the latest product software releases. Could you share details on Qualcomm Cloud AI100 instances and the software SDK you are using? The changes for the Llama 3.2 text models (11B and 90B) are currently under review. If you would like to run these models, you can cherry-pick the changes #134 onto the mainline and proceed with the validation. Regarding the Llama 3.2 multimodal model, it is still under evaluation. I will keep you updated on any progress. Additionally, I will address the issue with the broken README link. |
Support for Llama3.2 Multi Modality is merged as part of #267 PR, Please check. |
Closing as this has been addressed. Please reopen in case of any issue. |
Do you plan on supporting Llama3.2 (text/multimodal) on Qualcomm A100?
I saw this post (https://www.qualcomm.com/news/onq/2024/09/qualcomm-partners-with-meta-to-support-llama-3-point-2-big-deal-for-on-device-ai) , but it seems to have compatibility only for Snapdragon chips.
The text was updated successfully, but these errors were encountered: