-
Notifications
You must be signed in to change notification settings - Fork 8.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add support for Qwen MoE (Qwen2MoeForCausalLM) #6380
Labels
Comments
I tried
|
@maziyarpanahi (#6074) This PR has not been merged yet |
Thanks a lot! I missed the |
I would like to also leave this PR on the record since it's about MoE and Qwen models: #6453 |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Feature Description
Please provide a detailed written description of what you were trying to do, and what you expected
llama.cpp
to do as an enhancement.https://qwenlm.github.io/blog/qwen-moe/
Motivation
Please provide a detailed written description of reasons why this feature is necessary and how it is useful to
llama.cpp
users.Not sure if it's safe to use
--skip-unknown
here:Possible Implementation
If you have an idea as to how it can be implemented, please write a detailed description. Feel free to give links to external sources or share visuals that might be helpful to understand the details better.
Haven't looked at it yet, but could be more similar to Mixtral than DBRX
The text was updated successfully, but these errors were encountered: