-
Notifications
You must be signed in to change notification settings - Fork 56
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
您们会开源DeepSeekMoE 2B模型吗? #16
Comments
暂无计划开源小尺度的moe模型(2B),但我们会开源更大的145B MoE。 |
我需要能在安卓手机上快速运行,并且能在rtx3050上快速微调的小型号,用于个人化的助手实验。 |
这个场景建议试试 1-2B 的 dense 模型 |
I am looking for the 2B model as well. I need to deploy fast models in production on domain specific tasks and I don't need the reasoning capabilities of larger models. Right now I am running GPT-2-medium(440M) dense. Would be great to move to a MoE model that is the same speed, but higher parameter count. |
希望开源2B模型。目前社区小规模的MoE模型几乎没有,高校实验室几乎无法做任何关于MoE模型finetune的研究。DeepseekMoE-2B将解决这一需求。 |
您们会开源2b参数的moe模型吗?
The text was updated successfully, but these errors were encountered: