Replies: 1 comment
-
目前20B的推理性能没有测试。下面的链接是 lmdeploy 在 llama2-7b 上的benchmark。后续,我们会更新更多的模型和设备上的测试结果 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
想知道interLM尤其是20B的token生成速度,跟Qwen比如何
Beta Was this translation helpful? Give feedback.
All reactions