New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
V100推理internVL-1.5问题 #870
Comments
拉下最新代码
|
感谢回复,但是还是报错,麻烦您再指点下:
命令如下: CUDA_VISIBLE_DEVICES=6 swift infer --model_type internvl-chat-v1_5 --model_id_or_path /data/InternVL-Chat-V1-5-Int8/ --use_flash_attn false |
int8版本还没来得及做接入兼容 原版的应该没问题 你可以尝试修改本地模型的 |
Int8按您说的改了
然后不加载量化的,加载原版 CUDA_VISIBLE_DEVICES=1,2 swift infer --model_type internvl-chat-v1_5 --model_id_or_path /data/InternVL-Chat-V1-5/ --use_flash_attn false 还是会报错:
但是把原版 |
感谢回馈 我明天再修改下 |
int8模型已兼容 对于不支持flash attention的gpu,可以使用 |
您好,使用V100按您教程里的来做,最后还是会遇到这个问题,请问V100显卡能绕过这个问题吗?
The text was updated successfully, but these errors were encountered: