-
Notifications
You must be signed in to change notification settings - Fork 46
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
V2.2 support for Custom-Converted OpenVINO Models? #213
Comments
Does it have to be placed in this specific folder, unlike other models where I can set the file location in the AI Playground backend? In AI Playground, my other models are saved in different folders, and I haven't encountered any issues by changing the folder location in the settings. If it's mandatory to use this folder, I'll try it after work tonight. Thank you. |
It appears you are out of memory. The model may be too large to fit in available vram. Check openvino documentation on quantizing the model |
I tested the DeepSeek-R1-Distill-Qwen-7B model converted to int8 using OpenVINO again. This time there still seems to be plenty of VRAM remaining, but the same error occurred. |
Do you use safetensor to transform directly? I think openvino does not support the direct conversion of safetnsor to openvino format. Do you have files in deepseek-R1 format supported by openvino |
https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B |
In version V2.2, there appears to be support for OpenVINO. Following the instructions in the OpenVINO documentation, I attempted to convert the deepseek-ai/DeepSeek-R1-Distill-Qwen-14B model to OpenVINO format. However, neither using the parameter --weight-format fp16 nor keeping the default int8 configuration worked - the converted models cannot be used in the AI Playground.
To clarify, the issue of 'cannot be used' means that after specifying the model path, these models still do not appear in the model list. Only the default four models are shown.
I am a Chinese user utilizing DeepSeek translation. Please forgive any translation errors that may occur.
The text was updated successfully, but these errors were encountered: