-
Notifications
You must be signed in to change notification settings - Fork 353
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Out of vram and reboot #153
Comments
Hey @tdzz1102 and sorry for the late reply! Could you try with You might need to play with your value of |
@sanchit-gandhi I have solved this question by downgrading the nividia-driver and cuda version(but forgot what exactly the version was 😢). Now the server has been expired and I can't try this solution anylonger. The faster whisper has helped me a lot, and thank you anyway! |
Machine info
When I set up the environment and called
FlaxWhisperPipline('openai/whisper-xxx')
method to load the model, the server rebooted without any error. Only for 'openai/whisper-tiny' can it work correctly, and it will crush loading the 'openai/whisper-small' model and larger. I've triedXLA_PYTHON_CLIENT_PREALLOCATE=false
mentioned in issue 7 but it didn't work.The image below shows the vRAM usage of my machine. Missing data means machine rebooted.
Is there any way to prevent linux to reboot automatically when vRAM usage is high?
The text was updated successfully, but these errors were encountered: