You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The docs say one needs 80GB for a normal 7b model, however when I try to train the DeepSeek VL 7B on the research rig with an A100 I get an OOM. When trying to split across 4 GPUs (1 A100 and 3 4090s), it does not utilize the A100 and OOMs with the 3 4090s before training can start
The text was updated successfully, but these errors were encountered:
SinanAkkoyun
changed the title
VRAM requirement for full sft deepseek 7B
VRAM requirement for full sft deepseek VL 7B
May 1, 2024
Describe the bug
How much VRAM is needed to finetune the 7b VL model?
The docs say one needs 80GB for a normal 7b model, however when I try to train the DeepSeek VL 7B on the research rig with an A100 I get an OOM. When trying to split across 4 GPUs (1 A100 and 3 4090s), it does not utilize the A100 and OOMs with the 3 4090s before training can start
The text was updated successfully, but these errors were encountered: