Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training with 2 GPUs is much slower than with 1 GPU #294

Open
ivanlado opened this issue Oct 2, 2024 · 0 comments
Open

Training with 2 GPUs is much slower than with 1 GPU #294

ivanlado opened this issue Oct 2, 2024 · 0 comments

Comments

@ivanlado
Copy link

ivanlado commented Oct 2, 2024

I compared the time it took to train the models using 2 GPUs vs. using 1 GPU, an the result was that training with 2 GPUs is much slower. In fact, training with 2 GPUs takes at least two times the amount of time compared to using a 1GPU. What is happening? What is wrong?

I have looked the messages displayed after every iteration, an although "data" time does not vary with respect to the single GPU case, the "time" time is at least twice bigger in the 2 GPUs case.

  • "data" time: The time it takes to load the data.
  • "time" time: The time it take to do a whole iteration, including loading the data, forward and backward props.
    Disclaimer: These confusing terms are the ones uses in the code.

The comparisons have been made using the same hardware configurations.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant