We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
https://huggingface.co/docs/accelerate/main/en/usage_guides/big_modeling#designing-a-device-map
i try device_map=auto. but it seems not recongize the xla device tpu all goto cpu. please help and provide the minimum working example. @JackCaoG
The text was updated successfully, but these errors were encountered:
The feature sounds like host offloading. It's currently in experimental stage by @tengyifei .
host offloading
Sorry, something went wrong.
@zpcore @tengyifei could you give me more detailed example?how to use host offloading?
You should be able to find an example of this implementation here: https://github.com/AI-Hypercomputer/tpu-recipes/tree/main/training/trillium/Llama3-405B-PyTorch
On a related note, we are working on rolling out a much cleaner reference model.
No branches or pull requests
❓ Questions and Help
https://huggingface.co/docs/accelerate/main/en/usage_guides/big_modeling#designing-a-device-map
i try device_map=auto. but it seems not recongize the xla device tpu all goto cpu. please help and provide the minimum working example. @JackCaoG
The text was updated successfully, but these errors were encountered: