Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Encountered GPU OOM when trainning. #141

Open
pigmeetsomebody opened this issue Jul 9, 2024 · 1 comment
Open

Encountered GPU OOM when trainning. #141

pigmeetsomebody opened this issue Jul 9, 2024 · 1 comment

Comments

@pigmeetsomebody
Copy link

Is training this mode consuming large memory?
When I finally run the train code, it is stopped due to GPU OOM.
image
The console output says this process has 23.38 GiB memory in use. Of the allocated memory 21.43 GiB is allocated by PyTorch, and 1.50 GiB is reserved by PyTorch but unallocated.
How to reduce the memory occupation when running the train code?

@alexLIUMinhao
Copy link

Was the batch size large during your training? This would likely consume a significant amount of GPU memory.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants