If you OOM once you OOM forever
See original GitHub issueAssume while using inductor that you have batch sizes b_1
and b_2
where b_1 < b2
and If you run b_1
first and if the model doesn’t OOM and then run b_2
then it OOMs.
The problem is if you run b_2
first and it OOMS and then if you run b_1
it does also OOM even though it’s not supposed to.
Potentially using dynamo.reset()
might help until the OOM issues are all fixed https://github.com/pytorch/pytorch/blob/master/torch/_dynamo/__init__.py#L31
Issue Analytics
- State:
- Created a year ago
- Comments:8 (8 by maintainers)
Top Results From Across the Web
Using Personal Meeting ID (PMI) - Zoom Support
Your Personal Meeting Room is a virtual meeting room permanently reserved for you that you can access with your Personal Meeting ID (PMI)......
Read more >The Psychology of Zoom Fatigue - The Atlantic
By now, you have no doubt heard of “Zoom fatigue,” the range of maladies, including exhaustion and headaches, that are associated with hours...
Read more >The most common Zoom problems and how to fix them now
Solution 3: Test the audio and video. If your webcam or audio issues persist, you can test your audio and video in Zoom...
Read more >HOW TO Keep the Same Meeting ID in Zoom! - YouTube
As with any task, setting up dozens of Zoom meetings every week for your company or clients can get exhausting, especially when you...
Read more >Never click on this kind of Zoom invite. You'll thank us forever
To avoid falling for this Zoom phishing scam, the BBB advises the following: always check to see that the message is coming from...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
I can reproduce this locally. If I print out
torch.cuda.memory_allocated()
during each iteration, this is what dynamo looks like:However, this is native PyTorch:
It seems dynamo doesn’t free some memory after each iteration, so the memory keeps growing.
SG, let me check if I can reproduce with this repro.