Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Out of memory with RTX3090 #19

Open
PyxAI opened this issue Jan 12, 2022 · 4 comments
Open

Out of memory with RTX3090 #19

PyxAI opened this issue Jan 12, 2022 · 4 comments

Comments

@PyxAI
Copy link

PyxAI commented Jan 12, 2022

Hi,
I'm trying to train gpt2xl, but keep getting OOM, even when I set batch size to 1 and gradient_accumulation to 8\16\512, contigous_gradients false and allgather_bucket_size \ reduce_bucket_size 2e2.
I can see in nvidia-smi that I'm only reaching half the memory capacity - around 12GB
My system is as stated - 3090 with 24GB memory
80 GB RAM
5600x cpu if that matters
running WSL2 on windows 10
Thanks.

@PyxAI
Copy link
Author

PyxAI commented Jan 14, 2022

So working with WSL is just a no-go
I installed dual boot ubuntu and now the problem disappeared

@BrandonKoerner
Copy link

dual boot only, huh... that sucks. I was really hoping I could use this on win10 or wsl(2)

@PyxAI
Copy link
Author

PyxAI commented May 17, 2022

I was however, able to run the model under WSL2 windows 11
Didn't check training, it's worth a shot @ReewassSquared

@uahmad235
Copy link

Hi @PyxAI . Which ubuntu version did you run this code on?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants