-
Notifications
You must be signed in to change notification settings - Fork 9
Unable to allocate enough memory #32
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Hi JC, Thank you for the detailed message and the traceback — this isn’t an error on your side. Neural-Admixture is indeed built to handle large-scale datasets, including biobank-level data. So your dataset — nearly 500,000 samples and 161k markers — is well within the expected range. That said, the memory issue you’re encountering stems from a known bug in the data loading pipeline. The good news: we’re releasing an update later this week that resolves this issue. The new version significantly reduces GPU memory usage during training, especially with large datasets like yours. I’ll follow up here as soon as the update is live! Best regards, |
Hello @joansaurina, That's a very good news! I will wait for the update and hope it will resolve our issue! Best, JC |
Hello @joansaurina, any updates about the new release? Thanks a lot again! |
It's ready it will come out any time soon. Joan |
Hey @jcgrenier — the new version Make sure to reinstall, and let us know how it goes. :) Joan |
Thanks for letting me know! Is there new requirements for that new version? Do we need another python version? Furthermore, when I try to install it from the git, I have multiple issues with some dependencies, but particularly with numpy, for which version 2.2.5 seems needed, but later on during the installation, some other dependency require a previous version. Is it normal?
Thanks for your help! |
Hey @jcgrenier, It's strange — we tested with Could you try again with a fresh Python 3.12 environment? Feel free to reach out at [[email protected]] to schedule a call if you're still having trouble. Joan |
Hello @joansaurina, It started with an error regarding numpy while trying to install it from the git :
So I changed the setup.cfg file so it could work with numpy>=1.21.0,<2.0.0. But then torch had also an issue while running the training :
So I extended the requirements so it can take torch 2.4.1 (because 2.4.0 was not available on the wheels on our system). Hope these changes won't create any issues thought. Thanks. |
Uh oh!
There was an error while loading. Please reload this page.
Hello! I've been trying to run neural-admixture in train mode on a big dataset containing almost 500,000 samples and 161k markers but I am not able to make it run in GPU mode. It looks like it tries to send everything in the GPU memory. Do you have any idea how to handle such cases?
Quick note, I've been able to generate the PCA with CPUs only, using more than 1.3Tb of RAM to do so.
I also tried reducing the batch size, but I'm still having the same issue :
Here's the trace :
neural-admixture train --num_cpus 12 --num_gpus 1 --k 2 --name neuralAdmixture --data_path dataset.bed --save_dir neural_admixture_gpus --pca_path neural_admixture_gpus/neuralAdmixture_pca.pt --batch_size 400
Thanks for your help!
JC
The text was updated successfully, but these errors were encountered: