-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reduce memory usage! #50
Comments
|
I should note that there are extra memory leaks when trying current tensorflow version (that is no longer tensorflow-gpu and that is installed with pip rather than conda) |
Check for easy improvement when loading data, especially if we don't want to use the whole dataset. In particular inside the Plasim_Field object |
At the moment I create a separate dataset with |
Yeah, indeed. I was planning to do line by line evaluation of the code monitoring the RAM to see exactly when we load the data into memory and if we can do something about it, but I cannot guarantee I'll have time for that |
I was just doing some runs and I noticed that now with 1000 years of data training a CNN uses 1.3TB of virtual memory... Did you change anything? If yes is not going in the right direction 😆 |
Found the reason: before the network had ~200.000 trainable parameters... now it has ~8 millions... |
And that is because the MaxPool layers have disappeared |
Yes, sorry forgot to re-implement them. As modified the way |
Although virtual memory normally doesn't matter. |
I will re-implement the MaxPool for backward compatibility |
My view is that we will need to implement tensorflow datasets. Operations such as shuffling (balancing) and train/validation have to be done virtually by permuting only indices. The dataset has to somehow know which portions of the full dataset it has to provide for the |
There are not so many machines that can handle all the data in memory. We need to find a way to deal with this issue. There are many materials online which show how it can be achieved in tensorflow (pipelines etc). The problem is that I still get
out of memory
issue (the process simply gets killed).The text was updated successfully, but these errors were encountered: