Greetings,
I have seen several references to "memory leaks" and the thread "Memory “leaking” when writing lots of data - Stefan Scherfke" is most similar to my situation.
When using chunked storage memory usage grows linearly with read and write operations and is never released until the process is killed by the kernel.
When I run the same test with no chunking memory use (as measured) is static.
My system specs are :
os : Red Hat Linux Server 6.5
python : 2.7.6
h5py : 2.3.0
hdf5 : 1.8.12
My test case involves 100 files with dataset shape (250, 400, 300) and iterating through each slice of (x, y, :) reading from one file and writing to another.
Attached are plots of memory usage for :
- chunked storage;
- chunked with cache size set to 0 which appears not to have an effect;
- no chunking ie contiguous storage
For our application chunked storage is a requirement so I am hoping there is a solution, maybe in latest release?
I can post code but essentially it is just read and write done enough times to make memory usage obvious.
Any suggestions would be greatly appreciated.
Cheers
Stuart