Open felixcremer opened 1 year ago
This is still using all available memory and is running into the EOFError even so we now have a VM with 256 GB of memory. This was the htop usage shortly before it crashed. It was running on full power and had an expected runtime of 31 minutes.
I just reran it with 8 threads and 8 workers and it seems to be using all processing power at least in bursts and in between the cpu usage drops down, I suspect, when the new chunks are loaded. The memory usage is slowly increasing in small steps and in the beginning it was around 60 GB and it is slowely moving up with small increases and in between some drops in the memory usage.
In this setup the memory usage seems to be capped at 145 Gigabyte. The analysis went through just now.
The switch to do the computation of the diagonal recurrence density seems to reduce the memory usaage by a lot, but it seems as if the cpus are not fully used and I am now IO bound for my analysis: These are the benchmarking results for the inner function for a single random pixel:
Running the analysis on a single tile with a max_cache of 5e8 leads to a memory usage of 90Gb roughly, with loopchunks of
15000,20
but the CPUs seem to be very much underused. Is there a possibility to speed up the inner loop by using Threads inside? This is the current implementation: