Open luiztauffer opened 1 month ago
weirdly, this error stopped happening after I restarted the spark local cluster. But good to have it here for reference, in case it happens again
reopening because this error is happening consistently for the 4D dataset, both in my local machine and remote machines running with docker.
Spark keeps having issues of memory at that point in the code, we should probably improve that operation.
setting parallel_volume=False
seems avoid the problem... but this might be inefficient?
maybe related
a similar error happens at step 5 - clean_cells. Similarly, the error is avoided by setting parallel_clean=False
Should we consider changing the default values of parallel_volume
and parallel_clean
to False? @mikarubi
So, just to clarify -- this is an out-of-memory error, correct? In general, we expect people to start with a lot of RAM for these analyses, so I am inclined to keep these on (so that the jobs run faster without people needing to manually turn them on). Is it possible, at all, to catch this error and return a more meaningful error message to the user? That would probably be ideal.
the error is possibly due to an out of memory error allocated to the worker subprocesses. One possible solution would be to configure spark to increase this limit.
Ok, looking at this again.
parallels
to zero that would be probably be enough.
Short Java error:
Short Python error track:
some references: