Closed GabrielHoffman closed 2 years ago
HI @GabrielHoffman
Thanks for giving {zellkonverter} a go! Memory usage is something we haven't really looked into much yet so I'm not surprised there are some issues with big datasets. It would be useful to try and work out exactly which parts are using more memory. Possibly it's from somewhere we don't have a lot of control over but we will have to see.
I was able to run your example on my laptop with ~4-5 GB memory usage so potentially there are some system specific things as well.
I suspect is due to
Running gc()
after this line frees up memory. But for large datasets, my R session crashing for insufficient memory here.
Thats all I have time for right now.
Gabriel
Thanks. If this is the issue there's probably not a lot we can do. This conversion is handled by {reticulate} and I think there is always going to be an overhead moving between environments. We could see if forcing garbage collection helps but I generally prefer not to mess with that.
I have been using
zellkonverter::writeH5AD()
to convert text files from the Single Cell Portal to H5AD for downstream analysis. I import the single cell counts from text as asparseMatrix
, and I noticed thatwriteH5AD()
can use a huge amount of memory writing to disk. When I need to format large datasets, I need to use a high-memory machine since conversion can use more than 128 Gb.Is this on the user-side or the backend?
Reproducible example:
sessionInfo