Closed rbutleriii closed 3 years ago
You are right. In the sctransform package we use future_lapply without the future.seed parameters. I've just changed this in the develop branch. However, I don't think that's related to the error you are seeing. For a more informative error message, you might want to process the elements of your sc.list individually without multiprocessing.
I was also out of memory, and did get it to run w/o errors by running it w/o future
loaded (and more memory). The future.seed warnings also popped up for IntegrateData
and FindClusters
, but no errors w/ sufficient memory.
Worth noting this is big-ish batch analysis 175k cells and four groups; SCT + rPCA + reference. This can be closed w/ warnings handled, but perhaps a more informative postmortem message for killed worker? Though when I exited the interactive session on the cluster it gave me a memory error, so it will be more obvious to batch users & local users.
Hi @rbutleriii, besides using the develop
branch of sctransform where @ChristophH's has addressed the warning issue, you could set conserve.memory=TRUE
to prevent creating the residual matrix (which is non-sparse) in full (would increase the overall run time). We'll try to address it by providing a more informative postmortem message in the future. Feel free to reopen with any follow-up issues.
SCtransform doesn't use future.seed=T in its lapply functions, resulting in an error and warnings. Cannot pass it in as an argument, goes to unused argument in vst().
From
warnings()
sessionInfo