Open Yun-Ching-Chen opened 1 year ago
Thanks for using our tool!
The reason why ENIGNA trace norm need to perform SVD for each CSE in each round of gradient calculation. Therefore it takes time to optimise when apply ENIGMA trace norm on large datasets (>1000 samples or >5000 spots). Even though current version need to cost a long time on large dataset, it should not cost such long time (over 24 hrs) on ~500 samples. Here is my current thoughts to fix the issue:
verbose = TRUE
to check if the Kappa Score
is keeping decreasing. If not, and Kappa score is increasing. The algorithm is not converge. And I suggest to set a smaller gradient step tao_k
Kappa Score
is decreasing, then I suggest to set a relative bigger gradient step tao_k
, but keep in-mind, too big step size would lead algorithm not converge. Or, you could set a bigger max_ks
(e.g. 2-5), to relax the end condition.The parallelized ENIGMA is under development and I would upload soon. Hope above information is helpful, and please let me know if you suppose have any new questions.
Best, Ken
Hey
Have you fixed your question, if you still have problem, could share the data with me (omit some important information) and I could help you to address
Best, Ken
Hi,
I've tried to run Enigma trace norm for ~500 TCGA samples using my own scRNA data (15 cell types with ~ 10000 genes) as the reference (in the aggregated 10000x15 matrix but not the Seurat object). It has been running over 24 hrs. I wonder if there is any tip to speed up the calculation or if it is possible to make a multi-core version?
Thanks, YC