Closed ShaelynXU closed 1 year ago
Hi, I think 60 GB may not be enough for such a (quite) big dataset. You may either process the samples in chunks (say from barcode 1 to barcode 6, then from barcode 7 to barcode 12 etc) by moving the fastq.gz files to different folders. Then, you can merge the feature tables afterwards, using some QIIME2 commands. As an alternative, you may downsample the number of reads for each sample. I would recommend this solution in case the samples are very unbalanced in terms of numbers of reads (see dataQC/demux_summary.qzv file for evaluating the number of reads for each sample). SM
Hello Simone,
Thank you for your suggestions. Yes my samples are not balanced and I would like to try with downsampling. Could you please tell which part should I modify in the conf or nf file?
Thank you, Shaelyn
Hi,
you should set a lower value to maxNumReads parameter (line 17 in the conf file) and leave downsampleFastq=true
(line 50 in the conf file). You may start trying with up to 5k reads per sample.
SM
Thank you Simone, I will try with your recommendation accordingly and keep you updated!
Hi, I am going to close the issue. If you have any further questions, or update on the results, feel free to reopen it or comment here! SM
Hello Simone,
Thank you for your help last time.
I am running 24 barcoded samples at one time, which includes 229,182 reads in my rep-seqs.qza file. I encountered issues in the "assignTaxonomy" step and assumed it is because of the RAM memory. After modifying the conf file (change it to 60G), the problem still occurs.
Would you think 60G would be sufficient to process this input file? Or do you have other solutions to solve this error?
Thank you! Shaelyn