Closed shengxinzhuan closed 2 years ago
Hi thanks for sharing this! Was there anything you felt was missing from the quick start? We recently updated this to include detailed parallelization instructions, showing how to run ccs with --chunk, so please take a look!
@MariaNattestad Hi, MariaNattestad! Thanks to reply! I know how to run ccs with --chunk. But in China, we always got the ccs.bam from the sequencing company. So I don't need to filter the ccs.bam again, and move --chunk in actc. By the way, the ccs with --chunk and without --chunk are spend the same time on my test. If we need to filter the ccs.bam by ourself, --chunk may be more recommanded cause it can run again from the breakpoint.
Thanks for this context. I want to make sure to mention that in the quick start we do recommend running ccs yourself with --all
so that DeepConsensus has the chance to rescue some reads that would have gotten filtered out as being below (usually) Q20 when ccs was run with default settings. Your yield above Q20 will often increase significantly if you are able to rerun ccs with --all
.
@MariaNattestad Thanks to reply! I will try again if the reads depth do not satisficated to assembly an genome.
@shengxinzhuan sounds good! I'll close this issue, but feel free to open a new one (so we get notified) if you have any new issues or questions. We want to understand the problems that our users face using DeepConsensus in practice!
@MariaNattestad ok, I will open a new one if i meet new problem. Thanks!
My computer hardware equipment look like this:
Install the requirement packages
Create an environment for deepconsensus using conda
Download the ACTC for reads mapping
Install the Deepconsensus[GPU] by using pip
Prepare all the needed input file for Deepconsensus
Get the ccs.bam
Get the subreads_to_ccs.bam
Tips
If you use the actc to map the subreads to ccs without chunks, then you may encounter this error when running the deepconsensus.
This error is caused by the number of stream processors reaching an upper limit as the iteration process increases. To avoid this mistake, the right way is chunking the data when using actc.
Chunking your subreads.bam
Get the model for Deepconsensus
Run the Deepconsensus
Merge the output