Closed olgabot closed 4 years ago
Doing autocommits right now because the test data (https://github.com/czbiohub/test-datasets/pull/8) is local to czbiohub machines so I can't test any of the pipeline locally, need to push to servers to test. Once the pipeline is semi-working, I will subset the 22GB bam to just the relevant data so I can test locally.
this was a bad direction.. killing this PR
Not doing the below anymore, moved to https://github.com/czbiohub/nf-predictorthologs/pull/41.
On the way to doing this, also needed to separate aligned/unaligned hashes and do sourmash search on only one of those. To have some semblance of sanity and not millions of commits per PR, I'm separating those into separate PRs.
~For differentially expressed hashes or provided hashes, if the
--csv
csv containsbam
column then filter the input bams for read ids of sequences containing hashes, then do featurecounts to figure out if that hash is in one of seven categories:~~1. Not aligned
~### Example
featureCounts.txt.summary
~~### Example
orthology_counts_mqc.txt
~PR checklist
nextflow run . -profile test,docker
).nf-core lint .
).docs
is updatedCHANGELOG.md
is updatedREADME.md
is updatedLearn more about contributing: CONTRIBUTING.md