Nextomics / NextPolish

Fast and accurately polish the genome generated by long reads.
GNU General Public License v3.0
200 stars 27 forks source link

db split1 issue #89

Open kcl58759 opened 2 years ago

kcl58759 commented 2 years ago

Describe the bug Hi I am trying to run NextPolish on Hifi pacbio sequences. I keep running into an unspecified error with db split. My workflow is:

ls /scratch/kcl58759/Eco_pacbio_kendall/pb_css_474/cromwell-executions/pb_ccs/c7a3dc30-7f94-40de-ac16-2445f965bfad/call-export_fasta/execution/m64060_210804_174320.hifi_reads.fasta.gz > lgs.fofn

Make config file:

creat config file run.cfg

job_type = slurm job_prefix = nextPolish task = best rewrite = yes rerun = 3 parallel_jobs = 6 multithread_jobs = 5 genome = /scratch/kcl58759/Eco_pacbio_kendall/474.Primary.Hifi.asm/474.Primary.HiFi.asm.p_ctg.fa genome_size = auto workdir = ./01_rundir polish_options = -p {multithread_jobs}

[lgs_option] lgs_fofn = ./lgs.fofn lgs_options = -min_read_len 1k -max_depth 100 lgs_minimap2_options = -x map-ont

then I submit with

!/bin/bash

SBATCH --job-name=NextPolish

SBATCH --partition=batch

SBATCH --ntasks=1

SBATCH --cpus-per-task=6

SBATCH --mem=90gb

SBATCH --time=99:00:00

SBATCH --output=nextpolish.out

SBATCH --error=nextpolish.err

SBATCH --mail-user=kcl58759@uga.edu

SBATCH --mail-type=END,FAIL

nextPolish run.cfg

Error message [60417 INFO] 2022-03-03 13:02:46 NextPolish start... [60417 INFO] 2022-03-03 13:02:46 version:v1.4.0 logfile:pid60417.log.info [60417 WARNING] 2022-03-03 13:02:46 Re-write workdir [60417 WARNING] 2022-03-03 13:03:05 Delete task: 1 due to missing sgs_fofn. [60417 WARNING] 2022-03-03 13:03:05 Delete task: 1 due to missing sgs_fofn. [60417 WARNING] 2022-03-03 13:03:05 Delete task: 2 due to missing sgs_fofn. [60417 WARNING] 2022-03-03 13:03:05 Delete task: 2 due to missing sgs_fofn. [60417 WARNING] 2022-03-03 13:03:05 Delete task: 6 due to missing hifi_fofn. [60417 WARNING] 2022-03-03 13:03:05 Delete task: 6 due to missing hifi_fofn. [60417 INFO] 2022-03-03 13:03:05 scheduled tasks: [5, 5] [60417 INFO] 2022-03-03 13:03:05 options: [60417 INFO] 2022-03-03 13:03:05 rerun: 3 rewrite: 1 kill: None cleantmp: 0 task: [5, 5] use_drmaa: 0 submit: None job_type: sge sgs_unpaired: 0 sgs_rm_nread: 1 parallel_jobs: 6 align_threads: 5 check_alive: None job_id_regex: None sgs_max_depth: 100 lgs_max_depth: 100 lgs_read_type: clr multithread_jobs: 5 lgs_max_read_len: 0 hifi_max_depth: 100 polish_options: -p 5 lgs_min_read_len: 1k hifi_max_read_len: 0 genome_size: 36224976 hifi_block_size: 500M hifi_min_read_len: 1k job_prefix: nextPolish sgs_block_size: 500000000 lgs_block_size: 500000000 sgs_use_duplicate_reads: 0 sgs_align_options: bwa mem hifi_minimap2_options: -x map-pb lgs_minimap2_options: -x map-pb -t 5 lgs_fofn: /scratch/kcl58759/Eco_pacbio_kendall/./lgs.fofn workdir: /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir snp_phase: /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/%02d.snp_phase snp_valid: /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/%02d.snp_valid lgs_polish: /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/%02d.lgs_polish kmer_count: /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/%02d.kmer_count hifi_polish: /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/%02d.hifi_polish score_chain: /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/%02d.score_chain genome: /scratch/kcl58759/Eco_pacbio_kendall/474.Primary.Hifi.asm/474.Primary.HiFi.asm.p_ctg.fa [60417 INFO] 2022-03-03 13:03:05 step 0 and task 5 start: [60417 INFO] 2022-03-03 13:03:19 Total jobs: 2 [60417 CRITICAL] 2022-03-03 13:03:19 Command 'qsub -pe smp 5 -l vf=2.5G -o /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.o -e /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.e /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh' returned non-zero exit status 1, error info: . Traceback (most recent call last): File "/apps/eb/NextPolish/1.4.0-GCCcore-8.3.0-Python-3.8.2/nextPolish", line 515, in main(args) File "/apps/eb/NextPolish/1.4.0-GCCcore-8.3.0-Python-3.8.2/nextPolish", line 369, in main task.run.start() File "/apps/eb/NextPolish/1.4.0-GCCcore-8.3.0-Python-3.8.2/lib/python3.8/site-packages/paralleltask/task_control.py", line 344, in start self._start() File "/apps/eb/NextPolish/1.4.0-GCCcore-8.3.0-Python-3.8.2/lib/python3.8/site-packages/paralleltask/task_control.py", line 398, in _start self.submit(job) File "/apps/eb/NextPolish/1.4.0-GCCcore-8.3.0-Python-3.8.2/lib/python3.8/site-packages/paralleltask/taskcontrol.py", line 252, in submit , stdout, _ = self.run(job.cmd) File "/apps/eb/NextPolish/1.4.0-GCCcore-8.3.0-Python-3.8.2/lib/python3.8/site-packages/paralleltask/task_control.py", line 288, in run log.critical("Command '%s' returned non-zero exit status %d, error info: %s." % (cmd, p.returncode, stderr)) File "/apps/eb/Python/3.8.2-GCCcore-8.3.0/lib/python3.8/logging/init.py", line 1481, in critical self._log(CRITICAL, msg, args, kwargs) File "/apps/eb/Python/3.8.2-GCCcore-8.3.0/lib/python3.8/logging/init.py", line 1577, in _log self.handle(record) File "/apps/eb/Python/3.8.2-GCCcore-8.3.0/lib/python3.8/logging/init.py", line 1587, in handle self.callHandlers(record) File "/apps/eb/Python/3.8.2-GCCcore-8.3.0/lib/python3.8/logging/init.py", line 1649, in callHandlers hdlr.handle(record) File "/apps/eb/Python/3.8.2-GCCcore-8.3.0/lib/python3.8/logging/init.py", line 950, in handle self.emit(record) File "/apps/eb/NextPolish/1.4.0-GCCcore-8.3.0-Python-3.8.2/lib/python3.8/site-packages/paralleltask/kit.py", line 42, in emit raise Exception(record.msg) Exception: Command 'qsub -pe smp 5 -l vf=2.5G -o /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.o -e /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.e /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh' returned non-zero exit status 1, error info: . Operating system** I am running on UGA's cluster system Sapelo2, I used slurm.

GCC What version of GCC are you using? You can use the command gcc -v to get it.

Python What version of Python are you using? You can use the command python --version to get it.

NextPolish NextPolish/1.4.0-GCCcore-8.3.0-Python-3.8.2

moold commented 2 years ago

Could you try to submit task qsub -pe smp 5 -l vf=2.5G -o /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.o -e /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.e /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh manually and see what happens?

kcl58759 commented 2 years ago

I am trying to rework it for slurm like this sbatch --mem=90 -pe smp 5 -l vf=2.5G -o /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.o -e /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.e /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh

However, it keeps saying cant open smp. Is this an option?

moold commented 2 years ago

If you are using slurm but the log show you are using job_type: sge SGE, so check what happened?

kcl58759 commented 2 years ago

Hmm, I am confused about that.

This is what happened when I tried the qsub command:

The command was: '/opt/apps/slurm/21.08.5/bin/sbatch -e /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.e -o /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.o /scratch/kcl58759/Eco_pacbio_kendall/Nextpolish_dir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh 2>&1' and the output was: 'sbatch: error: You must request some amount of memory. sbatch: error: Batch job submission failed: Job size specification needs to be provided '

kcl58759 commented 2 years ago

I believe I got this to work with: sbatch --partition=batch --ntasks=1 --cpus-per-task=5 --mem-per-cpu=537 --time=99:00:00 -o /scratch/kcl58759/Eco_pacbio_kendall/01_rundir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.o -e /scratch/kcl58759/Eco_pacbio_kendall/01_rundir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh.e /scratch/kcl58759/Eco_pacbio_kendall/01_rundir/00.lgs_polish/01.db_split.sh.work/db_split1/nextPolish.sh

However, I am unsure if there are further steps since I cant seem to find the output files.

moold commented 2 years ago

Hi, see here and ParallelTask to change the submit command template.