Closed sadikmu closed 3 years ago
Thanks Bas Tolhuis from from PacBio techsupport @PacificBiosciences
I got feedback to check 0-rawreads/build for stderr file and the following error message noticed
2020-04-13 11:31:50,050 - root - ERROR - Was waiting for '../../subreads.fasta.fofn'
Checking my files again I found that subreads.fasta.fofn file was missing.
After creating the subreads.fasta.fofn file containing absolute path to the input tutorial F1_bull_test.subreads.fasta.gz file job now submitted to slurm. Progress will be updated.
resubmitted job failed again.
I couldn’t see stderr file in … 0-rawreads/build path.
List of files in in …0-rawreads/build
lrwxrwxrwx 1 sadikm sadikm 108 Apr 20 23:27 pwatcher.dir -> /home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/jobs/Pac163ff68e5c9f
-rw-rw-r-- 1 sadikm sadikm 269 Apr 20 23:27 run.sh
-rw-rw-r-- 1 sadikm sadikm 610 Apr 20 23:27 task.sh
-rw-rw-r-- 1 sadikm sadikm 368 Apr 20 23:27 task.json
-rw-rw-r-- 1 sadik sadikm 296 Apr 20 23:27 template.sh
contents of fc.2467347.err
falcon-kit 1.8.1 (pip thinks "falcon-kit 1.8.1")
pypeflow 2.3.0
[INFO]Setup logging from file "None".
[INFO]$ lfs setstripe -c 12 /home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data >
[INFO]Apparently '/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data' is not in lustre filesystem, which is fine.
[INFO]fc_run started with configuration fc_run.cfg
[WARNING]Option contains flags with "_":
"falcon_sense_option=--output_multi --min_idt 0.70 --min_cov 4 --max_n_read 200 --n_core 24". Those should be "-", as in
"falcon_sense_option=--output-multi --min-idt 0.70 --min-cov 4 --max-n-read 200 --n-core 24". Auto-replaced.
[WARNING]Option contains flags with "_":
"overlap_filtering_setting=--max_diff 120 --max_cov 120 --min_cov 4 --n_core 24". Those should be "-", as in
"overlap_filtering_setting=--max-diff 120 --max-cov 120 --min-cov 4 --n-core 24". Auto-replaced.
[WARNING]Unexpected keys in input config: {'falcon_greedy'}
[WARNING]Please supply a default for "njobs" (aka concurrency) in section [job.defaults]. For now, we will use 8
[INFO]cfg=
{
"General": {
"LA4Falcon_preload": false,
"avoid_text_file_busy": true,
"bestn": 12,
"dazcon": false,
"falcon_greedy": "False",
"falcon_sense_greedy": false,
"falcon_sense_option": "--output-multi --min-idt 0.70 --min-cov 4 --max-n-read 200 --n-core 24",
"falcon_sense_skip_contained": false,
"fc_ovlp_to_graph_option": " --min-len 12000",
"genome_size": "20000000",
"input_fofn": "subreads.fasta.fofn",
"input_type": "raw",
"length_cutoff": "-1",
"length_cutoff_pr": "12000",
"overlap_filtering_setting": "--max-diff 120 --max-cov 120 --min-cov 4 --n-core 24",
"ovlp_DBdust_option": "",
"ovlp_DBsplit_option": "-s400",
"ovlp_HPCdaligner_option": "-v -B128 -M24",
"ovlp_daligner_option": "-k24 -h480 -e.95 -l1800 -s100",
"pa_DBdust_option": "",
"pa_DBsplit_option": "-x500 -s200",
"pa_HPCREPmask_option": "-k18 -h480 -w8 -e.8 -s100",
"pa_HPCTANmask_option": "-k18 -h480 -w8 -e.8 -s100",
"pa_HPCdaligner_option": "-v -B128 -M24",
"pa_REPmask_code": "0,300/0,300/0,300",
"pa_daligner_option": "-e0.76 -l1200 -k18 -h480 -w8 -s100",
"pa_dazcon_option": "-j 4 -x -l 500",
"pa_fasta_filter_option": "streamed-internal-median",
"pa_subsample_coverage": 0,
"pa_subsample_random_seed": 12345,
"pa_subsample_strategy": "random",
"seed_coverage": "40",
"skip_checks": false,
"target": "assembly"
},
"job.defaults": {
"JOB_QUEUE": "bigmem",
"MB": "30000",
"NPROC": "6",
"job_type": "slurm",
"njobs": 8,
"pwatcher_type": "fs_based",
"submit": "srun --wait=0 -p myqueue \\\n-J ${JOB_NAME} \\\n-o ${JOB_STDOUT} \\\n-e ${JOB_STDERR} \\\n--mem-per-cpu=${MB}M \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n\"${CMD}\"",
"use_tmpdir": false
},
"job.step.asm": {
"MB": "192000",
"NPROC": "24",
"njobs": "1"
},
"job.step.cns": {
"MB": "64000",
"NPROC": "8",
"njobs": "200"
},
"job.step.da": {
"MB": "32000",
"NPROC": "4",
"njobs": "300"
},
"job.step.dust": {},
"job.step.la": {
"MB": "64000",
"NPROC": "8",
"njobs": "200"
},
"job.step.pda": {
"MB": "64000",
"NPROC": "8",
"njobs": "200"
},
"job.step.pla": {
"MB": "32000",
"NPROC": "4",
"njobs": "300"
}
}
[INFO]In simple_pwatcher_bridge, pwatcher_impl=<module 'pwatcher.fs_based' from '/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py'>
[INFO]job_type='slurm', (default)job_defaults={'JOB_QUEUE': 'bigmem', 'MB': '30000', 'NPROC': '6', 'job_type': 'slurm', 'submit': 'srun --wait=0 -p myqueue \\\n-J ${JOB_NAME} \\\n-o ${JOB_STDOUT} \\\n-e ${JOB_STDERR} \\\n--mem-per-cpu=${MB}M \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n"${CMD}"', 'pwatcher_type': 'fs_based', 'njobs': 8, 'use_tmpdir': False}, use_tmpdir=False, squash=False, job_name_style=0
[INFO]Setting max_jobs to 8; was None
[INFO]Num unsatisfied: 2, graph: 2
[INFO]About to submit: Node(0-rawreads/build)
[INFO] starting job Job(jobid='Pac163ff68e5c9f', cmd='/bin/bash run.sh', rundir='/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/0-rawreads/build', options={'JOB_QUEUE': 'bigmem', 'MB': 4000, 'NPROC': 1, 'job_type': 'slurm', 'submit': 'srun --wait=0 -p myqueue \\\n-J ${JOB_NAME} \\\n-o ${JOB_STDOUT} \\\n-e ${JOB_STDERR} \\\n--mem-per-cpu=${MB}M \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n"${CMD}"', 'pwatcher_type': 'fs_based', 'njobs': 8, 'use_tmpdir': False}) w/ job_type=SLURM
[INFO]$ srun --wait=0 -p myqueue \
-J Pac163ff68e5c9f \
-o stdout \
-e stderr \
--mem-per-cpu=4000M \
--cpus-per-task=1 \
/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash
"/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash" >
[ERROR]In pwatcher.fs_based.cmd_run(), failed to submit background-job:
MetaJobSlurm(MetaJob(job=Job(jobid='Pac163ff68e5c9f', cmd='/bin/bash run.sh', rundir='/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/0-rawreads/build', options={'JOB_QUEUE': 'bigmem', 'MB': 4000, 'NPROC': 1, 'job_type': 'slurm', 'submit': 'srun --wait=0 -p myqueue \\\n-J ${JOB_NAME} \\\n-o ${JOB_STDOUT} \\\n-e ${JOB_STDERR} \\\n--mem-per-cpu=${MB}M \\\n--cpus-per-task=${NPROC} \\\n${JOB_SCRIPT}\n"${CMD}"', 'pwatcher_type': 'fs_based', 'njobs': 8, 'use_tmpdir': False}), lang_exe='/bin/bash'))
Traceback (most recent call last):
File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py", line 474, in cmd_run
state.submit_background(bjob)
File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py", line 121, in submit_background
bjob.submit(self, exe, script_fn) # Can raise
File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pwatcher/fs_based.py", line 325, in submit
self.submit_capture = capture(sge_cmd)
File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pypeflow/io.py", line 50, in capture
raise Exception(msg)
Exception: 126 <- 'srun --wait=0 -p myqueue \\\n-J Pac163ff68e5c9f \\\n-o stdout \\\n-e stderr \\\n--mem-per-cpu=4000M \\\n--cpus-per-task=1 \\\n/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash\n"/home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash"':
srun: error: cnode10: tasks 0-1: Exited with exit code 13
srun: Terminating job step 247347.0
/bin/sh: line 7: /home/toolsTest.data/FAlCON-Unzip/F1_bull_test_data/mypwatcher/wrappers/run-Pac163ff68e5c9f.bash: Permission denied
[ERROR]Failed to enqueue 1 of 1 jobs: {Node(0-rawreads/build)}
[INFO]CALL:
scancel -n Pac163ff68e5c9f
Traceback (most recent call last):
File "/home/miniconda3/envs/falcon/bin/fc_run", line 11, in <module>
load_entry_point('falcon-kit==1.8.1', 'console_scripts', 'fc_run')()
File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/mains/run1.py", line 706, in main
main1(argv[0], args.config, args.logger)
File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/mains/run1.py", line 73, in main1
input_fofn_fn=input_fofn_fn,
File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/mains/run1.py", line 235, in run
dist=Dist(NPROC=4, MB=4000, job_dict=config['job.step.da']),
File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/falcon_kit/pype.py", line 106, in gen_parallel_tasks
wf.refreshTargets()
File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pypeflow/simple_pwatcher_bridge.py", line 278, in refreshTargets
self._refreshTargets(updateFreq, exitOnFailure)
File "/home/miniconda3/envs/falcon/lib/python3.7/site-packages/pypeflow/simple_pwatcher_bridge.py", line 323, in _refreshTargets
raise Exception(msg)
Exception: Failed to enqueue 1 of 1 jobs: {Node(0-rawreads/build)}
Hi @sadikmu this appears to be a problem I'm running into as well. I believe it is due to the permissions of the .bash
scripts that are created in the mypwatcher/wrappers/
directory.
If you change the permissions of the file to be executable and then submit your run script again it will work for that particular step (pertaining to the bash script you just made into an executable). However, it seems we'll have to do this for each individual step & I'd like it to just run. Do any of the PacBio folks have any suggestions/workarounds?
hi @shannonekj, It would be unrealistic to rerun on slurm queue every time it fails modifying in the file permission to executable which should be autogenerated executable the job is submitted.
The F1_bull_test data (~462Mb) and the fc_run.cfg script here worked successfully on local server. However, running falcon on actual data on the local sever after modifying genome size in fc_run.cfg used above just failed.
I tried to used various parameter options within fc_run.cfg including parameters in #708 -unfortunately it did not work at all and throws the same error shown below. I have been struggling on this for a while and could not figure it out.
Unable to move forward on this and highly appreciate anyones suggestion.
[INFO]Setup logging from file "None".
[INFO]$ lfs setstripe -c 12 /home/assembly/falcon >
[INFO]Apparently '/home/assembly/falcon' is not in lustre filesystem, which is fine.
[INFO]fc_run started with configuration fc_run.cfg
[WARNING]Unexpected keys in input config: set(['ovlp_concurrent_jobs', 'falcon_greedy'])
[INFO]cfg=
{
"General": {
"LA4Falcon_preload": false,
"avoid_text_file_busy": true,
"bestn": 12,
"dazcon": false,
"falcon_greedy": "False",
"falcon_sense_greedy": false,
"falcon_sense_option": "--output-multi --min-idt 0.70 --min-cov 4 --max-n-read 200 --n-core 24",
"falcon_sense_skip_contained": false,
"fc_ovlp_to_graph_option": " --min-len 10000",
"genome_size": "570000000",
"input_fofn": "subreads.fasta.fofn",
"input_type": "raw",
"length_cutoff": "-1",
"length_cutoff_pr": "10000",
"overlap_filtering_setting": "--max-diff 100 --max-cov 100 --min-cov 4 --n-core 24",
"ovlp_DBdust_option": "",
"ovlp_DBsplit_option": "-s400",
"ovlp_HPCdaligner_option": "-v -B128 -M24 -k24 -h600 -e.95 -l1800 -s100",
"ovlp_concurrent_jobs": "288 -s50",
"ovlp_daligner_option": "-k24 -h480 -e.95 -l1800 -s100",
"pa_DBdust_option": "",
"pa_DBsplit_option": "-x500 -s50",
"pa_HPCREPmask_option": "-k18 -h480 -w8 -e.8 -s100",
"pa_HPCTANmask_option": "-k18 -h480 -w8 -e.8 -s100",
"pa_HPCdaligner_option": "-v -B128 -e0.75 -M24 -l1200 -k14 -h256 -w8 -s100 -t16",
"pa_REPmask_code": "0,300/0,300/0,300",
"pa_daligner_option": "-e0.75 -l1200 -k14 -h256 -w8 -s100",
"pa_dazcon_option": "-j 4 -x -l 500",
"pa_fasta_filter_option": "streamed-internal-median",
"pa_subsample_coverage": 0,
"pa_subsample_random_seed": 12345,
"pa_subsample_strategy": "random",
"seed_coverage": "40",
"skip_checks": false,
"target": "assembly"
},
"job.defaults": {
"JOB_QUEUE": "default",
"MB": "32768",
"NPROC": "6",
"job_type": "local",
"njobs": "32",
"pwatcher_type": "blocking",
"submit": "/bin/bash -c \"${JOB_SCRIPT}\" > \"${JOB_STDOUT}\" 2> \"${JOB_STDERR}\"",
"use_tmpdir": false
},
"job.step.asm": {
"MB": "192000",
"NPROC": "24",
"njobs": "1"
},
"job.step.cns": {
"MB": "64000",
"NPROC": "8",
"njobs": "200"
},
"job.step.da": {
"MB": "32000",
"NPROC": "4",
"njobs": "300"
},
"job.step.dust": {},
"job.step.la": {
"MB": "64000",
"NPROC": "8",
"njobs": "200"
},
"job.step.pda": {
"MB": "64000",
"NPROC": "8",
"njobs": "200"
},
"job.step.pla": {
"MB": "32000",
"NPROC": "4",
"njobs": "300"
}
}
[INFO]In simple_pwatcher_bridge, pwatcher_impl=<module 'pwatcher.blocking' from '/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/pwatcher/blocking.pyc'>
[INFO]job_type='local', (default)job_defaults={'JOB_QUEUE': 'default', 'pwatcher_type': 'blocking', 'use_tmpdir': False, 'MB': '32768', 'job_type': 'local', 'submit': '/bin/bash -c "${JOB_SCRIPT}" > "${JOB_STDOUT}" 2> "${JOB_STDERR}"', 'NPROC': '6', 'njobs': '32'}, use_tmpdir=False, squash=False, job_name_style=0
[INFO]Setting max_jobs to 32; was None
[INFO]Num unsatisfied: 2, graph: 2
[INFO]About to submit: Node(0-rawreads/build)
[INFO]Popen: '/bin/bash -c "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/pwatcher/mains/job_start.sh" > "/home/assembly/falcon/0-rawreads/build/run-P593808baa38b0f.bash.stdout" 2> "/home/assembly/falcon/0-rawreads/build/run-P593808baa38b0f.bash.stderr"'
[INFO](slept for another 0.0s -- another 1 loop iterations)
[INFO](slept for another 0.3s -- another 2 loop iterations)
[INFO](slept for another 1.2s -- another 3 loop iterations)
[INFO](slept for another 3.0s -- another 4 loop iterations)
[INFO](slept for another 6.0s -- another 5 loop iterations)
[INFO](slept for another 10.5s -- another 6 loop iterations)
[INFO](slept for another 16.8s -- another 7 loop iterations)
[INFO](slept for another 25.2s -- another 8 loop iterations)
[INFO](slept for another 36.0s -- another 9 loop iterations)
[INFO](slept for another 49.5s -- another 10 loop iterations)
[INFO](slept for another 66.0s -- another 11 loop iterations)
[INFO](slept for another 85.8s -- another 12 loop iterations)
[INFO](slept for another 109.2s -- another 13 loop iterations)
[INFO](slept for another 135.6s -- another 14 loop iterations)
[INFO](slept for another 150.0s -- another 15 loop iterations)
[ERROR]Task Node(0-rawreads/build) failed with exit-code=1
[ERROR]Some tasks are recently_done but not satisfied: set([Node(0-rawreads/build)])
[ERROR]ready: set([])
submitted: set([])
[ERROR]Noop. We cannot kill blocked threads. Hopefully, everything will die on SIGTERM.
Traceback (most recent call last):
File "/home/sadik.conda/envs/falcon/bin/fc_run", line 11, in <module>
load_entry_point('falcon-kit==1.3.0', 'console_scripts', 'fc_run')()
File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/falcon_kit/mains/run1.py", line 706, in main
main1(argv[0], args.config, args.logger)
File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/falcon_kit/mains/run1.py", line 73, in main1
input_fofn_fn=input_fofn_fn,
File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/falcon_kit/mains/run1.py", line 235, in run
dist=Dist(NPROC=4, MB=4000, job_dict=config['job.step.da']),
File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/falcon_kit/pype.py", line 109, in gen_parallel_tasks
wf.refreshTargets()
File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/pypeflow/simple_pwatcher_bridge.py", line 278, in refreshTargets
self._refreshTargets(updateFreq, exitOnFailure)
File "/home/sadik/.conda/envs/falcon/lib/python2.7/site-packages/pypeflow/simple_pwatcher_bridge.py", line 362, in _refreshTargets
raise Exception(msg)
Exception: Some tasks are recently_done but not satisfied: set([Node(0-rawreads/build)])
Apologise @pb-cdunn, Any suggestion please? I got stuck on this.
We've been concentrating on pbipa, which now works really well if you have HiFi data (i.e. ccs reads).
Try using an absolute path for
input_fofn = subreads.fasta.fofn
I think it's having trouble finding that.
Any feedback please?
I have been struggling for a while to setup Falcon (falcon-kit 1.8.1) on slum to get it run example F1_bull_test data before using Falcon configuration for my Pacbio data.
I have got the F1_bull_test data from here and sadly I have not succeeded it getting it successfully assembly the this F1_bull_test data.
Can you please help me?
Here is the detail below:
pb-assembly suite installed in a separate an environment called 'falcon'
conda create -n falcon source activate falcon conda install pb-assembly
Lists of packages within falcon environment
Contents of fc_run.cfg
contents of fc.246025.err
`
Thank you,