Hi, I have been trying to run zUMIs on a HPC (both through a sbatch job and on an interactive node), but did not succeed. It seems like it does not see my input files as the Smartseq3.zUMIs_YAMLerror.log says:
WARNING: ignoring environment value of R_HOME
$file1
NULL
Create the output and log directories if they don't exist
mkdir -p "$output_dir"
mkdir -p "$log_dir"
Run zUMIs using its own miniconda environment (-c)
and the prepared YAML file (input_dir)
$path_zUMIs/zUMIs.sh -c -y $input_dir
$path_zUMIs/zUMIs.sh -c -y $input_dir
I do not know what the problem is. My hypothesis is that maybe the micoconda environment makes it not see in the input files (that do exist and are not empty as verified by the du -sh command). But at the same time it seems to be able to detect a slight discrepancy in the STAR versions used for my index (which means it does not see it as NULL) vs the one used by zUMIs.
Can you help me, please?
I have also tried to make my own mamba (conda) environment to run zUMIs following the vignette https://github.com/sdparekh/zUMIs/wiki/Installation#dependencies but I have not been able to complete the last part of the dependencies installation: devtools::install_github('VPetukhov/ggrastr') (some issues with Cairo)... And Docker is not accepted by HPCs (for security reasons)... Is there anyway you could make it into a Singularity file, please? This would make it much easier to deploy and in particular into pipelines (Nextflow / Snakemake)...
@sdparekh I have noticed the YAMLerror.log has the same error even going back to a few years. Do you know how I may solve this issue, please? Thank you
Hi, I have been trying to run zUMIs on a HPC (both through a sbatch job and on an interactive node), but did not succeed. It seems like it does not see my input files as the Smartseq3.zUMIs_YAMLerror.log says:
WARNING: ignoring environment value of R_HOME $file1 NULL
$file2 NULL
$file3 NULL
$file4 NULL
[1] "" [1] "" [1] "" [1] "" [1] "" "" "" "" [1] "" [1] "" [1] "" [1] "" [1] "" "" "" "" [1] "NULL" "NULL" "NULL" "NULL" $file1 NULL
$file2 NULL
$file3 NULL
$file4 NULL
$file1 NULL
$file2 NULL
$file3 NULL
$file4 NULL
[1] 0
This is my YAML file:
project: Smartseq3 sequence_files: file1: name: /home/ubaruchel/smart-seq3/data/240814/exp1/1a_cutadapt/Undetermined_S0_L001_trim_R1.fastq.gz base_definition:
samtools_exec: samtools pigz_exec: pigz STAR_exec: STAR Rscript_exec: Rscript
I ran this command through a .sh file that is called through a sbatch script (SLURM):
!/bin/bash
Always add these two commands to your scripts when using a environment
eval "$(conda shell.bash hook)" source $CONDA_PREFIX/etc/profile.d/mamba.sh
Source the parameters file
source ./params_bioinfo_experiments/0_params.sh
Set variables
input_dir=$input_dir_2b output_dir=$output_dir_2b log_dir=$log_dir_2b
Create the output and log directories if they don't exist
mkdir -p "$output_dir" mkdir -p "$log_dir"
Run zUMIs using its own miniconda environment (-c)
and the prepared YAML file (input_dir)
$path_zUMIs/zUMIs.sh -c -y $input_dir
$path_zUMIs/zUMIs.sh -c -y $input_dir
I do not know what the problem is. My hypothesis is that maybe the micoconda environment makes it not see in the input files (that do exist and are not empty as verified by the du -sh command). But at the same time it seems to be able to detect a slight discrepancy in the STAR versions used for my index (which means it does not see it as NULL) vs the one used by zUMIs.
Can you help me, please?
I have also tried to make my own mamba (conda) environment to run zUMIs following the vignette https://github.com/sdparekh/zUMIs/wiki/Installation#dependencies but I have not been able to complete the last part of the dependencies installation: devtools::install_github('VPetukhov/ggrastr') (some issues with Cairo)... And Docker is not accepted by HPCs (for security reasons)... Is there anyway you could make it into a Singularity file, please? This would make it much easier to deploy and in particular into pipelines (Nextflow / Snakemake)...
Thank you very much,
Best wishes,
Ulysse