Open ivr900 opened 4 years ago
Hello, can I get the specific commit you are using?
I use public versions. Can I get access to development git?
@ivr900 I will send out an invite and an email.
@JorgeG94 latest GAMESS public release commit info:
# Commit ID
45b4b41d774952a53dc86ce851faae85e580e87e
# Version
v2020.1
Sarom, thank you for that. I shall try the development version through. But from my first preliminary sight after uploading, there is no differences in the LIBCCHEM part between the development and latest public versions. I may be wrong though. I am compiling the development version. Then we'll see.
Hi @saromleang ,
I recently started with GAMESS and landed in same issue with latest GAMESS release.
I'm using the below versions of dependancies.
spack load gcc@10.3.0
spack load gcc@10.3.0
spack load globalarrays@5.8.2 ^mpich +cuda +slurm
spack load boost@1.85.0 +math+thread %gcc@10
spack load hdf5@1.14.3 +cxx+fortran +hl ^mpich +cuda +slurm %gcc@10
spack load eigen@3.4.0 %gcc@10
spack load intel-oneapi-mkl@2024.0.0
I appreciate the help.
Hi @samcom12 the libcchem code has been effectively deprecated in favour of a coming soon upgrade to it. At the moment you'd be better off using the CPU only code or the OMP gpu offloading cappabilities. The libcchem gpu code had seen little to no upgrade in many years and had fallen behind in quality and portability.
Thanks @JorgeG94 for the clarification.
Today, I tried using OpenMP offload compiled code. Still I dont see any GPU cards being used during simulation.
Can you help me chosing a sample testcase using openmp offload effectively? Im using below SLURM script.
#!/bin/bash
#SBATCH --nodes=1
#SBATCH --ntasks-per-node=40 ## Max. Cores = 48
#SBATCH --cpus-per-task=1
#SBATCH -p gpu ## gpu/standard
#SBATCH --time=07:00:00
#SBATCH --exclusive
#SBATCH --gres=gpu:2
ulimit -s unlimited
ulimit -c unlimited
source /home/apps/spack/share/spack/setup-env.sh
spack load cuda@11.4 %gcc@10
spack load intel-oneapi-mkl /b2a4omb
spack load mpich +cuda +slurm
spack load gcc +nvptx
export CUDA_VISIBLE_DEVICES='0,1'
export LD_LIBRARY_PATH=/home/apps/spack/opt/spack/linux-centos7-cascadelake/gcc-10.3.0/mpich-4.2.1-d5fsc72unjuz3byuzzmkjqfm5qbmf4sv/lib:$LD_LIBRARY_PATH
ulimit -s unlimited
time /scratch/samir/gamess_offload/rungms-dev /home/samir/gamess/tests/libcchem/paper/paper-cocaine.inp 00 $SLURM_NTASKS $SLURM_NTASKS_PER_NODE
The install.info
is attached FYR.
install.info.txt
Hi there!
I want to bring to attentions of developers that something not quite right is happening in the LIBCCHEM part of GAMESS after 2018-09-30 R2 version.That is that performance in HF, CC and MP2 calculations using GAMESS binary built with LIBCCHEM dropped significantly. Moreover, I noticed that GPU is not actually used in HF at all when it was requested, when monitoring job via 'nvidia-smi' functionality. The jobs finish and give correct results but performance in terms of elapsed time is horrible when not a trivial in size system used, as only cpus were used in calculation. Only RI-MP2 part of GPU code seems to be working as expected.
I was buiding GAMESS-LIBCCHEM on my linux workstation (linux64, Ubuntu 18.04LTS) as below
My "Makefile" was
My "install.info" was
I tried to use the same versions of compilers, MPI, GA, Boost and other 3d party dependencies, except CUDA where I can't use -lcublas_device as it was deprecated in CUDA>=10, therefore used cuda/9.2, with version 2018-09-30-R3. The 2018-09-30-R3 version of GAMESS works as expected. But all 2019 and 2020 versions get something wrong in LIBCCHEM code for HF, CC and MP2 calculations, as GPU is not actually used where it ought to.
Kind regards, Ivan Rostov NCI Australia, Canberra