SPARC is an open-source software package for the accurate, effcient, and scalable solution of the Kohn-Sham density functional theory (DFT) problem. The main features of SPARC currently include
SPARC is straightforward to install, use, and modify, with minimal external library dependencies. It has shown to be an order of magnitude faster than state-of-the-art planewave codes, with a range of exchange-correlation functionals, and with increasing advantages as the number of processors is increased. In particular, SPARC efficiently scales to thousands of processors in regular operation, bringing solution times down to about a minute for systems with O(500-1000) atoms, and a few seconds for O(100-500) atoms. Using the O(N) SQ method, SPARC has been scaled to system sizes of over a million atoms (https://doi.org/10.1088/1361-651X/acdf06).
Prerequisite: C compiler, MPI.
There are several options to compile SPARC, depending on the available external libraries.
Option 1: Compile with BLAS and LAPACK.
Step 1: Install/Load OpenBLAS/BLAS and LAPACK.
Step 2: Change directory to src/
directory, there is an available makefile
.
Step 3 (optional): Edit makefile
. If the BLAS library path and LAPACK library path are not in the search path, edit the BLASROOT
and LAPACKROOT
variables, and add them to LDFLAGS
. If you are using BLAS instead of OpenBLAS, replace all -lopenblas
flags with -lblas
.
Step 4 (optional): To turn on DEBUG
mode, set DEBUG_MODE
to 1 in the makefile
.
Step 5: Within the src/
directory, compile the code by
$ make clean; make
Remark: make sure in the makefile USE_MKL = 0
, USE_SCALAPACK = 0
, and USE_DP_SUBEIG = 1
for option 1.
Option 2 (default): Compile with MKL.
Step 1: Install/Load MKL.
Step 2: Change directory to src/
directory, there is an available makefile
.
Step 3: Edit makefile
. Set USE_MKL
to 1 to enable compilation with MKL. If the MKL library path is not in the search path, edit the MKLROOT
variable to manually set the MKL path.
Step 4 (optional): For the projection/subspace rotation step, to use SPARC routines for matrix data distribution rather than ScaLAPACK (through MKL), set USE_DP_SUBEIG
to 1. We found on some machines this option is faster.
Step 5 (optional): To turn on DEBUG
mode, set DEBUG_MODE
to 1 in the makefile
.
Step 6: Within the src/
directory, compile the code by
$ make clean; make
Remark: make sure in the makefile USE_MKL = 1
and USE_SCALAPACK = 0
for option 2.
Option 3: Compile with BLAS, LAPACK, and ScaLAPACK.
Step 1: Install/Load OpenBLAS/BLAS, LAPACK, and ScaLAPACK.
Step 2: Change directory to src/
directory, there is an available makefile
.
Step 3 (optional): Edit makefile
. If the BLAS library path, LAPACK library path, and/or ScaLAPACK library path are not in the search path, edit the BLASROOT
, LAPACKROOT
, and/or SCALAPACKROOT
variables accordingly, and add them to LDFLAGS
. If you are using BLAS instead of OpenBLAS, replace all -lopenblas
flags with -lblas
.
Step 4 (optional): For the projection/subspace rotation step, to use SPARC routines for matrix data distribution rather than ScaLAPACK, set USE_DP_SUBEIG
to 1. We found on some machines this option is faster.
Step 5 (optional): To turn on DEBUG
mode, set DEBUG_MODE
to 1 in the makefile
.
Step 6: Within the src/
directory, compile the code by
$ make clean; make
Remark: make sure in the makefile USE_MKL = 0
and USE_SCALAPACK = 1
for option 3.
Once compilation is done, a binary named sparc
will be created in the lib/
directory.
sparc
binaries distributed by conda-forge
Pre-compiled sparc
package can be installed on x86_64 or aarch64 Linux platforms using anaconda
or miniconda
.
The binary is compiled with OpenBLAS and OpenMPI and flags USE_MKL=0 USE_SCALAPACK=1 USE_FFTW=1
.
sparc-env
)
conda create -n sparc-env
conda activate sparc-env
sparc-x
conda install -c conda-forge sparc-x
echo sparc binary is located at: $(which sparc)
echo .psp files installed at: $SPARC_PSP_PATH
echo SPARC doc files installed at: $SPARC_DOC_PATH
The required input files to run a simulation with SPARC are (with shared names)
(a) ".inpt" -- User options and parameters.
(b) ".ion" -- Atomic information.
It is required that the ".inpt" and ".ion" files are located in the same directory and share the same name. A detailed description of the input options is provided in the documentation located in doc/. Examples of input files can be found in the SPARC/tests
directory .
In addition, SPARC requires pseudopotential files of psp8 format which can be generated by D. R. Hamann's open-source pseudopotential code ONCVPSP. The SPMS table of soft and transferable pseudopotentials is provided within the package. For access to more pseudopotentials, the user is referred to pseudoDOJO ONCV potentials and the SG15 ONCV potentials. Note that using the ONCVPSP input files included in the SG15 ONCV potentials, one can easily convert the SG15 ONCV potentials from upf format to psp8 format. Paths to the pseudopotential files are specified in the ".ion" file.
SPARC can be executed in parallel using the mpirun
command. Sample PBS script files are available in "SPARC/tests" folder. It is required that the ".inpt" and ".ion" files are located in the same directory and share the same name. For example, to run a simulation with 8 processes with input files as "filename.inpt" and "filename.ion" in the root directory (SPARC/
), use the following command:
$ mpirun -np 24 ./lib/sparc -name filename
As an example, one can run one of the tests located in SPARC/tests/
. First go to SPARC/tests/Example_tests/
directory:
$ cd tests/Example_tests/
The input file is available inside the folder. Run a DC silicon system by
$ mpirun -np 24 ../../lib/sparc -name Si8_kpt
The result is printed to output file "Si8_kpt.out", located in the same directory as the input files. If the file "Si8_kpt.out" is already present, the result will be printed to "Si8_kpt.out_1" instead. The max number of ".out" files allowed with the same name is 100. Once this number is reached, the result will instead overwrite the "Si8_kpt.out" file. One can compare the result with the reference out file named "Si8_kpt.refout".
In the tests/
directory, we also provide a suite of tests which are arranged in a hierarchy of folders. Each test system has its own directory. A python script is also provided which launches the suite of test systems. To run a set of four quick tests locally on the CPU, simply run:
$ python SPARC_testing_script.py quick_run
The result is stored in the corresponding directory of the tests. A message is also printed in the terminal showing if the tests passed or failed. The tests can also be launched in parallel on a cluster by using the Python script. Detailed information on using the python script can be found in the 'ReadMe' file in the tests/
directory.
Upon successful execution of the sparc
code, depending on the calculations performed, some output files will be created in the same location as the input files.
".out" file
General information about the test, including input parameters, SCF convergence progress, ground state properties and timing information.
".static" file
Atomic positions and atomic forces if the user chooses to print these information.
".out" file
See above.
".geopt" file
Atomic positions and atomic forces for atomic relaxation, cell lengths and stress tensor for volume relaxation, and atomic positions, atomic forces, cell lengths , and stress tensor for full relaxation.
".restart" file
Information necessary to perform a restarted structural relaxation calculation. Only created if atomic relaxation is performed.
Ab initio molecular dynamics (AIMD) calculations
.out
file
See above.
.aimd
file
Atomic positions, atomic velocities, atomic forces, electronic temperature, ionic temperature and total energy for each QMD step.
.restart
file
Information necessary to perform a restarted QMD calculation.
for SPARC C/C++ developers, check the maintenance guide for details including:
If you publish work using/regarding SPARC, please cite some of the following articles, particularly those that are most relevant to your work: