pghysels / STRUMPACK

Structured Matrix Package (LBNL)
http://portal.nersc.gov/project/sparse/strumpack/
Other
164 stars 38 forks source link

distributed memory version #31

Open puso1 opened 3 years ago

puso1 commented 3 years ago

Hi, my name is Mike Puso and I work with a number of finite element code developers at LLNL and are interested in you sparse direct solver in particular. I noticed in your latest paper that you mention you have a version of that but the results were from a shared memory version. I was wondering what the status and availability of distributed memory solver is. We would be very interested in collaborating with you and running it on our latest sierra platform. We currently use PWSMP (leased) from the IBM group of Anshul Gupta, Super LU, MUMPS and Pastix. The PWSMP solver is what we use the most due to its superior performance but we would prefer to have a very good open source direct solver.

pghysels commented 3 years ago

Hi Mike, The sparse direct solver does support distributed memory, relying on MPI and ScaLAPACK. Sierra is a GPU machine, much like summit, correct? We also have GPU support. For a single GPU, we use CUDA and cuBLAS/cuSOLVER. To run on multiple nodes, with multiple GPUs, we use SLATE (https://bitbucket.org/icl/slatehttps://bitbucket.org/icl/slate) as well as CUDA. Here is more info on the installation process: https://portal.nersc.gov/project/sparse/strumpack/master/installation.html

Apart from the sparse direct solver, we also have distributed memory preconditioners, based on approximate factorization. This uses rank-structured matrix approximations, like Hierarchically Semi-Separable, Hierarchically Off-Diagonal Low-Rank, Block Low Rank and Butterfly matrix decompositions. These preconditioners are also aimed at large sparse problems from for instance FEM codes. See for instance here: https://portal.nersc.gov/project/sparse/strumpack/master/HSS_Preconditioning.html

We'd be happy to work with you.