The smooth particle hydrodynamics (SPH) technique is a purely Lagrangian method. SPH discretizes a fluid in a series of interpolation points (SPH particles) whose distribution follows the mass density of the fluid and their evolution relies on a weighted interpolation over close neighboring particles.
SPH simulations represent computationally demanding calculations. Therefore, trade-offs are made between temporal and spatial scales, resolution, dimensionality (3-D or 2-D), and approximated versions of the physics involved. The parallelization of SPH codes is not trivial due to their boundless nature and the absence of a structured particle grid. SPHYNX, ChaNGa, and SPH-flow are the three SPH codes selected in the PASC SPH-EXA project proposal. The performance of these codes is negatively impacted by factors, such as multiple time-stepping and gravity. Therefore, the goal is to extrapolate their common basic SPH features, which are consolidated in a fully optimized, Exascale-ready, MPI+X, pure-SPH, mini-app.
SPH-EXA mini-app is a C++17 headers-only code with no external software dependencies. The parallelism is currently expressed via the following models: MPI, OpenMP, CUDA and HIP.
Check our wiki for more details
SPH-EXA
├── README.md
├── docs
├── domain - cornerstone octree and domain
│ ├── include
│ │ └── cstone
│ │ ├── CMakeLists.txt
│ │ ├── cuda
│ │ ├── domain
│ │ ├── findneighbors.hpp
│ │ ├── halos
│ │ ├── primitives
│ │ ├── sfc
│ │ ├── tree
│ │ └── util
│ └── test - cornerstone unit- performance-
│ ├── integration_mpi and integration tests
│ ├── performance
│ ├── unit
│ └── unit_cuda
├── include - legacy octree implementation (superseded by cornerstone octree)
│ └─── sph - SPH kernel functions
│ ├── cuda
│ └─── kernel
├── scripts
├── src - test case main function
│ ├── evrard
│ ├── sedov
│ └── sqpatch
├── test
└── tools
Use the following commands to compile the Sedov blast wave example:
Minimal CMake configuration:
mkdir build
cd build
cmake <GIT_SOURCE_DIR>Recommended CMake configuration on Piz Daint:
module load daint-gpu
module load cudatoolkit
# Current production version is 3.14.5 which is too old
# module load CMake
# workaround
module load EasyBuild-custom/cscs
module use /apps/daint/UES/sandbox/sebkelle/easybuild/gpu/modules/all
module load CMake/3.20.0
mkdir build
cd build
cmake -DCMAKE_CXX_COMPILER=CC <GIT_SOURCE_DIR>- Build everything:
make -j10 - MPI + OpenMP:
make sedov - MPI + OpenMP + CUDA:
make sedov-cuda
The Sedov test case binaries are located in build/src/sedov/
Possible arguments:
-n NUM: Run the simulation with NUM^3 (NUM to the cube) number of particles-s NUM: Run the simulation with NUM of iterations (time-steps)-w NUM: Dump particle data every NUM iterations (time-steps)--quiet: Don't print any output to stdout
Example usage:
OMP_NUM_THREADS=4 ./src/sedov/sedov -n 100 -s 1000 -w 10Runs Sedov with 1 million particles for 1000 iterations (time-steps) with 4 OpenMP threads and dumps particles data every 10 iterationsOMP_NUM_THREADS=4 ./src/sedov/sedov-cuda -n 100 -s 1000 -w 10Runs Sedov with 1 million particles for 1000 iterations (time-steps) with 4 OpenMP threads. Uses the GPU for most of the compute work.OMP_NUM_THREADS=4 mpiexec -np 2 ./src/sedov/sedov -n 100 -s 1000 -w 10Runs Sedov with 1 million particles for 1000 iterations (time-steps) with 2 MPI ranks of 4 OpenMP threads each. Works when using MPICH. For OpenMPI, usempiruninstead.OMP_NUM_THREADS=12 srun -Cgpu -A<your account> -n<nnodes> -c12 ./src/sedov/sedov-cuda -n 100 -s 1000 -w 10Optimal runtime configuration on Piz Daint fornnodesGPU compute nodes. Launches 1 MPI rank with 12 OpenMP threads per node.
Currently, only the cornerstone octree and domain are fully unit tested:
./domain/test/unit/component_unitsGPU-enabled unit tests:
./domain/test/unit_cuda/component_units_cudaMPI-enabled integration and regression tests:
mpiexec -np 2 ./domain/test/integration_mpi/domain_2ranks
mpiexec -np 2 ./domain/test/integration_mpi/exchange_focus
mpiexec -np 2 ./domain/test/integration_mpi/exchange_halos
mpiexec -np 2 ./domain/test/integration_mpi/globaloctree
mpiexec -np 5 ./domain/test/integration_mpi/domain_nranks
mpiexec -np 5 ./domain/test/integration_mpi/exchange_domain
mpiexec -np 5 ./domain/test/integration_mpi/exchange_keys
mpiexec -np 5 ./domain/test/integration_mpi/focus_tree
mpiexec -np 5 ./domain/test/integration_mpi/treedomain- Ruben Cabezon**
- Aurelien Cavelan**
- Florina Ciorba**
- Michal Grabarczyk**
- Danilo Guerrera**
- David Imbert**
- Sebastian Keller**
- Lucio Mayer**
- Ali Mohammed**
- Jg Piccinali**
- Tom Quinn**
- Darren Reed**
This project is licensed under the MIT License - see the LICENSE file for details
- PASC SPH-EXA project