Skip to main content
The Message Passing Interface (MPI) is the de facto standard for distributed memory parallelism in high performance computing (HPC). MPI is the dominant programming model for modern day supercomputers and will continue to be critical in enabling researchers to scale up their HPC workloads to forthcoming pre-exascale and exascale systems within EuroHPC and elsewhere.

This workshop targets programmers in both academia and industry who already have experience with basic MPI and are ready to take the next step to more advanced usage. Topics which will be covered include communicators, groups, derived data types, one-sided communication, non-blocking collectives and hybrid MPI+threading approaches. Lectures will be interleaved with hands-on exercises. All exercises will be written in C, but the instructors will be able to answer questions about MPI in Fortran and Python.


• Familiarity with MPI in C/C++, Fortran or Python, either from introductory courses (e.g. PDC’s Introduction to MPI or the SNIC course An introduction to parallel programming using Message Passing with MPI which will be taught December 1-4) or through self-taught usage.
• Familiarity with C/C++
• Basic Linux command line skills
• Existing access to a SNIC cluster or own computer with MPI and compilers installed.


Registration and further information: