Most HPC systems are clusters of shared memory nodes. To use such systems efficiently both memory consumption and communication time has to be optimized. Therefore, hybrid programming may combine the distributed memory parallelization on the node interconnect (e.g., with MPI) with the shared memory parallelization inside of each node (e.g., with OpenMP or MPI-3.0 shared memory). This course analyzes the strengths and weaknesses of several parallel programming models on clusters of SMP nodes. Multi-socket-multi-core systems in highly parallel environments are given special consideration. MPI-3.0 has introduced a new shared memory programming interface, which can be combined with inter-node MPI communication. It can be used for direct neighbor accesses similar to OpenMP or for direct halo copies, and enables new hybrid programming models. These models are compared with various hybrid MPI+OpenMP approaches and pure MPI. Numerous case studies and micro-benchmarks demonstrate the performance-related aspects of hybrid programming.
Hands-on sessions are included on all days. Tools for hybrid programming such as thread/process placement support and performance analysis are presented in a "how-to" section. This course provides scientific training in Computational Science and, in addition, the scientific exchange of the participants among themselves.
It is organized by the HLRS in cooperation with the VSC Research Center, TU Wien and NHR@FAU.
This hybrid event will take place online and at HLRS, University of Stuttgart Nobelstraße 19 70569 Stuttgart, Germany Room 0.439 / Rühle Saal Location and nearby accommodations
23. Jan 2024
25. Jan 2024
Hybrid Event - Stuttgart, Germany
Programming Languages for Scientific Computing
Zurück zur Liste
Learn more about course curricula and content levels.
Georg Hager (NHR@FAU, Uni. Erlangen), Claudia Blaas-Schenner (VSC Research Center, TU Wien), Rolf Rabenseifner (HLRS, Uni. Stuttgart)
1st day – 23 January 2024
08:45 Join online 09:00 Welcome 09:05 Motivation 09:15 Introduction 09:30 Programming Models 09:35 - MPI + OpenMP 10:00 Practical (how to compile and start) 10:30 Break 10:45 - continue: MPI + OpenMP 11:30 Break 11:45 - continue: MPI + OpenMP 12:30 Practical (how to do pinning) 13:00 Lunch 14:00 Practical (hybrid through OpenMP parallelization) 15:30 Q & A, Discussion 16:00 End of first day
2nd day – 24 January 2024
08:45 Join online 09:00 - Overlapping Communication and Computation 09:30 Practical (taskloops) 10:30 Break 10:45 - MPI + OpenMP Conclusions 11:00 - MPI + Accelerators 11:30 Tools 11:45 Break 12:00 Programming Models (continued) 12:05 - MPI + MPI-3.0 Shared Memory 13:00 Lunch 14:00 Practical (replicated data) 15:30 Q & A, Discussion 16:00 End of second day
3rd day – 25 January 2024
08:45 Join online 09:00 - MPI Memory Models and Synchronization 09:40 - Pure MPI 10:00 Break 10:15 - Recap - MPI Virtual Topologies 10:45 - Topology Optimization 11:15 Break 11:30 Practical/Demo (application aware Cartesian topology) 12:30 - Topology Optimization (Wrap up) 12:45 Conclusions 13:00 Lunch 14:00 Finish the hands-on labs, Discussion, Q & A, Feedback 16:00 End of third day (course)
A link to the course material (slides and exercises) will be available at course start
Register via the button at the top of this page.
Registration closes on Tuesday 26 Dec 2023.
Students without Master's degree or equivalent. Participants from EU or EuroCC countries only: 0 EUR PhD students or employees at a German university or public research institute: 0 EUR PhD students or employees at a university or public research institute in an EU or EuroCC country other than Germany: 0 EUR. Other participants, e.g., from industry, other public service providers, or government. Participants from EU or EuroCC countries only: 0 EUR
Our course fee includes coffee breaks (in classroom courses only).
For lists of EU and EuroCC countries have a look at the Horizon Europe and EuroCC website.
Only participants from these countries can take part in this course.
Maksym Deliyergiyev phone 0711 685 87261, maksym.deliyergiyev(at)hlrs.de
HLRS is part of the Gauss Centre for Supercomputing (GCS), together with JSC in Jülich and LRZ in Garching near Munich. EuroCC@GCS is the German National Competence Centre (NCC) for High-Performance Computing. HLRS is also a member of the Baden-Württemberg initiative bwHPC.
This course is provided within the framework of EuroCC2 and the bwHPC training program.
Besides the content of the training itself, another important aspect of this event is the scientific exchange among the participants. We try to facilitate such communication by
https://www.hlrs.de/training/2024/HY-HLRS or ... (at VSC)
See the training overview and the Supercomputing Academy pages.
Oktober 13, 2023
November 17, 2023
Online by JSC
Zurück zur Liste
Nobelstraße 19, 70569 Stuttgart, Deutschland
Ihr Weg zu uns
+49 711 685-87209
Als Mitglied des Gauss Centre for Supercomputing ist das HLRS eines der drei Bundeshöchstleistungsrechenzentren.
Das HLRS ist eine zentrale Einrichtung der Universität Stuttgart.