Speaker
Description
Exascale computing is close to becoming a reality. As technology progresses, it has become clear that heterogeneous computing is going to stay and adapting to new hardware is an ongoing challenge. Since 2015 PIConGPU has paved the way to accelerating plasma simulations across compute platforms using the Alpaka framework. This has enabled early adaption to new compute hardware and readiness for Exascale compute capabilities.
However, experience has shown that the real challenges are of a different nature. The first is in detailed analysis of the data produced in simulations. Here, we present our current work on I/O, code coupling, visual analytics and large-scale data analytics.
The second, and more pressing challenge, is comparison to experiment. Here, not only has the increasing quality of experiments put more demand on simulation quality, but more and more the damnd for fast, close to real time analysis has grown. This puts high quality simulations to the test, as runs on supercomputers tend to be costly. We present workflows to match experiment and simulations and a future look on how feedback loops between experiment and simulation can be optimized.
Acknowledgments
This work was partly funded by the Center of Advanced Systems Understanding (CASUS) which is financed by Germany’s Federal Ministry of Education and
Research (BMBF) and by the Saxon Ministry for Science, Culture and Tourism (SMWK) with tax funds on
the basis of the budget approved by the Saxon State Parliament.