Sergei Bastrakov's research while affiliated with Helmholtz-Zentrum Dresden-Rossendorf and other places

Publications (10)

Article
Due to the recent announcement of the Frontier supercomputer, many scientific application developers are working to make their applications compatible with AMD (CPU-GPU) architectures, which means moving away from the traditional CPU and NVIDIA-GPU systems. Due to the current limitations of profiling tools for AMD GPUs, this shift leaves a void in...
Chapter
It is common in the HPC community that the achieved performance with just CPUs is limited for many computational cases. The EuroHPC pre-exascale and the coming exascale systems are mainly focused on accelerators, and some of the largest upcoming supercomputers such as LUMI and Frontier will be powered by AMD Instinct™ accelerators. However, these n...
Preprint
Full-text available
HPC systems employ a growing variety of compute accelerators with different architectures and from different vendors. Large scientific applications are required to run efficiently across these systems but need to retain a single code-base in order to not stifle development. Directive-based offloading programming models set out to provide the requir...
Preprint
Full-text available
Due to the recent announcement of the Frontier supercomputer, many scientific application developers are working to make their applications compatible with AMD architectures (CPU-GPU), which means moving away from the traditional CPU and NVIDIA-GPU systems. Due to the current limitations of profiling tools for AMD GPUs, this shift leaves a void in...
Article
Full-text available
In recent years, the prospects of performing fundamental and applied studies at the next-generation high-intensity laser facilities have greatly stimulated the interest in performing large-scale simulations of laser interaction with matter with the account for quantum electrodynamics (QED) processes such as emission of high energy photons and decay...
Preprint
Full-text available
In recent years, the prospects of performing fundamental and applied studies at the next-generation high-intensity laser facilities have greatly stimulated the interest in performing large-scale simulations of laser interaction with matter with the account for quantum electrodynamics (QED) processes such as emission of high energy photons and decay...
Technical Report
Full-text available
Three kernels, Current Deposition (also known as Compute Current), Particle Push (Move and Mark), and Shift Particles are known to be some of the most time-consuming kernels in PIConGPU. The Current Deposition kernel and Particle Push kernel both set up the particle attributes for running any physics simulation with PIConGPU, so it is crucial to im...
Technical Report
Full-text available
This is a technical report that summarizes findings on the analysis of PIConGPU's three most intensive kernels by using NVProf Profiler tool and Summit system at the Oak Ridge National Lab (ORML). The kernels, Current Deposition (also known as Compute Current), Particle Push (Move and Mark), and Shift Particles are known to be some of the biggest k...
Chapter
Full-text available
State-of-the-art numerical simulations of laser plasma by means of the Particle-in-Cell method are often extremely computationally intensive. Therefore there is a growing need for the development of approaches for the efficient utilization of resources of modern supercomputers. In this paper, we address the problem of a substantially non-uniform an...

Citations

... As a data-driven astronomical application, gridding should have portability on different GPU architectures. ROCm 56 is AMD's open-source software platform for GPU-accelerated high-performance computing and machine learning [1,15,18]. An efficient thread assignment scheme is designed for the cell update step on GPU, in Section 4.1.2, ...
... The growth of the nonlinear Compton rate at high intensity means that it becomes the most important limiting factor on the time step. Subcycling, i.e. dividing the global time step into smaller fractions and testing for emission at each stage, is an efficient means of scaling simulations of photon emission (Volokitin et al., 2020). This assumes that the EM field is approximately constant over a single time step, which is generic requirement for a simulation to be appropriately resolved. ...