Andreas Adelmann’s research while affiliated with Paul Scherrer Institute and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (182)


Uncertainty quantification of spent nuclear fuel with multifidelity Monte Carlo
  • Article

February 2025

·

2 Reads

Annals of Nuclear Energy

Arnau Albà

·

Andreas Adelmann

·

Dimitri Rochman

GAMBAS -- Fast Beam Arrangement Selection for Proton Therapy using a Nearest Neighbour Model
  • Preprint
  • File available

August 2024

·

17 Reads

Renato Bellotti

·

·

Antony J. Lomax

·

[...]

·

Purpose: Beam angle selection is critical in proton therapy treatment planning, yet automated approaches remain underexplored. This study presents and evaluates GAMBAS, a novel, fast machine learning model for automatic beam angle selection. Methods: The model extracts a predefined set of anatomical features from a patient's CT and structure contours. Using these features, it identifies the most similar patient from a training database and suggests that patient's beam arrangement. A retrospective study with 19 patients was conducted, comparing this model's suggestions to human planners' choices and randomly selected beam arrangements from the training dataset. An expert treatment planner evaluated the plans on quality (scale 1-5), ranked them, and guessed the method used. Results: The number of acceptable (score 4 or 5) plans was comparable between human-chosen 17 (89%) and model-selected 16(84%) beam arrangements. The fully automatic treatment planning took between 4 - 7 min (mean 5 min). Conclusion: The model produces beam arrangements of comparable quality to those chosen by human planners, demonstrating its potential as a fast tool for quality assurance and patient selection, although it is not yet ready for clinical use.

Download

Figure 1: Dataflow diagram of JulianA.jl code from a domain expert's perspective. Squares represent whether JulianA.jl, the TPS interface or the domain expert (i. e. user) provide the functionality. Bookmark-shaped boxes indicate data structures. The function scientific task() is a placeholder for additional steps required to obtain a scientific result.
Figure 2: Schematic of pencil beam proton radiotherapy. The beam enters the patient from left to right and is subsequently scanned over the tumour (polygon) at discrete positions (orange dots) called spots.
Figure 3: Schematic of the treatment workflow for radiotherapy.
JulianA.jl -- A Julia package for radiotherapy

July 2024

·

63 Reads

The importance of computers is continually increasing in radiotherapy. Efficient algorithms, implementations and the ability to leverage advancements in computer science are crucial to improve cancer care even further and deliver the best treatment to each patient. Yet, the software landscape for radiotherapy is fragmented into proprietary systems that do not share a common interface. Further, the radiotherapy community does not have access to the vast possibilities offered by modern programming languages and their ecosystem of libraries yet. We present JulianA.jl, a novel Julia package for radiotherapy. It aims to provide a modular and flexible foundation for the development and efficient implementation of algorithms and workflows for radiotherapy researchers and clinicians. JulianA.jl can be interfaced with any scriptable treatment planning system, be it commercial, open source or in-house developed. This article highlights our design choices and showcases the package's simplicity and powerful automatic treatment planning capabilities.






Lasso Monte Carlo, a Variation on Multi Fidelity Methods for High Dimensional Uncertainty Quantification

September 2023

·

111 Reads

Uncertainty quantification (UQ) is an active area of research, and an essential technique used in all fields of science and engineering. The most common methods for UQ are Monte Carlo and surrogate-modelling. The former method is dimensionality independent but has slowconvergence, while the latter method has been shown to yield large computational speedups with respect to Monte Carlo. However, surrogate models suffer from the so-called curse of dimensionality, and become costly to train for high-dimensional problems, where UQ might become computationally prohibitive. In this paper we present a new technique, Lasso Monte Carlo (LMC), which combines a Lasso surrogate model with the multifidelity Monte Carlo technique, in order to perform UQ in high-dimensional settings, at a reduced computational cost. We provide mathematical guarantees for the unbiasedness of the method, and show that LMC can be more accurate than simple Monte Carlo. The theory is numerically tested with benchmarks on toy problems, aswell as on a real example ofUQfrom the field of nuclear engineering. In all presented examples LMC is more accurate than simple Monte Carlo and other multifidelity methods. Thanks to LMC, computational costs are reduced by more than a factor of 5 with respect to simple MC, in relevant cases.


Figure 2: Estimation of mean and standard deviation of the decay heat of the UO2 assembly after 2 years of cooling. Each line is an independent UQ calculation with increasing N, and there are 20 lines for each method.
Uncertainty Quantification on Spent Nuclear Fuel with LMC

September 2023

·

91 Reads

The recently developed method Lasso Monte Carlo (LMC) for uncertainty quantification is applied to the characterisation of spent nuclear fuel. The propagation of nuclear data uncertainties to the output of calculations is an often required procedure in nuclear computations. Commonly used methods such as Monte Carlo, linear error propagation, or surrogate modelling suffer from being computationally intensive, biased, or ill-suited for high-dimensional settings such as in the case of nuclear data. The LMC method combines multilevel Monte Carlo and machine learning to compute unbiased estimates of the uncertainty, at a lower computational cost than Monte Carlo, even in high-dimensional cases. Here LMC is applied to the calculations of decay heat, nuclide concentrations, and criticality of spent nuclear fuel placed in disposal canisters. The uncertainty quantification in this case is crucial to reduce the risks and costs of disposal of spent nuclear fuel. The results show that LMC is unbiased and has a higher accuracy than simple Monte Carlo.


Fast Uncertainty Quantification of Spent Nuclear Fuel with Neural Networks

August 2023

·

90 Reads

The accurate calculation and uncertainty quantification of the characteristics of spent nuclear fuel (SNF) play a crucial role in ensuring the safety, efficiency, and sustainability of nuclear energy production, waste management, and nuclear safeguards. State of the art physics-based models, while reliable, are computationally intensive and time-consuming. This paper presents a surrogate modeling approach using neural networks (NN) to predict a number of SNF characteristics with reduced computational costs compared to physics-based models. An NN is trained using data generated from CASMO5 lattice calculations. The trained NN accurately predicts decay heat and nuclide concentrations of SNF, as a function of key input parameters, such as enrichment, burnup, cooling time between cycles, mean boron concentration and fuel temperature. The model is validated against physics-based decay heat simulations and measurements of different uranium oxide fuel assemblies from two different pressurized water reactors. In addition, the NN is used to perform sensitivity analysis and uncertainty quantification. The results are in very good alignment to CASMO5, while the computational costs (taking into account the costs of generating training samples) are reduced by a factor of 10 or more. Our findings demonstrate the feasibility of using NNs as surrogate models for fast characterization of SNF, providing a promising avenue for improving computational efficiency in assessing nuclear fuel behavior and associated risks.


Citations (49)


... With this new architecture we will be able to make efficient use of Exascale-Architecture that will come online soon. The core algorithms of OPAL are already performance portable as demonstrated in [16]. ...

Reference:

Computational models for high-power cyclotrons and FFAs
Scaling and performance portability of the particle-in-cell scheme for plasma physics applications through mini-apps targeting exascale architectures
  • Citing Chapter
  • February 2024

... kruszywo drobne metalowe, baryt, hematyt i pył ołowiany, limonit, ilmenit, szkło oraz żużel stalowniczy [16][17][18][19][20][21][22][23]. Uwzględniając szereg wyspecyfikowanych wymagań projektowych oraz technologicznych, dążymy w kierunku uzyskania materiału konstrukcyjnego o podwyższonej trwałości dla basenów schładzania przechowywanych zużytych prętów paliwowych odseparowanych siatkami dystansującymi w zestawach (kasetach) paliwowych [24][25][26][27][28][29][30][31][32][33][34][35][36][37][38][39][40]. ...

Fast uncertainty quantification of spent nuclear fuel with neural networks
  • Citing Article
  • February 2024

Annals of Nuclear Energy

... This paper is a concise summary of the more detailed white-paper report of the workshop [25]. Subsequent to the workshop, an issue dedicated to the dynamics of high-power and high-energy cyclotrons was also published in International Committee for Future Accelerators (ICFA) Newsletter series [26][27][28][29][30][31][32][33][34][35][36][37][38][39][40]. ...

Computational models for high-power cyclotrons and FFAs

Journal of Instrumentation

... In this section, we briefly highlight some recent fault prediction research with application to particle accelerators (see also [12] for a more thorough literature review of time series forecasting methods in accelerators). Blokland et al [13] present a ML method that incorporates uncertainty aware predictions, employing the Siamese neural network model to predict upcoming errant beam pulses. ...

Time series forecasting methods and their applications to particle accelerators

Physical Review Accelerators and Beams

... IsoDAR@Yemilab is a proposed experiment aimed at resolving neutrino oscillation anomalies and searching for physics beyond Standard Model [122,123]. This can be achieved by producing neutrinos from the β-decay of 8 Li, produced by bombarding a 9 Be target with 60 MeV protons accelerated by a cyclotron. ...

IsoDAR@Yemilab: A report on the technology, capabilities, and deployment
  • Citing Article
  • September 2022

Journal of Instrumentation

... Accounting for the wiggler focusing and transverse space charge effects necessitated slight modifications to the matched Twiss functions and focusing strengths of the triplet's quadrupoles compared to a lattice in which drifts are used instead of the wigglers. This task was carried out with the particle tracking through one cell in OPAL-FEL [15]. As a result, small rms beam sizes in the cell and a beam crossover in the middle of the wiggler were obtained, see Fig. 8. ...

Benchmarking Collective Effects of Electron Interactions in a Wiggler with OPAL-FEL
  • Citing Article
  • July 2022

Computer Physics Communications

... GENE [5] and AMReX [6] would be examples for the first approach while the latter porting technique is used by ORB5 [7]. Our approach is related to the Cabana Toolkit [8] or the Alpine miniapps [9] for particle applications. Both create datastructures based on the performance portability framework Kokkos [2] to use multiple shared memory techniques. ...

ALPINE: A set of performance portable plasma physics particle-in-cell mini-apps for exascale computing

... Although efforts have been made to overcome this limitation for specific physical processes [12], the full optimization of particle detectors remains an open challenge. To date, only a handful of studies have addressed this problem [12][13][14][15][16][17]. ...

Input Beam Matching and Beam Dynamics Design Optimizations of the IsoDAR RFQ Using Statistical and Machine Learning Techniques

Frontiers in Physics

... Machine learning has become a powerful tool in high energy physics, where the computational cost of large-scale simulations is often prohibitively high. One promising approach involves the use of neural surrogate models, which serve as cheaper approximations to the exact theory and can significantly reduce computational cost [1]. In particular, gauge covariant/equivariant neural networks 1 are attracting considerable attention because they allow flexible and differentiable mappings between gauge fields, controlled by learnable parameters. ...

New directions for surrogate models and differentiable programming for High Energy Physics detector simulation

... Machine learning and neural networks have seen rapidly growing applications across various fields in recent years. Aerosol sensing based on light scattering and holography has been significantly enhanced by machine learning [43,45,50]. Machine learning has also been implemented to classify biological aerosols and atmospheric aerosols [51,52]. ...

Retrieval of aerosol properties from in situ, multi-angle light scattering measurements using invertible neural networks
  • Citing Article
  • March 2022

Journal of Aerosol Science