Pub Date : 2024-10-24DOI: 10.1016/j.cpc.2024.109406
David J. Lusher , Andrea Sansica , Neil D. Sandham , Jianping Meng , Bálint Siklósi , Atsushi Hashimoto
OpenSBLI is an automatic code-generation framework for compressible Computational Fluid Dynamics (CFD) simulations on heterogeneous computing architectures (previous release: Lusher et al. (2021) [4]). OpenSBLI is coupled to the Oxford Parallel Structured (OPS) Domain Specific Language (DSL), which uses source-to-source translation to enable parallel execution of the code on large-scale supercomputers, including multi-GPU clusters. To date, OpenSBLI has largely been applied to compressible turbulence and shock-wave/boundary-layer interactions on very simple geometries comprised of single mesh blocks with essentially orthogonal grid lines. OpenSBLI has been extended in this new release to target strongly curvilinear cases, including transonic aerofoils using multi-block grids. In addition to multi-block mesh support, more efficient numerical shock-capturing methods and filters have been added to the codebase. Improvements to post-processing, reduced-dimension data output, and coupling to a modal decomposition library are also included. A set of validation cases are presented to showcase the new code features. Furthermore, state-of-the-art wide-span transonic aerofoil simulations on up to grid points demonstrate that wider aspect ratios can alter buffet predictions and increase the regularity of the low-frequency shock oscillations by accommodating fully-developed trailing edge flow separation. Spectral Proper Orthogonal Decomposition (SPOD) analysis showed that overly-narrow aerofoil simulations contain additional domain-dependent energy content at a Strouhal number of associated with wake modes.
{"title":"OpenSBLI v3.0: High-fidelity multi-block transonic aerofoil CFD simulations using domain specific languages on GPUs","authors":"David J. Lusher , Andrea Sansica , Neil D. Sandham , Jianping Meng , Bálint Siklósi , Atsushi Hashimoto","doi":"10.1016/j.cpc.2024.109406","DOIUrl":"10.1016/j.cpc.2024.109406","url":null,"abstract":"<div><div>OpenSBLI is an automatic code-generation framework for compressible Computational Fluid Dynamics (CFD) simulations on heterogeneous computing architectures (previous release: Lusher et al. (2021) <span><span>[4]</span></span>). OpenSBLI is coupled to the Oxford Parallel Structured (OPS) Domain Specific Language (DSL), which uses source-to-source translation to enable parallel execution of the code on large-scale supercomputers, including multi-GPU clusters. To date, OpenSBLI has largely been applied to compressible turbulence and shock-wave/boundary-layer interactions on very simple geometries comprised of single mesh blocks with essentially orthogonal grid lines. OpenSBLI has been extended in this new release to target strongly curvilinear cases, including transonic aerofoils using multi-block grids. In addition to multi-block mesh support, more efficient numerical shock-capturing methods and filters have been added to the codebase. Improvements to post-processing, reduced-dimension data output, and coupling to a modal decomposition library are also included. A set of validation cases are presented to showcase the new code features. Furthermore, state-of-the-art wide-span transonic aerofoil simulations on up to <span><math><mi>N</mi><mo>=</mo><mn>2.5</mn><mo>×</mo><msup><mrow><mn>10</mn></mrow><mrow><mn>9</mn></mrow></msup></math></span> grid points demonstrate that wider aspect ratios can alter buffet predictions and increase the regularity of the low-frequency shock oscillations by accommodating fully-developed trailing edge flow separation. Spectral Proper Orthogonal Decomposition (SPOD) analysis showed that overly-narrow aerofoil simulations contain additional domain-dependent energy content at a Strouhal number of <span><math><mi>S</mi><mi>t</mi><mo>≈</mo><mn>3</mn></math></span> associated with wake modes.</div></div>","PeriodicalId":285,"journal":{"name":"Computer Physics Communications","volume":"307 ","pages":"Article 109406"},"PeriodicalIF":7.2,"publicationDate":"2024-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142554022","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-24DOI: 10.1016/j.cpc.2024.109404
Clément Scherding , Georgios Rigas , Denis Sipp , Peter J. Schmid , Taraneh Sayadi
Many engineering applications rely on the evaluation of expensive, non-linear high-dimensional functions. In this paper, we propose the RONAALP algorithm (Reduced Order Nonlinear Approximation with Adaptive Learning Procedure) to incrementally learn, on-the-fly as the application progresses, a fast and accurate reduced-order surrogate model of a target function. First, a combination of nonlinear auto-encoder, community clustering, and radial basis function networks allows us to learn an efficient and compact surrogate model with limited training data. Secondly, an active learning procedure overcomes any extrapolation issues during the online stage by adapting the surrogate model with high-fidelity evaluations that fall outside its current validity range. This approach results in generalizable, fast, and accurate reduced-order models of high-dimensional functions. The method is demonstrated on three direct numerical simulations of hypersonic flows in chemical nonequilibrium. Accurate simulations of these flows rely on detailed thermochemical gas models that dramatically increase the cost of such calculations. Using RONAALP to learn a reduced-order thermodynamic model surrogate on-the-fly, the cost of such simulations was reduced by up to 75% while maintaining an error of less than 10% on relevant quantities of interest.
许多工程应用依赖于对昂贵的非线性高维函数进行评估。在本文中,我们提出了 RONAALP 算法(Reduced Order Nonlinear Approximation with Adaptive Learning Procedure,自适应学习程序的降阶非线性逼近算法),随着应用的进展,逐步学习目标函数的快速、准确的降阶代理模型。首先,结合非线性自动编码器、群集聚类和径向基函数网络,我们可以利用有限的训练数据学习高效、紧凑的代用模型。其次,主动学习程序克服了在线阶段的任何外推问题,通过对当前有效范围之外的高保真评估来调整代理模型。通过这种方法,可以建立可通用、快速、准确的高维函数降阶模型。该方法在化学非平衡态高超音速流动的三次直接数值模拟中得到了验证。对这些流动的精确模拟依赖于详细的热化学气体模型,这大大增加了计算成本。使用 RONAALP 实时学习减阶热力学代用模型,此类模拟的成本最多可降低 75%,而相关相关量的误差却保持在 10%以下。
{"title":"An adaptive learning strategy for surrogate modeling of high-dimensional functions - Application to unsteady hypersonic flows in chemical nonequilibrium","authors":"Clément Scherding , Georgios Rigas , Denis Sipp , Peter J. Schmid , Taraneh Sayadi","doi":"10.1016/j.cpc.2024.109404","DOIUrl":"10.1016/j.cpc.2024.109404","url":null,"abstract":"<div><div>Many engineering applications rely on the evaluation of expensive, non-linear high-dimensional functions. In this paper, we propose the RONAALP algorithm (<strong>R</strong>educed <strong>O</strong>rder <strong>N</strong>onlinear <strong>A</strong>pproximation with <strong>A</strong>daptive <strong>L</strong>earning <strong>P</strong>rocedure) to incrementally learn, on-the-fly as the application progresses, a fast and accurate reduced-order surrogate model of a target function. First, a combination of nonlinear auto-encoder, community clustering, and radial basis function networks allows us to learn an efficient and compact surrogate model with limited training data. Secondly, an active learning procedure overcomes any extrapolation issues during the online stage by adapting the surrogate model with high-fidelity evaluations that fall outside its current validity range. This approach results in generalizable, fast, and accurate reduced-order models of high-dimensional functions. The method is demonstrated on three direct numerical simulations of hypersonic flows in chemical nonequilibrium. Accurate simulations of these flows rely on detailed thermochemical gas models that dramatically increase the cost of such calculations. Using RONAALP to learn a reduced-order thermodynamic model surrogate on-the-fly, the cost of such simulations was reduced by up to 75% while maintaining an error of less than 10% on relevant quantities of interest.</div></div>","PeriodicalId":285,"journal":{"name":"Computer Physics Communications","volume":"307 ","pages":"Article 109404"},"PeriodicalIF":7.2,"publicationDate":"2024-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142561012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-23DOI: 10.1016/j.cpc.2024.109405
Inga S. Ulusoy , Lucas E. Aebersold , Cong Wang , Angela K. Wilson
The software MCEND (Multi-Configuration Electron-Nuclear Dynamics) is a free open-source program package which simulates the quantum dynamics of electron-nuclei simultaneously for diatomic molecules. Its formulation, implementation, and usage are described in detail. MCEND uses a grid-based basis representation for the nuclei, and the electronic basis is derived from standard electronic structure basis sets on the nuclear grid. The wave function is represented as a sum over products of electronic and nuclear wave functions, thus capturing correlation effects between electrons, nuclei, and electrons and nuclei. The LiH molecule was used as an example for simulating the molecular properties such as the dipole moment and absorption spectrum.
PROGRAM SUMMARY
Program Title MCEND, v.2.6.0
CPC Library link to program files:https://doi.org/10.17632/tkb9dwf85t.1
Developer's repository link:https://github.com/MCEND-hub/MCEND (https://github.com/MCEND-hub/MCEND-library and https://github.com/MCEND-hub/MCEND-tools are git submodules of MCEND)
Nature of problem: MCEND is to simulate the quantum dynamics of electrons and nuclei simultaneously at multiconfiguration levels.
Solution method: The presented program package solves the time-dependent Schrödinger equation with the wave function represented as sum over configuration products using an 8th-order adaptive step size Runge-Kutta ordinary differential equation (ODE) solver. The software can be extended by supplementing modules on the existing infrastructure.
MCEND(Multi-Configuration Electron-Nuclear Dynamics)软件是一个免费的开放源码程序包,可同时模拟二原子分子的电子核量子动力学。本文详细介绍了该软件的开发、实施和使用。MCEND 采用基于网格的原子核基础表示法,电子基础来自核网格上的标准电子结构基础集。波函数表示为电子波函数与核波函数乘积之和,从而捕捉电子、原子核以及电子与原子核之间的相关效应。以 LiH 分子为例,模拟了偶极矩和吸收光谱等分子特性。PROGRAM SUMMARYProgram Title MCEND, v.2.6.0CPC Library link to program files: https://doi.org/10.17632/tkb9dwf85t.1Developer's repository link: https://github.com/MCEND-hub/MCEND (https://github.com/MCEND-hub/MCEND-library and https://github.com/MCEND-hub/MCEND-tools are git submodules of MCEND)Licensing provisions:MIT 编程语言:Fortran 90 和 Python 3Fortran 90 和 Python 3外部例程/库:FFTW、OpenMP、BLAS、LAPACK、PSI4、Matplotlib、mendeleev、NumPy、Pandas、SciPy、PyTables问题性质:MCEND 是在多配置水平上同时模拟电子和原子核的量子动力学:所介绍的程序包使用 8 阶自适应步长 Runge-Kutta 常微分方程求解器求解与时间相关的薛定谔方程,波函数表示为配置乘积之和。该软件可通过在现有基础架构上补充模块进行扩展。
{"title":"MCEND: An open-source program for quantum electron-nuclear dynamics","authors":"Inga S. Ulusoy , Lucas E. Aebersold , Cong Wang , Angela K. Wilson","doi":"10.1016/j.cpc.2024.109405","DOIUrl":"10.1016/j.cpc.2024.109405","url":null,"abstract":"<div><div>The software MCEND (Multi-Configuration Electron-Nuclear Dynamics) is a free open-source program package which simulates the quantum dynamics of electron-nuclei simultaneously for diatomic molecules. Its formulation, implementation, and usage are described in detail. MCEND uses a grid-based basis representation for the nuclei, and the electronic basis is derived from standard electronic structure basis sets on the nuclear grid. The wave function is represented as a sum over products of electronic and nuclear wave functions, thus capturing correlation effects between electrons, nuclei, and electrons and nuclei. The LiH molecule was used as an example for simulating the molecular properties such as the dipole moment and absorption spectrum.</div><div><strong>PROGRAM SUMMARY</strong></div><div><em>Program Title</em> MCEND, v.2.6.0</div><div><em>CPC Library link to program files:</em> <span><span>https://doi.org/10.17632/tkb9dwf85t.1</span><svg><path></path></svg></span></div><div><em>Developer's repository link:</em> <span><span>https://github.com/MCEND-hub/MCEND</span><svg><path></path></svg></span> (<span><span>https://github.com/MCEND-hub/MCEND-library</span><svg><path></path></svg></span> and <span><span>https://github.com/MCEND-hub/MCEND-tools are git submodules of MCEND</span><svg><path></path></svg></span>)</div><div><em>Licensing provisions:</em> MIT</div><div><em>Programming language:</em> Fortran 90 and Python 3</div><div><em>External routines/libraries:</em> FFTW, OpenMP, BLAS, LAPACK, PSI4, Matplotlib, mendeleev, NumPy, Pandas, SciPy, PyTables</div><div><em>Nature of problem:</em> MCEND is to simulate the quantum dynamics of electrons and nuclei simultaneously at multiconfiguration levels.</div><div><em>Solution method:</em> The presented program package solves the time-dependent Schrödinger equation with the wave function represented as sum over configuration products using an 8th-order adaptive step size Runge-Kutta ordinary differential equation (ODE) solver. The software can be extended by supplementing modules on the existing infrastructure.</div></div>","PeriodicalId":285,"journal":{"name":"Computer Physics Communications","volume":"307 ","pages":"Article 109405"},"PeriodicalIF":7.2,"publicationDate":"2024-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142554027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-17DOI: 10.1016/j.cpc.2024.109403
José M. Rodríguez-Borbón , Xian Wang , Adrián P. Diéguez , Khaled Z. Ibrahim , Bryan M. Wong
We present an open-source software package, VAN-DAMME (Versatile Approaches to Numerically Design, Accelerate, and Manipulate Magnetic Excitations), for massively-parallelized quantum optimal control (QOC) calculations of multi-qubit systems. To enable large QOC calculations, the VAN-DAMME software package utilizes symmetry-based techniques with custom GPU-enhanced algorithms. This combined approach allows for the simultaneous computation of hundreds of matrix exponential propagators that efficiently leverage the intra-GPU parallelism found in high-performance GPUs. In addition, to maximize the computational efficiency of the VAN-DAMME code, we carried out several extensive tests on data layout, computational complexity, memory requirements, and performance. These extensive analyses allowed us to develop computationally efficient approaches for evaluating complex-valued matrix exponential propagators based on Padé approximants. To assess the computational performance of our GPU-accelerated VAN-DAMME code, we carried out QOC calculations of systems containing 10 - 15 qubits, which showed that our GPU implementation is 18.4× faster than the corresponding CPU implementation. Our GPU-accelerated enhancements allow efficient calculations of multi-qubit systems, which can be used for the efficient implementation of QOC applications across multiple domains.
Program summary
Program Title: VAN-DAMME
CPC Library link to program files::https://doi.org/10.17632/zcgw2n5bjf.1
Licensing provisions: GNU General Public License 3
Programming language: C++ and CUDA
Nature of problem: The VAN-DAMME software package utilizes GPU-accelerated routines and new algorithmic improvements to compute optimized time-dependent magnetic fields that can drive a system from a known initial qubit configuration to a specified target state with a large (≈1) transition probability.
{"title":"VAN-DAMME: GPU-accelerated and symmetry-assisted quantum optimal control of multi-qubit systems","authors":"José M. Rodríguez-Borbón , Xian Wang , Adrián P. Diéguez , Khaled Z. Ibrahim , Bryan M. Wong","doi":"10.1016/j.cpc.2024.109403","DOIUrl":"10.1016/j.cpc.2024.109403","url":null,"abstract":"<div><div>We present an open-source software package, VAN-DAMME (Versatile Approaches to Numerically Design, Accelerate, and Manipulate Magnetic Excitations), for massively-parallelized quantum optimal control (QOC) calculations of multi-qubit systems. To enable large QOC calculations, the VAN-DAMME software package utilizes symmetry-based techniques with custom GPU-enhanced algorithms. This combined approach allows for the simultaneous computation of hundreds of matrix exponential propagators that efficiently leverage the intra-GPU parallelism found in high-performance GPUs. In addition, to maximize the computational efficiency of the VAN-DAMME code, we carried out several extensive tests on data layout, computational complexity, memory requirements, and performance. These extensive analyses allowed us to develop computationally efficient approaches for evaluating complex-valued matrix exponential propagators based on Padé approximants. To assess the computational performance of our GPU-accelerated VAN-DAMME code, we carried out QOC calculations of systems containing 10 - 15 qubits, which showed that our GPU implementation is 18.4× faster than the corresponding CPU implementation. Our GPU-accelerated enhancements allow efficient calculations of multi-qubit systems, which can be used for the efficient implementation of QOC applications across multiple domains.</div></div><div><h3>Program summary</h3><div><em>Program Title:</em> VAN-DAMME</div><div><em>CPC Library link to program files::</em> <span><span>https://doi.org/10.17632/zcgw2n5bjf.1</span><svg><path></path></svg></span></div><div><em>Licensing provisions:</em> GNU General Public License 3</div><div><em>Programming language:</em> C++ and CUDA</div><div><em>Nature of problem:</em> The VAN-DAMME software package utilizes GPU-accelerated routines and new algorithmic improvements to compute optimized time-dependent magnetic fields that can drive a system from a known initial qubit configuration to a specified target state with a large (≈1) transition probability.</div><div><em>Solution method:</em> Quantum control, GPU acceleration, analytic gradients, matrix exponential, and gradient ascent optimization.</div></div>","PeriodicalId":285,"journal":{"name":"Computer Physics Communications","volume":"307 ","pages":"Article 109403"},"PeriodicalIF":7.2,"publicationDate":"2024-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142534983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-16DOI: 10.1016/j.cpc.2024.109398
Y. Dong , S.M. Valle , G. Battistoni , I. Mattei , C. Finck , V. Patera , A. Alexandrov , B. Alpat , G. Ambrosi , S. Argirò , M. Barbanera , N. Bartosik , M.G. Bisogni , V. Boccia , F. Cavanna , P. Cerello , E. Ciarrocchi , A. De Gregorio , G. De Lellis , A. Di Crescenzo , S. Muraro
The FOOT experiment of INFN is devoted to the measurement of the nuclear fragmentation double differential cross sections useful for the improvement of calculation models adopted in hadrontherapy and radioprotection. A detailed Monte Carlo simulation of the FOOT magnetic spectrometer has been implemented in order to optimize the design and to guide data analysis. This task has been accomplished by means of the FLUKA Monte Carlo code. The input files of the FLUKA simulations are created from the software framework of the experiment, in order to have a consistent generation and description of geometry and materials in both simulation and data analysis. In addition, this ensures the possibility of processing both simulated and real data with the same data analysis procedures. Databases containing specific parameters describing the setup employed in each different data taking campaign are used. A customized event-by-event output of the Monte Carlo code has been developed. It can be read out by the general software framework of FOOT, enabling access to the generation history of all particles in the same event. This output structure therefore gives the possibility to perform a detailed analysis and study of all relevant processes, allowing the detailed tracking reconstruction of all individual particles. Examples of results are presented.
{"title":"The FLUKA Monte Carlo simulation of the magnetic spectrometer of the FOOT experiment","authors":"Y. Dong , S.M. Valle , G. Battistoni , I. Mattei , C. Finck , V. Patera , A. Alexandrov , B. Alpat , G. Ambrosi , S. Argirò , M. Barbanera , N. Bartosik , M.G. Bisogni , V. Boccia , F. Cavanna , P. Cerello , E. Ciarrocchi , A. De Gregorio , G. De Lellis , A. Di Crescenzo , S. Muraro","doi":"10.1016/j.cpc.2024.109398","DOIUrl":"10.1016/j.cpc.2024.109398","url":null,"abstract":"<div><div>The FOOT experiment of INFN is devoted to the measurement of the nuclear fragmentation double differential cross sections useful for the improvement of calculation models adopted in hadrontherapy and radioprotection. A detailed Monte Carlo simulation of the FOOT magnetic spectrometer has been implemented in order to optimize the design and to guide data analysis. This task has been accomplished by means of the FLUKA Monte Carlo code. The input files of the FLUKA simulations are created from the software framework of the experiment, in order to have a consistent generation and description of geometry and materials in both simulation and data analysis. In addition, this ensures the possibility of processing both simulated and real data with the same data analysis procedures. Databases containing specific parameters describing the setup employed in each different data taking campaign are used. A customized event-by-event output of the Monte Carlo code has been developed. It can be read out by the general software framework of FOOT, enabling access to the generation history of all particles in the same event. This output structure therefore gives the possibility to perform a detailed analysis and study of all relevant processes, allowing the detailed tracking reconstruction of all individual particles. Examples of results are presented.</div></div>","PeriodicalId":285,"journal":{"name":"Computer Physics Communications","volume":"307 ","pages":"Article 109398"},"PeriodicalIF":7.2,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142535613","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-16DOI: 10.1016/j.cpc.2024.109401
Xinyang Dong , Emanuel Gull
The exploitation of space group symmetries in numerical calculations of periodic crystalline solids accelerates calculations and provides physical insight. We present results for a space-group symmetry adaptation of electronic structure calculations within the finite-temperature self-consistent GW method along with an efficient parallelization scheme on accelerators. Our implementation employs the simultaneous diagonalization of the Dirac characters of the orbital representation. Results show that symmetry adaptation in self-consistent many-body codes results in substantial improvements of the runtime, and that block diagonalization on top of a restriction to the irreducible wedge results in additional speedup.
{"title":"Symmetry adaptation for self-consistent many-body calculations","authors":"Xinyang Dong , Emanuel Gull","doi":"10.1016/j.cpc.2024.109401","DOIUrl":"10.1016/j.cpc.2024.109401","url":null,"abstract":"<div><div>The exploitation of space group symmetries in numerical calculations of periodic crystalline solids accelerates calculations and provides physical insight. We present results for a space-group symmetry adaptation of electronic structure calculations within the finite-temperature self-consistent GW method along with an efficient parallelization scheme on accelerators. Our implementation employs the simultaneous diagonalization of the Dirac characters of the orbital representation. Results show that symmetry adaptation in self-consistent many-body codes results in substantial improvements of the runtime, and that block diagonalization on top of a restriction to the irreducible wedge results in additional speedup.</div></div>","PeriodicalId":285,"journal":{"name":"Computer Physics Communications","volume":"307 ","pages":"Article 109401"},"PeriodicalIF":7.2,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142535614","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-15DOI: 10.1016/j.cpc.2024.109396
Guilherme Catumba, Alberto Ramos, Bryan Zaldivar
Monte Carlo methods represent a cornerstone of computer science. They allow sampling high dimensional distribution functions in an efficient way. In this paper we consider the extension of Automatic Differentiation (AD) techniques to Monte Carlo processes, addressing the problem of obtaining derivatives (and in general, the Taylor series) of expectation values. Borrowing ideas from the lattice field theory community, we examine two approaches. One is based on reweighting while the other represents an extension of the Hamiltonian approach typically used by the Hybrid Monte Carlo (HMC) and similar algorithms. We show that the Hamiltonian approach can be understood as a change of variables of the reweighting approach, resulting in much reduced variances of the coefficients of the Taylor series. This work opens the door to finding other variance reduction techniques for derivatives of expectation values.
{"title":"Stochastic automatic differentiation for Monte Carlo processes","authors":"Guilherme Catumba, Alberto Ramos, Bryan Zaldivar","doi":"10.1016/j.cpc.2024.109396","DOIUrl":"10.1016/j.cpc.2024.109396","url":null,"abstract":"<div><div>Monte Carlo methods represent a cornerstone of computer science. They allow sampling high dimensional distribution functions in an efficient way. In this paper we consider the extension of Automatic Differentiation (AD) techniques to Monte Carlo processes, addressing the problem of obtaining derivatives (and in general, the Taylor series) of expectation values. Borrowing ideas from the lattice field theory community, we examine two approaches. One is based on reweighting while the other represents an extension of the Hamiltonian approach typically used by the Hybrid Monte Carlo (HMC) and similar algorithms. We show that the Hamiltonian approach can be understood as a change of variables of the reweighting approach, resulting in much reduced variances of the coefficients of the Taylor series. This work opens the door to finding other variance reduction techniques for derivatives of expectation values.</div></div>","PeriodicalId":285,"journal":{"name":"Computer Physics Communications","volume":"307 ","pages":"Article 109396"},"PeriodicalIF":7.2,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142441928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-15DOI: 10.1016/j.cpc.2024.109395
B. Chargeishvili , L. Fekésházy , G. Somogyi , S. Van Thurenhout
We present LinApart, a routine designed for efficiently performing the univariate partial fraction decomposition of large symbolic expressions. Our method is based on an explicit closed formula for the decomposition of rational functions with fully factorized denominators. We provide an implementation in the Wolfram Mathematica language, which can lead to very significant performance gains over the built-in Apart command. Furthermore, a C language library implementing the core functionality and suitable for interfacing with other software is also provided. Both codes are made available at https://github.com/fekeshazy/LinApart.
{"title":"LinApart: Optimizing the univariate partial fraction decomposition","authors":"B. Chargeishvili , L. Fekésházy , G. Somogyi , S. Van Thurenhout","doi":"10.1016/j.cpc.2024.109395","DOIUrl":"10.1016/j.cpc.2024.109395","url":null,"abstract":"<div><div>We present <span>LinApart</span>, a routine designed for efficiently performing the univariate partial fraction decomposition of large symbolic expressions. Our method is based on an explicit closed formula for the decomposition of rational functions with fully factorized denominators. We provide an implementation in the <span>Wolfram Mathematica</span> language, which can lead to very significant performance gains over the built-in <span>Apart</span> command. Furthermore, a <span>C</span> language library implementing the core functionality and suitable for interfacing with other software is also provided. Both codes are made available at <span><span>https://github.com/fekeshazy/LinApart</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":285,"journal":{"name":"Computer Physics Communications","volume":"307 ","pages":"Article 109395"},"PeriodicalIF":7.2,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142535615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-15DOI: 10.1016/j.cpc.2024.109399
Laurits Tani , Nalong-Norman Seeba , Hardi Vanaveski , Joosep Pata , Torben Lange
Tau leptons serve as an important tool for studying the production of Higgs and electroweak bosons, both within and beyond the Standard Model of particle physics. Accurate reconstruction and identification of hadronically decaying tau leptons is a crucial task for current and future high energy physics experiments. Given the advances in jet tagging, we demonstrate how tau lepton reconstruction can be decomposed into tau identification, kinematic reconstruction, and decay mode classification in a multi-task machine learning setup. Based on an electron-positron collision dataset with full detector simulation and reconstruction, we show that common jet tagging architectures can be effectively used for these sub-tasks. We achieve comparable momentum resolutions of 2–3% with all the tested models, while the precision of reconstructing individual decay modes is between 80–95%. We find ParticleTransformer to be the best-performing approach, significantly outperforming the heuristic baseline. This paper also serves as an introduction to a new publicly available Fuτure dataset for the development of tau reconstruction algorithms. This allows to further study the resilience of ML models to domain shifts and the efficient use of foundation models for such tasks.
头轻子是研究希格斯玻色子和电弱玻色子产生的重要工具,无论是在粒子物理标准模型之内还是之外。精确重建和识别强子衰变的头轻子是当前和未来高能物理实验的关键任务。鉴于射流标签技术的进步,我们展示了如何在多任务机器学习设置中将头轻子重构分解为头识别、运动学重构和衰变模式分类。基于电子-正电子碰撞数据集的完整探测器模拟和重构,我们展示了常见的射流标记架构可以有效地用于这些子任务。所有测试模型的动量分辨率都在 2-3% 之间,而重建单个衰变模式的精度则在 80-95% 之间。我们发现粒子转换器(ParticleTransformer)是表现最好的方法,明显优于启发式基线。本文还介绍了用于开发 tau 重建算法的新的公开 Fuτure 数据集。这有助于进一步研究 ML 模型对领域变化的适应能力,以及在此类任务中对基础模型的有效使用。
{"title":"A unified machine learning approach for reconstructing hadronically decaying tau leptons","authors":"Laurits Tani , Nalong-Norman Seeba , Hardi Vanaveski , Joosep Pata , Torben Lange","doi":"10.1016/j.cpc.2024.109399","DOIUrl":"10.1016/j.cpc.2024.109399","url":null,"abstract":"<div><div>Tau leptons serve as an important tool for studying the production of Higgs and electroweak bosons, both within and beyond the Standard Model of particle physics. Accurate reconstruction and identification of hadronically decaying tau leptons is a crucial task for current and future high energy physics experiments. Given the advances in jet tagging, we demonstrate how tau lepton reconstruction can be decomposed into tau identification, kinematic reconstruction, and decay mode classification in a multi-task machine learning setup. Based on an electron-positron collision dataset with full detector simulation and reconstruction, we show that common jet tagging architectures can be effectively used for these sub-tasks. We achieve comparable momentum resolutions of 2–3% with all the tested models, while the precision of reconstructing individual decay modes is between 80–95%. We find ParticleTransformer to be the best-performing approach, significantly outperforming the heuristic baseline. This paper also serves as an introduction to a new publicly available <span>Fu</span><em>τ</em><span>ure</span> dataset for the development of tau reconstruction algorithms. This allows to further study the resilience of ML models to domain shifts and the efficient use of foundation models for such tasks.</div></div>","PeriodicalId":285,"journal":{"name":"Computer Physics Communications","volume":"307 ","pages":"Article 109399"},"PeriodicalIF":7.2,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142438238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-15DOI: 10.1016/j.cpc.2024.109391
S.J.P. Pamela , N. Carey , J. Brandstetter , R. Akers , L. Zanisi , J. Buchanan , V. Gopakumar , M. Hoelzl , G. Huijsmans , K. Pentland , T. James , G. Antonucci , JOREK Team
The fusion research facility ITER is currently being assembled to demonstrate that fusion can be used for industrial energy production, while several other programmes across the world are also moving forward, such as EU-DEMO, CFETR, SPARC and STEP. The high engineering complexity of a tokamak makes it an extremely challenging device to optimise, and test-based optimisation would be too slow and too costly. Instead, digital design and optimisation must be favoured, which requires strongly-coupled suites of multi-physics, multi-scale High-Performance Computing calculations. Safety regulation, uncertainty quantification, and optimisation of fusion digital twins is undoubtedly an Exascale grand challenge. In this context, having surrogate models to provide quick estimates with uncertainty quantification is essential to explore and optimise new design options. But there lies the dilemma: accurate surrogate training first requires simulation data. Extensive work has explored solver-in-the-loop solutions to maximise the training of such surrogates. Likewise, innovative methods have been proposed to accelerate conventional HPC solvers using surrogates. Here, a novel approach is designed to do both. By bootstrapping neural operators and HPC methods together, a self-improving framework is achieved. As more simulations are being run within the framework, the surrogate improves, while the HPC simulations get accelerated. This idea is demonstrated on fusion-relevant MHD simulations, where Fourier Neural Operator based surrogates are used to create neural coarse-solver for the Parareal (time-parallelisation) method. Parareal is particularly relevant for large HPC simulations where conventional spatial parallelisation has saturated, and the temporal dimension is thus parallelised as well. This Neural-Parareal framework is a step towards exploiting the convergence of HPC and AI, where scientists and engineers can benefit from automated, self-improving, ever faster simulations. All data/codes developed here are made available to the community.
{"title":"Neural-Parareal: Self-improving acceleration of fusion MHD simulations using time-parallelisation and neural operators","authors":"S.J.P. Pamela , N. Carey , J. Brandstetter , R. Akers , L. Zanisi , J. Buchanan , V. Gopakumar , M. Hoelzl , G. Huijsmans , K. Pentland , T. James , G. Antonucci , JOREK Team","doi":"10.1016/j.cpc.2024.109391","DOIUrl":"10.1016/j.cpc.2024.109391","url":null,"abstract":"<div><div>The fusion research facility ITER is currently being assembled to demonstrate that fusion can be used for industrial energy production, while several other programmes across the world are also moving forward, such as EU-DEMO, CFETR, SPARC and STEP. The high engineering complexity of a tokamak makes it an extremely challenging device to optimise, and test-based optimisation would be too slow and too costly. Instead, digital design and optimisation must be favoured, which requires strongly-coupled suites of multi-physics, multi-scale High-Performance Computing calculations. Safety regulation, uncertainty quantification, and optimisation of fusion digital twins is undoubtedly an Exascale grand challenge. In this context, having surrogate models to provide quick estimates with uncertainty quantification is essential to explore and optimise new design options. But there lies the dilemma: accurate surrogate training first requires simulation data. Extensive work has explored solver-in-the-loop solutions to maximise the training of such surrogates. Likewise, innovative methods have been proposed to accelerate conventional HPC solvers using surrogates. Here, a novel approach is designed to do both. By bootstrapping neural operators and HPC methods together, a self-improving framework is achieved. As more simulations are being run within the framework, the surrogate improves, while the HPC simulations get accelerated. This idea is demonstrated on fusion-relevant MHD simulations, where Fourier Neural Operator based surrogates are used to create neural coarse-solver for the Parareal (time-parallelisation) method. Parareal is particularly relevant for large HPC simulations where conventional spatial parallelisation has saturated, and the temporal dimension is thus parallelised as well. This Neural-Parareal framework is a step towards exploiting the convergence of HPC and AI, where scientists and engineers can benefit from automated, self-improving, ever faster simulations. All data/codes developed here are made available to the community.</div></div>","PeriodicalId":285,"journal":{"name":"Computer Physics Communications","volume":"307 ","pages":"Article 109391"},"PeriodicalIF":7.2,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142446884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}