Pub Date : 2024-05-22DOI: 10.5194/gmd-17-4135-2024
Ross Mower, Ethan D. Gutmann, Glen E. Liston, Jessica Lundquist, Soren Rasmussen
Abstract. SnowModel, a spatially distributed snow-evolution modeling system, was parallelized using Coarray Fortran for high-performance computing architectures to allow high-resolution (1 m to hundreds of meters) simulations over large regional- to continental-scale domains. In the parallel algorithm, the model domain was split into smaller rectangular sub-domains that are distributed over multiple processor cores using one-dimensional decomposition. All the memory allocations from the original code were reduced to the size of the local sub-domains, allowing each core to perform fewer computations and requiring less memory for each process. Most of the subroutines in SnowModel were simple to parallelize; however, there were certain physical processes, including blowing snow redistribution and components within the solar radiation and wind models, that required non-trivial parallelization using halo-exchange patterns. To validate the parallel algorithm and assess parallel scaling characteristics, high-resolution (100 m grid) simulations were performed over several western United States domains and over the contiguous United States (CONUS) for a year. The CONUS scaling experiment had approximately 70 % parallel efficiency; runtime decreased by a factor of 1.9 running on 1800 cores relative to 648 cores (the minimum number of cores that could be used to run such a large domain because of memory and time limitations). CONUS 100 m simulations were performed for 21 years (2000–2021) using 46 238 and 28 260 grid cells in the x and y dimensions, respectively. Each year was simulated using 1800 cores and took approximately 5 h to run.
{"title":"Parallel SnowModel (v1.0): a parallel implementation of a distributed snow-evolution modeling system (SnowModel)","authors":"Ross Mower, Ethan D. Gutmann, Glen E. Liston, Jessica Lundquist, Soren Rasmussen","doi":"10.5194/gmd-17-4135-2024","DOIUrl":"https://doi.org/10.5194/gmd-17-4135-2024","url":null,"abstract":"Abstract. SnowModel, a spatially distributed snow-evolution modeling system, was parallelized using Coarray Fortran for high-performance computing architectures to allow high-resolution (1 m to hundreds of meters) simulations over large regional- to continental-scale domains. In the parallel algorithm, the model domain was split into smaller rectangular sub-domains that are distributed over multiple processor cores using one-dimensional decomposition. All the memory allocations from the original code were reduced to the size of the local sub-domains, allowing each core to perform fewer computations and requiring less memory for each process. Most of the subroutines in SnowModel were simple to parallelize; however, there were certain physical processes, including blowing snow redistribution and components within the solar radiation and wind models, that required non-trivial parallelization using halo-exchange patterns. To validate the parallel algorithm and assess parallel scaling characteristics, high-resolution (100 m grid) simulations were performed over several western United States domains and over the contiguous United States (CONUS) for a year. The CONUS scaling experiment had approximately 70 % parallel efficiency; runtime decreased by a factor of 1.9 running on 1800 cores relative to 648 cores (the minimum number of cores that could be used to run such a large domain because of memory and time limitations). CONUS 100 m simulations were performed for 21 years (2000–2021) using 46 238 and 28 260 grid cells in the x and y dimensions, respectively. Each year was simulated using 1800 cores and took approximately 5 h to run.\u0000","PeriodicalId":12799,"journal":{"name":"Geoscientific Model Development","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-05-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141109196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-21DOI: 10.5194/gmd-17-4115-2024
Rene Gassmöller, J. Dannberg, W. Bangerth, E. Puckett, Cedric Thieulot
Abstract. Numerical models are a powerful tool for investigating the dynamic processes in the interior of the Earth and other planets, but the reliability and predictive power of these discretized models depends on the numerical method as well as an accurate representation of material properties in space and time. In the specific context of geodynamic models, particle methods have been applied extensively because of their suitability for advection-dominated processes and have been used in applications such as tracking the composition of solid rock and melt in the Earth's mantle, fluids in lithospheric- and crustal-scale models, light elements in the liquid core, and deformation properties like accumulated finite strain or mineral grain size, along with many applications outside the Earth sciences. There have been significant benchmarking efforts to measure the accuracy and convergence behavior of particle methods, but these efforts have largely been limited to instantaneous solutions, or time-dependent models without analytical solutions. As a consequence, there is little understanding about the interplay of particle advection errors and errors introduced in the solution of the underlying transient, nonlinear flow equations. To address these limitations, we present two new dynamic benchmarks for transient Stokes flow with analytical solutions that allow us to quantify the accuracy of various advection methods in nonlinear flow. We use these benchmarks to measure the accuracy of our particle algorithm as implemented in the ASPECT geodynamic modeling software against commonly employed field methods and analytical solutions. In particular, we quantify if an algorithm that is higher-order accurate in time will allow for better overall model accuracy and verify that our algorithm reaches its intended optimal convergence rate. We then document that the observed increased accuracy of higher-order algorithms matters for geodynamic applications with an example of modeling small-scale convection underneath an oceanic plate and show that the predicted place and time of onset of small-scale convection depends significantly on the chosen particle advection method. Descriptions and implementations of our benchmarks are openly available and can be used to verify other advection algorithms. The availability of accurate, scalable, and efficient particle methods as part of the widely used open-source code ASPECT will allow geodynamicists to investigate complex time-dependent geodynamic processes such as elastic deformation, anisotropic fabric development, melt generation and migration, and grain damage.
{"title":"Benchmarking the accuracy of higher-order particle methods in geodynamic models of transient flow","authors":"Rene Gassmöller, J. Dannberg, W. Bangerth, E. Puckett, Cedric Thieulot","doi":"10.5194/gmd-17-4115-2024","DOIUrl":"https://doi.org/10.5194/gmd-17-4115-2024","url":null,"abstract":"Abstract. Numerical models are a powerful tool for investigating the dynamic processes in the interior of the Earth and other planets, but the reliability and predictive power of these discretized models depends on the numerical method as well as an accurate representation of material properties in space and time. In the specific context of geodynamic models, particle methods have been applied extensively because of their suitability for advection-dominated processes and have been used in applications such as tracking the composition of solid rock and melt in the Earth's mantle, fluids in lithospheric- and crustal-scale models, light elements in the liquid core, and deformation properties like accumulated finite strain or mineral grain size, along with many applications outside the Earth sciences. There have been significant benchmarking efforts to measure the accuracy and convergence behavior of particle methods, but these efforts have largely been limited to instantaneous solutions, or time-dependent models without analytical solutions. As a consequence, there is little understanding about the interplay of particle advection errors and errors introduced in the solution of the underlying transient, nonlinear flow equations. To address these limitations, we present two new dynamic benchmarks for transient Stokes flow with analytical solutions that allow us to quantify the accuracy of various advection methods in nonlinear flow. We use these benchmarks to measure the accuracy of our particle algorithm as implemented in the ASPECT geodynamic modeling software against commonly employed field methods and analytical solutions. In particular, we quantify if an algorithm that is higher-order accurate in time will allow for better overall model accuracy and verify that our algorithm reaches its intended optimal convergence rate. We then document that the observed increased accuracy of higher-order algorithms matters for geodynamic applications with an example of modeling small-scale convection underneath an oceanic plate and show that the predicted place and time of onset of small-scale convection depends significantly on the chosen particle advection method. Descriptions and implementations of our benchmarks are openly available and can be used to verify other advection algorithms. The availability of accurate, scalable, and efficient particle methods as part of the widely used open-source code ASPECT will allow geodynamicists to investigate complex time-dependent geodynamic processes such as elastic deformation, anisotropic fabric development, melt generation and migration, and grain damage.\u0000","PeriodicalId":12799,"journal":{"name":"Geoscientific Model Development","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141118360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-17DOI: 10.5194/gmd-17-4031-2024
Federica Castino, F. Yin, V. Grewe, H. Yamashita, S. Matthes, S. Dietmüller, Sabine Baumann, M. Soler, Abolfazl Simorgh, Maximilian Mendiguchia Meuser, F. Linke, B. Lührs
Abstract. The optimization of aircraft trajectories involves balancing operating costs and climate impact, which are often conflicting objectives. To achieve compromised optimal solutions, higher-level information such as preferences of decision-makers must be taken into account. This paper introduces the SolFinder 1.0 module, a decision-making tool designed to identify eco-efficient aircraft trajectories, which allow for the reduction of the flight's climate impact with limited cost penalties compared to cost-optimal solutions. SolFinder 1.0 offers flexible decision-making options that allow users to select trade-offs between different objective functions, including fuel use, flight time, NOx emissions, contrail distance, and climate impact. The module is included in the AirTraf 3.0 submodel, which optimizes trajectories under atmospheric conditions simulated by the ECHAM/MESSy Atmospheric Chemistry model. This paper focuses on the ability of the module to identify eco-efficient trajectories while solving a bi-objective optimization problem that minimizes climate impact and operating costs. SolFinder 1.0 enables users to explore trajectory properties at varying locations of the Pareto fronts without prior knowledge of the problem results and to identify solutions that limit the cost of reducing the climate impact of a single flight.
{"title":"Decision-making strategies implemented in SolFinder 1.0 to identify eco-efficient aircraft trajectories: application study in AirTraf 3.0","authors":"Federica Castino, F. Yin, V. Grewe, H. Yamashita, S. Matthes, S. Dietmüller, Sabine Baumann, M. Soler, Abolfazl Simorgh, Maximilian Mendiguchia Meuser, F. Linke, B. Lührs","doi":"10.5194/gmd-17-4031-2024","DOIUrl":"https://doi.org/10.5194/gmd-17-4031-2024","url":null,"abstract":"Abstract. The optimization of aircraft trajectories involves balancing operating costs and climate impact, which are often conflicting objectives. To achieve compromised optimal solutions, higher-level information such as preferences of decision-makers must be taken into account. This paper introduces the SolFinder 1.0 module, a decision-making tool designed to identify eco-efficient aircraft trajectories, which allow for the reduction of the flight's climate impact with limited cost penalties compared to cost-optimal solutions. SolFinder 1.0 offers flexible decision-making options that allow users to select trade-offs between different objective functions, including fuel use, flight time, NOx emissions, contrail distance, and climate impact. The module is included in the AirTraf 3.0 submodel, which optimizes trajectories under atmospheric conditions simulated by the ECHAM/MESSy Atmospheric Chemistry model. This paper focuses on the ability of the module to identify eco-efficient trajectories while solving a bi-objective optimization problem that minimizes climate impact and operating costs. SolFinder 1.0 enables users to explore trajectory properties at varying locations of the Pareto fronts without prior knowledge of the problem results and to identify solutions that limit the cost of reducing the climate impact of a single flight.\u0000","PeriodicalId":12799,"journal":{"name":"Geoscientific Model Development","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141126611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-28DOI: 10.5194/gmd-17-2447-2024
M. Awais, A. Vinca, E. Byers, S. Frank, Oliver Fricko, E. Boere, Peter, Burek, Miguel Poblete Cazenave, P. Kishimoto, A. Mastrucci, Y. Satoh, A. Palazzo, Madeleine McPherson, K. Riahi, V. Krey
Abstract. The integrated assessment model (IAM) MESSAGEix-GLOBIOM developed by IIASA is widely used to analyze global change and socioeconomic development scenarios within energy and land systems across different scales. However, to date, the representation of impacts from climate effects and water systems in the IAM has been limited. We present a new nexus module for MESSAGEix-GLOBIOM that improves the representation of climate impacts and enables the analysis of interactions between population, economic growth, energy, land, and water resources in a dynamic system. The module uses a spatially resolved representation of water systems to retain hydrological information without compromising computational feasibility. It maps simplified water availability and key infrastructure assumptions with the energy and land systems. The results of this study inform on the transformation pathways required under climate change impacts and mitigation scenarios. The pathways include multi-sectoral indicators highlighting the importance of water as a constraint in energy and land-use decisions and the implications of global responses to limited water availability from different sources, suggesting possible shifts in the energy and land sectors.
{"title":"MESSAGEix-GLOBIOM nexus module: integrating water sector and climate impacts","authors":"M. Awais, A. Vinca, E. Byers, S. Frank, Oliver Fricko, E. Boere, Peter, Burek, Miguel Poblete Cazenave, P. Kishimoto, A. Mastrucci, Y. Satoh, A. Palazzo, Madeleine McPherson, K. Riahi, V. Krey","doi":"10.5194/gmd-17-2447-2024","DOIUrl":"https://doi.org/10.5194/gmd-17-2447-2024","url":null,"abstract":"Abstract. The integrated assessment model (IAM) MESSAGEix-GLOBIOM developed by IIASA is widely used to analyze global change and socioeconomic development scenarios within energy and land systems across different scales. However, to date, the representation of impacts from climate effects and water systems in the IAM has been limited. We present a new nexus module for MESSAGEix-GLOBIOM that improves the representation of climate impacts and enables the analysis of interactions between population, economic growth, energy, land, and water resources in a dynamic system. The module uses a spatially resolved representation of water systems to retain hydrological information without compromising computational feasibility. It maps simplified water availability and key infrastructure assumptions with the energy and land systems. The results of this study inform on the transformation pathways required under climate change impacts and mitigation scenarios. The pathways include multi-sectoral indicators highlighting the importance of water as a constraint in energy and land-use decisions and the implications of global responses to limited water availability from different sources, suggesting possible shifts in the energy and land sectors.\u0000","PeriodicalId":12799,"journal":{"name":"Geoscientific Model Development","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140370898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-28DOI: 10.5194/gmd-17-2427-2024
Daniel Giles, Matthew M. Graham, Mosé Giordano, Tuomas Koskela, Alexandros Beskos, Serge Guillas
Abstract. Digital twins of physical and human systems informed by real-time data are becoming ubiquitous across weather forecasting, disaster preparedness, and urban planning, but researchers lack the tools to run these models effectively and efficiently, limiting progress. One of the current challenges is to assimilate observations in highly non-linear dynamical systems, as the practical need is often to detect abrupt changes. We have developed a software platform to improve the use of real-time data in non-linear system representations where non-Gaussianity limits the applicability of data assimilation algorithms such as the ensemble Kalman filter and variational methods. Particle-filter-based data assimilation algorithms have been implemented within a user-friendly open-source software platform in Julia – ParticleDA.jl. To ensure the applicability of the developed platform in realistic scenarios, emphasis has been placed on numerical efficiency and scalability on high-performance computing systems. Furthermore, the platform has been developed to be forward-model agnostic, ensuring that it is applicable to a wide range of modelling settings, for instance unstructured and non-uniform meshes in the spatial domain or even state spaces that are not spatially organized. Applications to tsunami and numerical weather prediction demonstrate the computational benefits and ease of using the high-level Julia interface with the package to perform filtering in a variety of complex models.
摘要以实时数据为基础的物理和人类系统数字孪生正在天气预报、备灾和城市规划中变得无处不在,但研究人员缺乏有效运行这些模型的工具,从而限制了研究的进展。目前的挑战之一是如何在高度非线性动态系统中吸收观测数据,因为实际需求往往是检测突然的变化。我们开发了一个软件平台,以改进非线性系统表征中实时数据的使用,在这种系统中,非高斯性限制了数据同化算法的适用性,如集合卡尔曼滤波器和变分法。基于粒子滤波的数据同化算法是在一个用户友好的 Julia 开放源码软件平台--ParticleDA.jl--上实现的。为确保所开发平台在现实场景中的适用性,重点放在高性能计算系统的数值效率和可扩展性上。此外,该平台的开发与前向模型无关,确保其适用于各种建模环境,例如空间域中的非结构化和非均匀网格,甚至是非空间组织的状态空间。在海啸和数值天气预报中的应用证明了使用该软件包的高级 Julia 界面在各种复杂模型中进行过滤的计算优势和便捷性。
{"title":"ParticleDA.jl v.1.0: a distributed particle-filtering data assimilation package","authors":"Daniel Giles, Matthew M. Graham, Mosé Giordano, Tuomas Koskela, Alexandros Beskos, Serge Guillas","doi":"10.5194/gmd-17-2427-2024","DOIUrl":"https://doi.org/10.5194/gmd-17-2427-2024","url":null,"abstract":"Abstract. Digital twins of physical and human systems informed by real-time data are becoming ubiquitous across weather forecasting, disaster preparedness, and urban planning, but researchers lack the tools to run these models effectively and efficiently, limiting progress. One of the current challenges is to assimilate observations in highly non-linear dynamical systems, as the practical need is often to detect abrupt changes. We have developed a software platform to improve the use of real-time data in non-linear system representations where non-Gaussianity limits the applicability of data assimilation algorithms such as the ensemble Kalman filter and variational methods. Particle-filter-based data assimilation algorithms have been implemented within a user-friendly open-source software platform in Julia – ParticleDA.jl. To ensure the applicability of the developed platform in realistic scenarios, emphasis has been placed on numerical efficiency and scalability on high-performance computing systems. Furthermore, the platform has been developed to be forward-model agnostic, ensuring that it is applicable to a wide range of modelling settings, for instance unstructured and non-uniform meshes in the spatial domain or even state spaces that are not spatially organized. Applications to tsunami and numerical weather prediction demonstrate the computational benefits and ease of using the high-level Julia interface with the package to perform filtering in a variety of complex models.\u0000","PeriodicalId":12799,"journal":{"name":"Geoscientific Model Development","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140372305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-25DOI: 10.5194/gmd-17-2419-2024
Rolf Sander
Abstract. The open-source software MEXPLORER 1.0.0 is presented here. The program can be used to analyze, reduce, and visualize complex chemical reaction mechanisms. The mathematics behind the tool is based on graph theory: chemical species are represented as vertices, and each reaction is described as a set of edges. MEXPLORER is a community tool published under the GNU General Public License.
摘要本文介绍了开源软件 MEXPLORER 1.0.0。该软件可用于分析、还原和可视化复杂的化学反应机理。该工具背后的数学基础是图论:化学物种表示为顶点,每个反应描述为一组边。MEXPLORER 是一款基于 GNU 通用公共许可证发布的社区工具。
{"title":"MEXPLORER 1.0.0 – a mechanism explorer for analysis and visualization of chemical reaction pathways based on graph theory","authors":"Rolf Sander","doi":"10.5194/gmd-17-2419-2024","DOIUrl":"https://doi.org/10.5194/gmd-17-2419-2024","url":null,"abstract":"Abstract. The open-source software MEXPLORER 1.0.0 is presented here. The program can be used to analyze, reduce, and visualize complex chemical reaction mechanisms. The mathematics behind the tool is based on graph theory: chemical species are represented as vertices, and each reaction is described as a set of edges. MEXPLORER is a community tool published under the GNU General Public License.\u0000","PeriodicalId":12799,"journal":{"name":"Geoscientific Model Development","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140382602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-22DOI: 10.5194/gmd-17-2387-2024
S. Fiedler, V. Naik, F. O’Connor, Christopher J. Smith, R. Pincus, Paul Griffiths, R. Kramer, T. Takemura, R. Allen, U. Im, M. Kasoar, A. Modak, S. Turnock, A. Voulgarakis, D. Watson‐Parris, Daniel, M. Westervelt, L. Wilcox, Alcide Zhao, W. Collins, Michael Schulz, G. Myhre, P. Forster
Abstract. The climate science community aims to improve our understanding of climate change due to anthropogenic influences on atmospheric composition and the Earth's surface. Yet not all climate interactions are fully understood, and uncertainty in climate model results persists, as assessed in the latest Intergovernmental Panel on Climate Change (IPCC) assessment report. We synthesize current challenges and emphasize opportunities for advancing our understanding of the interactions between atmospheric composition, air quality, and climate change, as well as for quantifying model diversity. Our perspective is based on expert views from three multi-model intercomparison projects (MIPs) – the Precipitation Driver Response MIP (PDRMIP), the Aerosol Chemistry MIP (AerChemMIP), and the Radiative Forcing MIP (RFMIP). While there are many shared interests and specializations across the MIPs, they have their own scientific foci and specific approaches. The partial overlap between the MIPs proved useful for advancing the understanding of the perturbation–response paradigm through multi-model ensembles of Earth system models of varying complexity. We discuss the challenges of gaining insights from Earth system models that face computational and process representation limits and provide guidance from our lessons learned. Promising ideas to overcome some long-standing challenges in the near future are kilometer-scale experiments to better simulate circulation-dependent processes where it is possible and machine learning approaches where they are needed, e.g., for faster and better subgrid-scale parameterizations and pattern recognition in big data. New model constraints can arise from augmented observational products that leverage multiple datasets with machine learning approaches. Future MIPs can develop smart experiment protocols that strive towards an optimal trade-off between the resolution, complexity, and number of simulations and their length and, thereby, help to advance the understanding of climate change and its impacts.
{"title":"Interactions between atmospheric composition and climate change – progress in understanding and future opportunities from AerChemMIP, PDRMIP, and RFMIP","authors":"S. Fiedler, V. Naik, F. O’Connor, Christopher J. Smith, R. Pincus, Paul Griffiths, R. Kramer, T. Takemura, R. Allen, U. Im, M. Kasoar, A. Modak, S. Turnock, A. Voulgarakis, D. Watson‐Parris, Daniel, M. Westervelt, L. Wilcox, Alcide Zhao, W. Collins, Michael Schulz, G. Myhre, P. Forster","doi":"10.5194/gmd-17-2387-2024","DOIUrl":"https://doi.org/10.5194/gmd-17-2387-2024","url":null,"abstract":"Abstract. The climate science community aims to improve our understanding of climate change due to anthropogenic influences on atmospheric composition and the Earth's surface. Yet not all climate interactions are fully understood, and uncertainty in climate model results persists, as assessed in the latest Intergovernmental Panel on Climate Change (IPCC) assessment report. We synthesize current challenges and emphasize opportunities for advancing our understanding of the interactions between atmospheric composition, air quality, and climate change, as well as for quantifying model diversity. Our perspective is based on expert views from three multi-model intercomparison projects (MIPs) – the Precipitation Driver Response MIP (PDRMIP), the Aerosol Chemistry MIP (AerChemMIP), and the Radiative Forcing MIP (RFMIP). While there are many shared interests and specializations across the MIPs, they have their own scientific foci and specific approaches. The partial overlap between the MIPs proved useful for advancing the understanding of the perturbation–response paradigm through multi-model ensembles of Earth system models of varying complexity. We discuss the challenges of gaining insights from Earth system models that face computational and process representation limits and provide guidance from our lessons learned. Promising ideas to overcome some long-standing challenges in the near future are kilometer-scale experiments to better simulate circulation-dependent processes where it is possible and machine learning approaches where they are needed, e.g., for faster and better subgrid-scale parameterizations and pattern recognition in big data. New model constraints can arise from augmented observational products that leverage multiple datasets with machine learning approaches. Future MIPs can develop smart experiment protocols that strive towards an optimal trade-off between the resolution, complexity, and number of simulations and their length and, thereby, help to advance the understanding of climate change and its impacts.\u0000","PeriodicalId":12799,"journal":{"name":"Geoscientific Model Development","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140212234","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-22DOI: 10.5194/gmd-17-2359-2024
C. Kerry, M. Roughan, Shane Keating, D. Gwyther, G. Brassington, Adil Siripatana, Joao Marcos A. C. Souza
Abstract. Ocean models must be regularly updated through the assimilation of observations (data assimilation) in order to correctly represent the timing and locations of eddies. Since initial conditions play an important role in the quality of short-term ocean forecasts, an effective data assimilation scheme to produce accurate state estimates is key to improving prediction. Western boundary current regions, such as the East Australia Current system, are highly variable regions, making them particularly challenging to model and predict. This study assesses the performance of two ocean data assimilation systems in the East Australian Current system over a 2-year period. We compare the time-dependent 4-dimensional variational (4D-Var) data assimilation system with the more computationally efficient, time-independent ensemble optimal interpolation (EnOI) system, across a common modelling and observational framework. Both systems assimilate the same observations: satellite-derived sea surface height, sea surface temperature, vertical profiles of temperature and salinity (from Argo floats), and temperature profiles from expendable bathythermographs. We analyse both systems' performance against independent data that are withheld, allowing a thorough analysis of system performance. The 4D-Var system is 25 times more expensive but outperforms the EnOI system against both assimilated and independent observations at the surface and subsurface. For forecast horizons of 5 d, root-mean-squared forecast errors are 20 %–60 % higher for the EnOI system compared to the 4D-Var system. The 4D-Var system, which assimilates observations over 5 d windows, provides a smoother transition from the end of the forecast to the subsequent analysis field. The EnOI system displays elevated low-frequency (>1 d) surface-intensified variability in temperature and elevated kinetic energy at length scales less than 100 km at the beginning of the forecast windows. The 4D-Var system displays elevated energy in the near-inertial range throughout the water column, with the wavenumber kinetic energy spectra remaining unchanged upon assimilation. Overall, this comparison shows quantitatively that the 4D-Var system results in improved predictability as the analysis provides a smoother and more dynamically balanced fit between the observations and the model's time-evolving flow. This advocates the use of advanced, time-dependent data assimilation methods, particularly for highly variable oceanic regions, and motivates future work into further improving data assimilation schemes.
摘要。海洋模式必须通过观测数据同化(数据同化)进行定期更新,以正确反映涡旋的时间和位置。由于初始条件对短期海洋预报的质量起着重要作用,因此采用有效的数据同化方案来生成准确的状态估计值是改进预测的关键。西边界洋流区(如东澳大利亚洋流系统)是高度多变的区域,因此对其建模和预测尤其具有挑战性。本研究评估了两个海洋数据同化系统在东澳大利亚洋流系统中两年的表现。我们在一个共同的建模和观测框架内,比较了与时间相关的四维变分(4D-Var)数据同化系统和计算效率更高的与时间无关的集合优化插值(EnOI)系统。两个系统同化了相同的观测数据:卫星得出的海面高度、海面温度、温度和盐度的垂直剖面(来自 Argo 浮漂),以及来自消耗性水深测量仪的温度剖面。我们根据不公开的独立数据分析了两个系统的性能,从而对系统性能进行了全面分析。4D-Var 系统的成本是 EnOI 系统的 25 倍,但在地表和地下的同化观测数据和独立观测数据方面,4D-Var 系统的性能优于 EnOI 系统。在 5 d 的预报范围内,EnOI 系统的均方根预报误差比 4D-Var 系统高 20%-60%。4D-Var 系统吸收了 5 d 窗口的观测数据,从预报结束到后续分析领域的过渡更加平滑。在预报窗口开始时,EnOI 系统显示出较高的低频(>1 d)表面强化温度变化和长度尺度小于 100 km 的较高动能。4D-Var 系统在整个水体的近惯性范围内显示出能量升高,而同化后的波数动能谱保持不变。总之,这种比较从数量上表明,4D-Var 系统提高了可预测性,因为分析结果在观测数据和模式的时变流之间提供了更平滑、更动态平衡的拟合。这就提倡使用先进的、随时间变化的数据同化方法,特别是在高度多变的海洋区域,并推动了今后进一步改进数据同化方案的工作。
{"title":"Comparison of 4-dimensional variational and ensemble optimal interpolation data assimilation systems using a Regional Ocean Modeling System (v3.4) configuration of the eddy-dominated East Australian Current system","authors":"C. Kerry, M. Roughan, Shane Keating, D. Gwyther, G. Brassington, Adil Siripatana, Joao Marcos A. C. Souza","doi":"10.5194/gmd-17-2359-2024","DOIUrl":"https://doi.org/10.5194/gmd-17-2359-2024","url":null,"abstract":"Abstract. Ocean models must be regularly updated through the assimilation of observations (data assimilation) in order to correctly represent the timing and locations of eddies. Since initial conditions play an important role in the quality of short-term ocean forecasts, an effective data assimilation scheme to produce accurate state estimates is key to improving prediction. Western boundary current regions, such as the East Australia Current system, are highly variable regions, making them particularly challenging to model and predict. This study assesses the performance of two ocean data assimilation systems in the East Australian Current system over a 2-year period. We compare the time-dependent 4-dimensional variational (4D-Var) data assimilation system with the more computationally efficient, time-independent ensemble optimal interpolation (EnOI) system, across a common modelling and observational framework. Both systems assimilate the same observations: satellite-derived sea surface height, sea surface temperature, vertical profiles of temperature and salinity (from Argo floats), and temperature profiles from expendable bathythermographs. We analyse both systems' performance against independent data that are withheld, allowing a thorough analysis of system performance. The 4D-Var system is 25 times more expensive but outperforms the EnOI system against both assimilated and independent observations at the surface and subsurface. For forecast horizons of 5 d, root-mean-squared forecast errors are 20 %–60 % higher for the EnOI system compared to the 4D-Var system. The 4D-Var system, which assimilates observations over 5 d windows, provides a smoother transition from the end of the forecast to the subsequent analysis field. The EnOI system displays elevated low-frequency (>1 d) surface-intensified variability in temperature and elevated kinetic energy at length scales less than 100 km at the beginning of the forecast windows. The 4D-Var system displays elevated energy in the near-inertial range throughout the water column, with the wavenumber kinetic energy spectra remaining unchanged upon assimilation. Overall, this comparison shows quantitatively that the 4D-Var system results in improved predictability as the analysis provides a smoother and more dynamically balanced fit between the observations and the model's time-evolving flow. This advocates the use of advanced, time-dependent data assimilation methods, particularly for highly variable oceanic regions, and motivates future work into further improving data assimilation schemes.\u0000","PeriodicalId":12799,"journal":{"name":"Geoscientific Model Development","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140211832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-21DOI: 10.5194/gmd-17-2347-2024
Leonardo Olivetti, Gabriele Messori
Abstract. In recent years, deep learning models have rapidly emerged as a stand-alone alternative to physics-based numerical models for medium-range weather forecasting. Several independent research groups claim to have developed deep learning weather forecasts that outperform those from state-of-the-art physics-based models, and operational implementation of data-driven forecasts appears to be drawing near. However, questions remain about the capabilities of deep learning models with respect to providing robust forecasts of extreme weather. This paper provides an overview of recent developments in the field of deep learning weather forecasts and scrutinises the challenges that extreme weather events pose to leading deep learning models. Lastly, it argues for the need to tailor data-driven models to forecast extreme events and proposes a foundational workflow to develop such models.
{"title":"Advances and prospects of deep learning for medium-range extreme weather forecasting","authors":"Leonardo Olivetti, Gabriele Messori","doi":"10.5194/gmd-17-2347-2024","DOIUrl":"https://doi.org/10.5194/gmd-17-2347-2024","url":null,"abstract":"Abstract. In recent years, deep learning models have rapidly emerged as a stand-alone alternative to physics-based numerical models for medium-range weather forecasting. Several independent research groups claim to have developed deep learning weather forecasts that outperform those from state-of-the-art physics-based models, and operational implementation of data-driven forecasts appears to be drawing near. However, questions remain about the capabilities of deep learning models with respect to providing robust forecasts of extreme weather. This paper provides an overview of recent developments in the field of deep learning weather forecasts and scrutinises the challenges that extreme weather events pose to leading deep learning models. Lastly, it argues for the need to tailor data-driven models to forecast extreme events and proposes a foundational workflow to develop such models.\u0000","PeriodicalId":12799,"journal":{"name":"Geoscientific Model Development","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140223509","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-21DOI: 10.5194/gmd-17-2325-2024
V. Ogarko, Kim Frankcombe, Taige Liu, J. Giraud, Roland Martin, M. Jessell
Abstract. We present a major release of the Tomofast-x open-source gravity and magnetic inversion code that incorporates several functionalities enhancing its performance and applicability for both industrial and academic studies. The code has been re-designed with a focus on real-world mineral exploration scenarios, while offering flexibility for applications at regional scale or for crustal studies. This new version includes several major improvements: magnetisation vector inversion, inversion of multi-component magnetic data, wavelet compression, improved handling of topography with support for non-uniform grids, a new and efficient parallelisation scheme, a flexible parameter file, and optimised input–output operations. Extensive testing has been conducted on a large synthetic dataset and field data from a prospective area of the Eastern Goldfields (Western Australia) to explore new functionalities with a focus on inversion for magnetisation vectors and magnetic susceptibility, respectively. Results demonstrate the effectiveness of Tomofast-x 2.0 in real-world studies in terms of both the recovery of subsurface features and performances on shared and distributed memory machines. Overall, with its updated features, improved capabilities, and performances, the new version of Tomofast-x provides a free open-source, validated advanced and versatile tool for constrained gravity and magnetic inversion.
{"title":"Tomofast-x 2.0: an open-source parallel code for inversion of potential field data with topography using wavelet compression","authors":"V. Ogarko, Kim Frankcombe, Taige Liu, J. Giraud, Roland Martin, M. Jessell","doi":"10.5194/gmd-17-2325-2024","DOIUrl":"https://doi.org/10.5194/gmd-17-2325-2024","url":null,"abstract":"Abstract. We present a major release of the Tomofast-x open-source gravity and magnetic inversion code that incorporates several functionalities enhancing its performance and applicability for both industrial and academic studies. The code has been re-designed with a focus on real-world mineral exploration scenarios, while offering flexibility for applications at regional scale or for crustal studies. This new version includes several major improvements: magnetisation vector inversion, inversion of multi-component magnetic data, wavelet compression, improved handling of topography with support for non-uniform grids, a new and efficient parallelisation scheme, a flexible parameter file, and optimised input–output operations. Extensive testing has been conducted on a large synthetic dataset and field data from a prospective area of the Eastern Goldfields (Western Australia) to explore new functionalities with a focus on inversion for magnetisation vectors and magnetic susceptibility, respectively. Results demonstrate the effectiveness of Tomofast-x 2.0 in real-world studies in terms of both the recovery of subsurface features and performances on shared and distributed memory machines. Overall, with its updated features, improved capabilities, and performances, the new version of Tomofast-x provides a free open-source, validated advanced and versatile tool for constrained gravity and magnetic inversion.\u0000","PeriodicalId":12799,"journal":{"name":"Geoscientific Model Development","volume":null,"pages":null},"PeriodicalIF":5.1,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140221044","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}