ITS1.4/CL0.10 | Advancing Earth System Models using Machine Learning
EDI
Advancing Earth System Models using Machine Learning
Convener: Jack AtkinsonECSECS | Co-conveners: Will Chapman, Laura MansfieldECSECS
Orals
| Wed, 30 Apr, 14:00–15:45 (CEST)
 
Room -2.33
Posters on site
| Attendance Wed, 30 Apr, 16:15–18:00 (CEST) | Display Wed, 30 Apr, 14:00–18:00
 
Hall X5
Posters virtual
| Attendance Fri, 02 May, 14:00–15:45 (CEST) | Display Fri, 02 May, 08:30–18:00
 
vPoster spot 2
Orals |
Wed, 14:00
Wed, 16:15
Fri, 14:00
Machine learning (ML) is being used throughout the geophysical sciences with a wide variety of applications. Advances in big data, deep learning, and other areas of artificial intelligence (AI) have opened up a number of new approaches to traditional problems.

Many fields (climate, ocean, NWP, space weather etc.) make use of large numerical models and are now seeking to enhance these by combining them with scientific ML/AI techniques. Examples include ML emulation of computationally intensive processes, data-driven parameterisations for sub-grid processes, ML assisted calibration and uncertainty quantification of parameters, amongst other applications.

Doing this brings a number of unique challenges, however, including but not limited to:
- enforcing physical compatibility and conservation laws, and incorporating physical intuition,
- ensuring numerical stability,
- coupling of numerical models to ML frameworks and language interoperation,
- handling computer architectures and data transfer,
- adaptation/generalisation to different models/resolutions/climatologies,
- explaining, understanding, and evaluating model performance and biases.
- quantifying uncertainties and their sources
- tuning of physical or ML parameters after coupling to numerical models (derivative-free optimisation, Bayesian optimisation, ensemble Kalman methods, etc.)

Addressing these requires knowledge of several areas and builds on advances already made in domain science, numerical simulation, machine learning, high performance computing, data assimilation etc.

We solicit talks that address any topics relating to the above. Anyone working to combine machine learning techniques with numerical modelling is encouraged to participate in this session.

Orals: Wed, 30 Apr | Room -2.33

Chairpersons: Jack Atkinson, Laura Mansfield, Will Chapman
14:00–14:05
14:05–14:15
|
EGU25-13920
|
ECS
|
On-site presentation
Sambit Kumar Panda, Todd Jones, Muhammad Shahzad, Anna-Louise Ellis, and Bryan Lawrence

Accurate representation of turbulent processes remains a critical challenge in atmospheric modelling. Large Eddy Simulations (LES) serve as valuable tools for understanding atmospheric turbulence by explicitly resolving energy-containing eddies while parameterizing smaller-scale motions through subgrid-scale (SGS) models. In their most complex forms, these SGS parameterizations can significantly influence LES performance and computational efficiency, making their improvement useful for advancing atmospheric modelling capabilities. Neural Network based emulation of such parametrizations have proven effective in reducing the computational cost, while maintaining accuracy and stability.

Building upon recent advances in physics-informed neural networks (NN) for atmospheric modelling and emulation of physics-based processes, we present a physics-guided NN architecture for emulation of the SGS turbulence parameterizations that introduces several key innovations. Our approach uniquely combines scale-specific normalization with multi-scale feature extraction through parallel convolutional paths, distinguishing it from existing physics-guided machine learning frameworks. The deep learning-based (DL) model also incorporates physically-motivated constraints across different spatial scales while simultaneously ensuring conservation of momentum and energy.

Unlike earlier studies that focus on single aspects of physical conservation, our architecture implements a comprehensive physics-informed framework that combines Richardson number gradient handling for stability constraints, with explicit treatment of diffusion and viscosity coefficients, and scale-specific normalization for different atmospheric variables. The model was trained on limited high-resolution Radiative-Convective Equilibrium (RCE) simulations from the Met Office-Natural Environment Research Council (NERC) Cloud model (MONC), employing physics-based loss functions that enforce both conservation laws and stability constraints.

The training dataset consisted of 3-D diagnostics data from the RCE simulations, with a 64x64 km2 domain and 1 km grid spacing in the horizontal. While the original simulations had 99 vertical levels with varying vertical resolution, the DL model was trained on random slices (vertical levels) chosen from the original data volume. The inputs consisted of the resolved state variables like velocity components (u, v, w) from the previous time step, the perturbations to potential temperature, mixing ratios and Richardson number, whereas the targets for the DL model were the SGS tendencies of the model prognostic fields resulting from the Smagorisnky parameterization and the coefficients of viscosity and diffusion.

The DL model's cross-regime applicability was evaluated through multiple independent test cases: (200-second sampling frequency) and different atmospheric conditions from the Atmospheric Radiation Measurement (ARM) program. The simulations from ARM atmospheric settings were mainly targeted at simulating shallow convection, with different grid/domain configurations. Results from the off-line tests demonstrate promising performance in predicting SGS and transport coefficients (viscosity and diffusion) across these varied conditions, particularly in maintaining physical consistency during regime transitions.

Our preliminary findings indicate that this enhanced multi-scale, physics-informed architecture can effectively learn SGS parameterizations from limited training data while maintaining physical fidelity across different atmospheric conditions and spatio-temporal resolutions. This approach demonstrates the potential for the development of high-fidelity, generalizable parameterizations for weather and climate models, suggesting a route forward for reducing the greater computational costs associated with more complex SGS parameterization schemes.

How to cite: Panda, S. K., Jones, T., Shahzad, M., Ellis, A.-L., and Lawrence, B.: Physics-Guided Deep Learning-Based Emulation of Subgrid-Scale Turbulence Parameterization for Atmospheric Large Eddy Simulations, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13920, https://doi.org/10.5194/egusphere-egu25-13920, 2025.

14:15–14:25
|
EGU25-14742
|
ECS
|
On-site presentation
NORi: A Novel, Physically-Principled Approach to Parameterization of Upper Ocean Turbulence using Neural Ordinary Differential Equations
(withdrawn)
Xin Kai Lee, Ali Ramadhan, Andre Souza, Gregory Wagner, Simone Silvestri, John Marshall, and Raffaele Ferrari
14:25–14:35
|
EGU25-11747
|
ECS
|
On-site presentation
Helena Reid and Cyril Morcrette

Stochastic parameterisations have seen widespread use in atmospheric models. These schemes represent uncertainty by adding terms that include a random noise component directly to the equations that describe the time evolution of the model. Stochastic parameterisation development thus involves the following questions: what are the sources of uncertainty, how do we represent them, and how precisely should we formulate stochastic terms to quantify them? Common methods to quantify the uncertainty inherent in parameterisation include applying multiplicative perturbations to physics tendencies (such that the larger the tendency due to subgrid processes, the more uncertainty we should have in the tendency) or applying perturbations to physical parameters (our physics schemes often rely on parameters whose values we do not know precisely, and have complicated nonlinear responses to perturbing this set of parameters, so perturbing each one within its own specified range during the model run allows this uncertainty to feed back into the model state).

In this work we present a different approach to stochastic parameterisation. We perturb the thermodynamic profiles that constitute the inputs to parameterisation schemes. The perturbations are scaled by the degree of subgrid inhomogeneity. A representation of the subgrid inhomogeneity is estimated by a machine learning model which has been trained on coarse-grained high resolution (dx=~1.5km) model output from the Met Office Unified Model. The scheme is implemented in LFRic, the UK Met Office’s next generation modelling system, and we present results of experiments ran in single column model mode.

How to cite: Reid, H. and Morcrette, C.: A new stochastic physics scheme incorporating machine-learnt subgrid variability, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11747, https://doi.org/10.5194/egusphere-egu25-11747, 2025.

14:35–14:45
|
EGU25-15200
|
ECS
|
On-site presentation
Tom Beucler, Arthur Grundner, Sara Shamekh, Peter Ukkonen, Matthew Chantry, and Ryan Lagerquist

The added value of machine learning for weather and climate applications is measurable through performance metrics, but explaining it remains challenging, particularly for large deep learning models. Inspired by climate model hierarchies, we propose that a full hierarchy of Pareto-optimal models, defined within an appropriately determined error-complexity plane, can guide model development and help understand the models' added value. We demonstrate the use of Pareto fronts in atmospheric physics through three sample applications, with hierarchies ranging from semi-empirical models with minimal parameters (simplest) to deep learning algorithms (most complex). First, in cloud cover parameterization, we find that neural networks identify nonlinear relationships between cloud cover and its thermodynamic environment, and assimilate previously neglected features such as vertical gradients in relative humidity that improve the representation of low cloud cover. This added value is condensed into a ten-parameter equation that rivals deep learning models. Second, we establish a machine learning model hierarchy for emulating shortwave radiative transfer, distilling the importance of bidirectional vertical connectivity for accurately representing absorption and scattering, especially for multiple cloud layers. Third, we emphasize the importance of convective organization information when modeling the relationship between tropical precipitation and its surrounding environment. We discuss the added value of temporal memory when high-resolution spatial information is unavailable, with implications for precipitation parameterization. Therefore, by comparing data-driven models directly with existing schemes using Pareto optimality, we promote process understanding by hierarchically unveiling system complexity, with the hope of improving the trustworthiness of machine learning models in atmospheric applications.

Preprint: https://arxiv.org/abs/2408.02161

How to cite: Beucler, T., Grundner, A., Shamekh, S., Ukkonen, P., Chantry, M., and Lagerquist, R.: Distilling Machine Learning's Added Value: Pareto Fronts in Atmospheric Applications, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15200, https://doi.org/10.5194/egusphere-egu25-15200, 2025.

14:45–14:55
|
EGU25-12601
|
ECS
|
On-site presentation
Maximilian Gelbrecht, Milan Klöwer, and Niklas Boers

Differentiable programming enables automatic differentiation (AD) tools to compute gradients through code without manually defining derivatives. AD tools can differentiate through entire software stacks, composing many functions and algorithms via the chain rule. With models that incorporate differentiable programming long-standing challenges like systematic calibration, comprehensive sensitivity analyses, and uncertainty quantification can be tackled, and machine learning (ML) methods can be integrated directly into the process-based core of earth-system models (ESMs) to incorporate additional information from observations. Through the advent of ML, several AD tools are gaining traction. A new generation of powerful tools like JAX, Zygote and Enzyme enable differentiable programming for models of varying complexity including highly complex coupled ESMs. Here we present an overview about the perspectives of differentiable programming for ESMs, using experience from two of our applications in atmospheric modelling. First off, we set up PseudoSpectralNet, a differentiable quasi-geostrophic model in Julia with Zygote. This is a hybrid model combining neural networks with a dynamical core, showcasing how the stability and accuracy of ML models is improved by integrating a process-based dynamical core into our model. Additionally, ongoing work uses Enzyme to achieve a differentiable version of the significantly more complex SpeedyWeather.jl atmospheric model. We will discuss advantages of both approaches and give an outlook into future possibilities with differentiable models. 

How to cite: Gelbrecht, M., Klöwer, M., and Boers, N.: Differentiable Programming for Atmospheric Models: Experiences and Perspectives , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12601, https://doi.org/10.5194/egusphere-egu25-12601, 2025.

14:55–15:05
|
EGU25-16817
|
On-site presentation
Christian Lessig

The training of machine learning models for weather and climate on multiple datasets, including local high-resolution reanalyses and level-1 observations, is one of the frontiers of the field. It promises to allow for models that are no longer constrained by the capabilities of equation-based models, as is currently still largely the case when training on global reanalyses. For example, level-1 observations contain the feedback from arbitrary scale processes and hence do not suffer from the closure problem of equation-based models. Training on observations might hence lead to machine learning-based Earth system models with reduced systematic biases, in particular for long-term climate projections. Local reanalyses are only available for a small set of regions, mainly over Europe and North America. Appropriate training might allow one to generalize the detailed process information in these to other regions or even globally. 

In this talk, we present results on the effective training with a combination of global and local reanalysis as well as level-1 observations. We consider different pre-training protocols to learn the correlations between datasets, which is critical to obtain a benefit through their combination. We use a forecasting task as baseline and study the effectiveness of different variants of masked-token modeling and more sophisticated approaches that exploit the latent space of the machine learning models. We also study different fine-tuning strategies to extract a best state estimate from multiple datasets and to generalize regional datasets globally. For this, we build on the extensive results on fine-tuning of large language models that have been developed in the last years. Our results aim to determine general principles which combination of datasets is beneficial. We also perform a detailed analysis of the physical consistency and physical process representation in the model output. Through this, we believe our work provides an important stepping stone for the next generation of machine learning-based models for weather and climate.

How to cite: Lessig, C.: Towards next generation machine learning-based Earth system models that exploit a wide range of datasets, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16817, https://doi.org/10.5194/egusphere-egu25-16817, 2025.

15:05–15:15
|
EGU25-5159
|
ECS
|
Virtual presentation
Pritthijit Nath, Henry Moss, Emily Shuckburgh, and Mark Webb

This study explores integrating reinforcement learning (RL) with idealised climate models to address key parameterisation challenges in climate science. Current climate models rely on complex mathematical parameterisations to represent sub-grid scale processes, which can introduce substantial uncertainties. RL offers capabilities to enhance these parameterisation schemes, including direct interaction, handling sparse or delayed feedback, continuous online learning, and long-term optimisation. We evaluate the performance of eight RL algorithms on two idealised environments: one for temperature bias correction, another for radiative-convective equilibrium (RCE) imitating real-world computational constraints. Results show different RL approaches excel in different climate scenarios with exploration algorithms performing better in bias correction, while exploitation algorithms proving more effective for RCE. These findings support the potential of RL-based parameterisation schemes to be integrated into global climate models, improving accuracy and efficiency in capturing complex climate dynamics. Overall, this work represents an important first step towards leveraging RL to enhance climate model accuracy, critical for improving climate understanding and predictions. Code accessible at https://github.com/p3jitnath/climate-rl.

How to cite: Nath, P., Moss, H., Shuckburgh, E., and Webb, M.: RAIN: Reinforcement Algorithms for Improving Numerical Weather and Climate Models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5159, https://doi.org/10.5194/egusphere-egu25-5159, 2025.

15:15–15:25
|
EGU25-17333
|
ECS
|
On-site presentation
Marco Mariano De Carlo, Gabriele Accarino, Igor Atake, Donatello Elia, Italo Epicoco, and Giovanni Coppini

Accurate oil spill predictions are crucial for mitigating environmental and socioeconomic impacts. Numerical models, like MEDSLIK-II (De Dominicis et al., 2013), simulate oil advection, dispersion, and transformation, but their performance depends heavily on the configuration of physical parameters, often requiring labor-intensive manual tuning based on expert judgment.

To address this limitation, we integrate MEDSLIK-II with a Bayesian Optimization (BO) framework to systematically identify the optimal parameter configuration, ensuring simulations closely match observed spatiotemporal oil spill distributions. Our optimization focuses on horizontal diffusivity and drift factor parameters, using the Fraction Skill Score as the objective metric to maximize, thus reducing the overlap between simulations and observations.

The approach is validated on the 2021 Baniyas (Syria) oil spill, demonstrating improved accuracy, reduced biases and lower computational costs compared to the standalone numerical model.

By integrating BO with the MEDSLIK-II numerical model, our method enhances oil spill prediction capabilities and provides a transferable, physically consistent optimization framework applicable to a wide range of geophysical challenges.

This work is conducted within the framework of the iMagine European project, which leverages Artificial Intelligence, including AI-assisted image generation, to advance a series of use cases in marine and oceanographic science.

 

References

De Dominicis, M., Pinardi, N., Zodiatis, G., & Lardner, R. (2013). MEDSLIK-II, a Lagrangian marine surface oil spill model for short-term forecasting – Part 1: Theory. Geoscientific Model Development, 6, 1851–1869. https://doi.org/10.5194/gmd-6-1851-2013

How to cite: De Carlo, M. M., Accarino, G., Atake, I., Elia, D., Epicoco, I., and Coppini, G.: Improving Oil Spill Numerical Simulations through Bayesian Optimization, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17333, https://doi.org/10.5194/egusphere-egu25-17333, 2025.

15:25–15:35
|
EGU25-4565
|
On-site presentation
Darri Eythorsson, Kasra Keshavarz, Cyril Thébault, Mohamed Ahmed, Raymond Spiteri, Alain Pietroniro, and Martyn Clark

AI enhanced environmental modelling workflows: Towards Automated Scientific Exploration in Hydrology

Authors: Darri Eythorsson, Kasra Keshavarz, Cyril Thébault, Mohamed Ismaiel Ahmed, Raymond Spiteri, Alain Pietroniro and Martyn Clark

Modern hydrological modeling has evolved into a complex scientific endeavour requiring sophisticated workflows that span multiple scales, processes, and computational paradigms. While existing workflow solutions address specific technical challenges, the field lacks comprehensive frameworks that can support end-to-end modeling while maintaining reproducibility and scalability. This works introduces two complementary frameworks that aim to address these fundamental challenges: CONFLUENCE (Community Optimization Nexus For Large-domain Understanding of Environmental Networks and Computational Exploration) and INDRA (the Intelligent Network for Dynamic River Analysis).

CONFLUENCE implements a modular architecture that enforces workflow reproducibility through a unified configuration system while maintaining the flexibility needed to support diverse modeling applications. The framework provides comprehensive solutions for four critical workflow components: (1) flexible geospatial domain definition and discretization, (2) model-agnostic data acquisition and preprocessing, (3) extensible model setup and parameterization capabilities, and (4) comprehensive evaluation and optimization tools. This systematic approach enables efficient, reproducible, and transparent hydrological modeling across scales.

INDRA augments this foundation by implementing a network of specialized AI expert agents that support various components of the hydrological modeling workflow. Through structured dialogue between domain experts (including AI specialists in hydrology, hydrogeology, meteorology, data science, and geospatial analysis), INDRA provides context-aware guidance while maintaining complete provenance of modeling decisions and their justification. This AI-assisted approach helps address three critical challenges: (1) the growing complexity of modelling decisions, (2) the need for reproducible workflows and detailed documentation, and (3) the technical barriers limiting broader adoption of advanced modeling practices.

The integration of these frameworks aims to explore how automation and AI assistance can enhance rather than disrupt traditional modeling practices. By maintaining clear documentation of decisions and their justifications, these systems help build trust in model results while creating opportunities for recursive learning from previous modeling experiments. Our case studies, spanning scales from individual catchments to continental domains, showcase the frameworks' capabilities while highlighting their potential to transform how researchers’ interface with complex environmental modeling workflows.

This work aims to advance both operational and research oriented hydrological modeling practices, offering a foundation for reproducible, scalable, and interoperable modeling while maintaining scientific rigor and flexibility. The framework’s open-source nature and modular design create opportunities for community-driven development and extension, potentially accelerating scientific discovery in hydrological sciences.

How to cite: Eythorsson, D., Keshavarz, K., Thébault, C., Ahmed, M., Spiteri, R., Pietroniro, A., and Clark, M.: AI enhanced environmental modelling workflows: Towards Automated Scientific Exploration in Hydrology, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4565, https://doi.org/10.5194/egusphere-egu25-4565, 2025.

15:35–15:45
|
EGU25-17542
|
On-site presentation
Martin V. Butz, Matthias Karlbauer, Frank Beyrich, and Volker Wulfmeyer

Vertical energy transport from the land surface into the atmosphere in the form of sensible and latent heat flux must be well represented in numerical weather prediction models to allow accurate estimates of near-surface atmospheric variables. Traditionally, these heat fluxes are parameterized relying on Monin-Obukhov Similarity Theory (MOST), which is based on differences in wind speed, air temperature, and humidity between adjacent measurement or model levels. Recently, Wulfmeyer et al. (2024) estimated heat flux with machine learning at much higher accuracy compared to MOST. Their ML model proposed the incorporation of additional predictor variables when estimating latent heat flux (such as solar radiation), which stands in contrast to the classical MOST approach. However, the analysis in Wulfmeyer et al. (2024) is based on a rather short data period in August 2017 at three nearby locations in Oklahoma, USA, which limits the generalizability of the results. Here, we replicate and expand the findings from Wulfmeyer et al. (2024) on a dataset from the boundary layer field site (GM) Falkenberg of the German Meteorological Service over a period of twelve years, covering various seasons and synoptic weather situations. Our findings support the role of incoming shortwave radiation not only for latent but also for sensible heat flux estimates, particularly for other parts of the year. The results thus underline the potential to develop more advanced flux parameterizations beyond MOST. In future research, we intend to investigate the role of other predictor variables, such as vapor pressure deficit or soil moisture, to assess the generalizability of the relations, to judge their performance under extreme conditions, and to derive simple but universally applicable parameterizations.

How to cite: Butz, M. V., Karlbauer, M., Beyrich, F., and Wulfmeyer, V.: Replicating Sensible and Latent Heat Flux Diagnosis with Multilayer Perceptrons on Multi-Year Falkenberg Tower Data , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17542, https://doi.org/10.5194/egusphere-egu25-17542, 2025.

Posters on site: Wed, 30 Apr, 16:15–18:00 | Hall X5

Display time: Wed, 30 Apr, 14:00–18:00
Chairpersons: Jack Atkinson, Laura Mansfield, Will Chapman
X5.130
|
EGU25-288
|
ECS
Waleed AlGharbi, Rebecca Bell, and Cédric John

Seismic data forms the backbone of what we understand in the subsurface, and seismic data interpretation is still usually done by hand. Automatic seismic interpretation with deep learning is very promising, but there the problem is a lack of labelled training data. In this study, we use forward stratigraphic modelling and show how forward modelling can be advantageously used in deep learning.

Specifically, we focus on shelf-edge trajectories as the geological representations of lateral and vertical shifts in sediments’ position through time. They provide continuous tracks of changes in relative sea-level as well as sediment stacking patterns and depositional geometries. Mapping these trajectories and measuring their changing angles help in quantifying the sequence stratigraphic analysis and predicting ancient depositional environments.

Here, we evaluate the ability of deep learning models, trained on synthetic seismic data, to identify clinoforms and their rollover points for shelf-edge trajectories mapping. The synthetic training dataset generated using geological processed-based forward modelling represents different depositional slope scenarios. Controlling the different parameters that govern shelf-edges and shelf-edge trajectories (such as bathymetry, sediment supply, eustatic sea-level changes and subsidence) gave us a better chance to mimic realistic and diverse depositional setting, which helps in generalizing the deep learning model. In addition, the ground truth (labels) for the created synthetic seismic data is automatically generated by the forward model, without the need of manual labelling seismic data.

Higher accuracy score on both validation and testing datasets demonstrates the power and effectiveness of using synthetic as training dataset. This study shows that synthetic data can play a major role in bridging the gap between traditional seismic interpretation and automating the process using machine learning. It also shows that forward modelling is a powerful technique to combine with data modelling, such as machine learning.

How to cite: AlGharbi, W., Bell, R., and John, C.: Forward Stratigraphic Modelling to Generate Synthetic Seismic Training Dataset for Deep Learning: A Case Study to Predict Shelf-Edge Trajectories, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-288, https://doi.org/10.5194/egusphere-egu25-288, 2025.

X5.131
|
EGU25-2197
|
ECS
Lou Wangqi

This study aims to develop a CNN-LSTM hybrid network model integrated with a coupled self-attention mechanism, based on deep learning techniques, to simulate flood processes in the Inner Harbor area of Macau. With global climate change and accelerated urbanization, Macau, a low-lying coastal city, frequently experiences urban flooding due to typhoons and heavy rainfall. While traditional hydrological and hydrodynamic models can accurately predict flooding processes, their computational intensity and lack of real-time responsiveness make them unsuitable for emergency disaster warnings. To address these limitations, this paper proposes a convolutional long short-term memory (ConvLSTM) model enhanced with a coupled self-attention mechanism. The model leverages an encoder-decoder structure to predict the evolution of flood processes under 4–10 hours of heavy rainfall in the Inner Harbor area of Macau.

The model integrates CNN components for extracting spatial features, LSTM components for capturing temporal features, and a coupled self-attention mechanism to dynamically reweight spatial-temporal representations, improving the model's sensitivity to key flood patterns. The encoder encodes input sequences into fixed-length vectors, while the decoder translates these vectors into target sequences. The self-attention mechanism ensures the model focuses on critical spatial and temporal regions, further enhancing prediction accuracy and robustness.

The training and testing datasets were constructed from simulation data generated by hydrological-hydrodynamic models and static geographical information data, following preprocessing and normalization. Evaluation metrics, including mean squared error (MSE), Nash-Sutcliffe efficiency coefficient (NSE), and relative error, were used to assess model performance. Results demonstrate that the proposed hybrid model, augmented by the coupled self-attention mechanism, effectively simulates maximum water depth distribution and flood evolution processes, achieving high consistency with hydrodynamic simulation data while providing improved predictive performance.

How to cite: Wangqi, L.: Flood Process Simulation in Macau's Inner Harbor Area Based on CNN-LSTM, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2197, https://doi.org/10.5194/egusphere-egu25-2197, 2025.

X5.132
|
EGU25-5991
|
ECS
|
|
Gianmarco Guglielmo and Pietro Prestininzi

Machine Learning is gaining increasing attention from the scientific community in hydrological and hydraulic research. However, this field faces a consistent challenge in applying data-driven approaches due to the evident poor generalization capabilities, which are partly a result of inherent data scarcity.

We propose incorporating expert knowledge into data-driven models for river hydraulics and flood mapping by integrating physically-based information without relying on the underlying mathematical formulation (e.g., the calculation of the residuals of differential equations). This approach appears to be particularly valuable for flood simulations, where hydraulically relevant distributed parameters such as roughness, lithology, topography etc. pose significant uncertainties. The method is versatile and applicable to physical systems and scenarios in which the underlying mathematical formulation is not fully known, but expert knowledge enables the introduction of meaningful, physically-inspired constraints.

Specifically, the physical information is integrated into the model by including an additional term, weighted by the hyperparameter in the guise of a regularization term in the loss function :

 

Here, represents the data-driven error metric, while the physical loss term is an error metric that depends not only on the true and predicted outputs ( ), but also potentially on the inputs . Indeed, this term employs physical principles, laws, and quantities, which are not explicitly formulated in the original dataset. In this sense, we can note its similarity to data augmentation, a widely used technique in machine learning that extracts additional insights by offering alternative interpretations of the same dataset.

We clarify that this approach does not aim to replace numerical solvers or serve as an alternative numerical model, as Physics-Informed Neural Networks do: indeed, their similarity is limited to the formulation of the modified loss function.

We assessed the methodology and empirically quantified the effectiveness of the method in a simplified, well-controlled problem, evaluating the gain in generalisation capability of Neural Networks (NNs) in the reconstruction of the steady state, one-dimensional, water surface profile in a rectangular channel. We found improved predictive capabilities, even when extrapolating beyond the boundaries of the training dataset and in data-scarce scenarios. This kind of assessment is of great relevance to the application of NNs to flood mapping, where cases featuring values of the observed quantities falling out of the range of the recorded series need to be predicted.

New experiments have been also conducted on two-dimensional domains. The data-driven model was trained on a single catchment and tested on its ability to determine flooded areas in unseen catchments. Preliminary results show that an encoder-decoder model with convolutional layers exhibits improved generalization when a physical training strategy is employed. Future applications could include flood mapping for ungauged basins, leveraging similarities with other basins.

How to cite: Guglielmo, G. and Prestininzi, P.: Physically-Enhanced Training of Neural Networks for Hydraulic Modelling of Rivers and Flood Events, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5991, https://doi.org/10.5194/egusphere-egu25-5991, 2025.

X5.133
|
EGU25-9987
|
ECS
Eric Asamoah, Gerard Heuvelink, Ikram Chairi, Prem Bindraban, and Vincent Logah

Background: Agriculture is increasingly leveraging machine learning (ML) to enhance yield predictions and optimize agronomic practices. Maize, a staple crop in Ghana, offers a valuable case study for evaluating the effectiveness of diverse ML models in yield prediction and resource management.

Objective: This study aims to evaluate the predictive performance of four ML models namely Random Forest (RF), Support Vector Machine (SVM), K-Nearest Neighbours (KNN), and Extreme Gradient Boosting (XGBoost) for maize yield and agronomic efficiency prediction. It also compares variable importance across these models to determine key explanatory variables.

Methods: The study utilized 4,496 georeferenced maize trial datasets from various agroecological zones in Ghana. Thirty-five explanatory variables included soil properties, climate, topography, crop management practices, and fertilizer application datasets. Model performance was evaluated using leave-one-out, leave-site-out, and leave-agroecological-zone-out cross-validation techniques. Metrics including Mean Error (ME), Root Mean Squared Error (RMSE), and Model Efficiency Coefficient (MEC) were used to compare model accuracy, while a permutation-based approach was employed to assess variable importance.

Results: XGBoost emerged as the most accurate model, achieving the lowest RMSE for yield (639.5 kg ha⁻¹) and agronomic efficiency (11.6 kg kg⁻¹), particularly for nitrogen (AE-N). RF demonstrated competitive performance, while KNN and SVM yielded inconsistent results under rigorous cross-validation conditions. Key explanatory variables identified across models included nitrogen fertilizer, rainfall, and crop genotype, underscoring their critical role in yield and agronomic efficiency outcomes.

Conclusion: XGBoost was the most robust and accurate model for maize yield and agronomic efficiency predictions, offering a reliable tool for data-driven agricultural planning in diverse agroecological settings. The findings underscore the transformative role of advanced ML techniques in modern agriculture, particularly in optimizing staple crop production in sub-Saharan Africa.

How to cite: Asamoah, E., Heuvelink, G., Chairi, I., Bindraban, P., and Logah, V.: Modelling Maize Yield and Agronomic Efficiency Using Machine Learning Models: A Comparative Analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9987, https://doi.org/10.5194/egusphere-egu25-9987, 2025.

X5.134
|
EGU25-10915
|
ECS
|
|
Sebastián Garzón, Willem Dabekaussen, Eva De Boever, Freek Busschers, Siamak Mehrkanoon, and Derek Karssenberg

Expert interpretation of borehole data is a critical component of geological modelling, offering essential insights into the spatial distribution of geological units within the subsurface. For large-scale regional mapping efforts, expert interpretation of all available data is impractical due to the sheer volume of boreholes. Therefore, many 3D geological subsurface models rely only on a small portion of all available data. Machine learning (ML) models can be used to automate borehole data interpretation, increasing data density. However, these automated interpretations must adhere to strict spatial and stratigraphical relationships to be consistent with the established geological knowledge of the area. Using a dataset of 1,400 boreholes with expert interpretations from the Roer Valley Graben (Southeast Netherlands), we explore how ML models can be integrated into geological modelling workflows, highlighting the challenge of ensuring compatibility with geological principles and known spatial relationships. We evaluate the model performance using traditional metrics such as accuracy, Cohen's kappa and F1 Score and newly proposed geology-inspired metrics to quantify the ability of Random Forest and Neural Network models to interpret borehole data into lithostratigraphic units while preserving key geological relationships. Our results demonstrate that while many models achieve accuracy values of 75% to 80%, Neural Networks perform significantly better in capturing the expected sequential relationships between geological units, achieving up to 96% of geological transitions between geological units that are plausible, compared to 65% for the best-performing Random Forest model selected based on traditional metrics. This study underscores the need for domain-specific metrics in evaluating model performance and the potential for ML to increase the volume of data incorporated in subsurface models.

How to cite: Garzón, S., Dabekaussen, W., De Boever, E., Busschers, F., Mehrkanoon, S., and Karssenberg, D.: Assessing the Geological Plausibility of Machine Learning Borehole Interpretations: A Case Study in the Roer Valley Graben, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10915, https://doi.org/10.5194/egusphere-egu25-10915, 2025.

X5.135
|
EGU25-11031
|
ECS
Herbert Rakotonirina, Théophile Lohier, Anne Raingeard, Frédéric Lacquement, Julien Baptiste, and Hélène Tissoux

Abstract:
Alluvial terraces in watersheds are geomorphic features formed by river incision and sediment deposit, representing former floodplain levels. They serve as valuable records of fluvial dynamics, climatic changes, and tectonic activity. Mapping methods for terraces that rely on field-acquired data, often involving physical or chemical analyses, are not feasible for large-scale applications. When aiming to map at the national scale, the development of a methodology that eliminates the need for such detailed information enhances scalability and broadens applicability.

We proposed a semi-automatic predictive mapping method for watershed terraces using 25m Digital Earth Model (DEM) provided by the IGN (French geographical service) and derived variables such as curvature, slope, and the difference from a base level (Raingeard et al., 2019). This method demonstrated meaningful results in the Pyrenean Piedmont for the Baïse and Ousse rivers, with the predicted map showing strong alignment with the geological reality.

In this study, we propose an automated approach for identifying alluvial terraces using relative height. Relative height is defined as the difference between the elevation derived from a DEM and the base level. Our methodology is based on the hypothesis that terraces are represented as flat areas in the relative height, where pixels exhibit similar statistical distributions. To capture these patterns, we employ a Gaussian Mixture Model, a probabilistic framework that approximates data as a combination of multiple Gaussian distributions. In this context, each Gaussian distribution corresponds to a specific alluvial terrace.

We conducted experiments on the study areas used by Raingeard et al. (2019), and the results are consistent with both the semi-automatic method and the geological reality. These outcomes provide promising prospects for the predictive mapping of superficial deposits

Reference:

Raingeard A., Tourlière B., Lacquement. F, Baptiste. J, Tissoux. H. Semi-automatic quaternary alluvial deposits mapping - Methodology for the predictive mapping of flat terrains within a watershed, by semi-automatic analysis of the Digital Elevation Model. INQUA 2019, Jul 2019, Dublin, Ireland. 2019.



How to cite: Rakotonirina, H., Lohier, T., Raingeard, A., Lacquement, F., Baptiste, J., and Tissoux, H.: Mapping Alluvial Terraces in Watersheds Using Gaussian Mixture Model on Relative Height., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11031, https://doi.org/10.5194/egusphere-egu25-11031, 2025.

X5.136
|
EGU25-12464
|
ECS
Hussam Eldin Elzain, Osman Abdalla, Ali Al-Maktoumi, Anvar Kacimov, and Mingjie Chen

Accurately forecasting water table rise (WTR) is essential for effective water resource management, infrastructure development, flood risk mitigation, and environmental conservation. This research employed multiple machine learning (ML) models, namely Ridge Linear Regression (RLR), Radial Basis Function Support Vector Machine (RBF-SVM), Linear SVM (LSVM), Random Forest (RF), and a hybrid deep learning Transformer (TR) with Bi-Long Short-Term Memory (BiLSTM), to forecast WTR one and two weeks ahead in the Muscat Governorate, Oman. A total of 19,465 high-resolution datasets, measured at half-hour intervals between December 2017 and January 2019, were utilized. The data were divided into training and testing sets, with 90% (17,976 datasets) used for training and the remaining 10% (1,489 datasets) reserved for testing. A two-way time series analysis was employed to analyze dynamic interactions between two time-dependent behaviors over time. Additionally, the rolling forecasting method was used alongside the models to capture patterns and provide updated predictions based on the most recent data trends. The results demonstrated that RLR outperformed both the individual ML models and the hybrid deep learning TR-BiLSTM models, as indicated by the NSE and RSR statistical metrics applied to the testing data. Furthermore, the one-week step-ahead forecasting achieved greater accuracy in predicting WTR compared to the two-week step-ahead forecast. However, the average computational time of the hybrid deep learning TR-BiLSTM models was notably higher compared to the standalone models. Linear models such as RLR and LSVM demonstrated accurate forecasting results due to their ability to prevent overfitting in correlated features and effectively capture the simplicity of the relationship between the data. The approach presented in this research can be effectively useful to various arid regions worldwide that are influenced by WTR.

How to cite: Elzain, H. E., Abdalla, O., Al-Maktoumi, A., Kacimov, A., and Chen, M.: Water table rise forecasting using machine and deep learning models in arid regions, Oman, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12464, https://doi.org/10.5194/egusphere-egu25-12464, 2025.

X5.137
|
EGU25-17078
|
ECS
Erisa Ismaili, Tom Beucler, and Robert Jnglin Wills

Mesoscale atmospheric processes that are neither resolved nor parameterized in global climate models, such as slantwise convection, can have a significant impact on climate variability and change. An example of such mesoscale influence on the climate is shown by Wills et al. 2024,  who demonstrate that circulation responses to surface anomalies are increased through heat and momentum fluxes by mesoscale processes. To enable longer simulations in comprehensive global climate models that include information about subgrid mesoscale processes, a machine learning (ML) parameterization could be applied at relatively low computational cost. So far, such ML parameterizations have been primarily applied to idealized geographies (e.g., aquaplanets), and they have not targeted midlatitude mesoscale processes in particular.

In this work, we focus on midlatitude mesoscale processes over the Gulf Stream region, as simulated by variable resolution CESM2 simulations, which have 14-km resolution over the North Atlantic. Learning subgrid fluxes from this model allows a targeted parametrization of mesoscale processes leading to vertical fluxes, namely slantwise convection and frontogenesis. We use an artificial neural network to predict vertical profiles of subgrid fluxes of momentum, heat and moisture. The features (inputs) for the ML models in this work include coarse-grained atmospheric state variables at each grid point, such as the vertical profiles of horizontal winds, temperature and their horizontal shear as well as surface pressure. The vertical profile of the specific humidity and the value of convective available potential energy are included to assess the importance of moist dynamics in the determination of subgrid convectional fluxes. Our results show that moisture variables have a rather small impact, suggesting that the subgrid fluxes can be explained by dry dynamics. A greater importance is found in the horizontal differences of neighbouring momentum and temperature columns. This suggests that neighbouring column information may be essential in the prediction of subgrid-scale fluxes, e.g., through the action of shear instabilities or conditional symmetric instability. Combined with information about the vertical localization relationship of the inputs and outputs, the goal is to feed this information into an equation discovery approach, which could lead to deeper physical understanding of mesoscale momentum and energy fluxes in midlatitudes.

 

How to cite: Ismaili, E., Beucler, T., and Jnglin Wills, R.: Prediction and Understanding of Subgrid-Scale Vertical Fluxes by Missing Midlatitude Mesoscale Processes Using a Machine Learning Approach, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17078, https://doi.org/10.5194/egusphere-egu25-17078, 2025.

X5.138
|
EGU25-17958
|
ECS
Maha Badri, Philipp Hess, Yunan Lin, Sebastian Bathiany, Maximilian Gelbrecht, and Niklas Boers

Vegetation plays a crucial role in the Earth's climate system via a number of key processes, affecting the exchange of carbon, water and energy between surface and atmosphere. The complex relationship between climate change and vegetation highlights the importance of accurate and reliable vegetation models that fully capture these interactions. 

Traditional vegetation models are primarily designed to operate on CPU architectures, which restricts their ability to exploit advancements in modern parallel computing architectures such as GPUs. Furthermore, limited process knowledge and the absence of direct observations and/or quantitative theories for certain processes hinder accurate representation of these processes, which introduces uncertainties in the model results, leading to discrepancies when compared to observations. The rigid structure of these traditional models also makes integration of new processes challenging and hinders the application of advanced optimization techniques for automatic parameter tuning and objective calibration using abundant observational data due to their non-differentiable nature.

This work follows a new paradigm in vegetation modeling that integrates the robustness of traditional models with the adaptive power of machine learning techniques. The goal is to combine reliable physical components with machine learning components. As opposed to classical vegetation models, the resulting hybrid model is differentiable and the parameters of both the physical and the neural network components can be optimized jointly and efficiently using observational data.

In the proposed hybrid vegetation model, machine learning can be used to improve the computational efficiency of the model by emulating computationally expensive routines. We have implemented the key processes related to photosynthesis in LPJ in Julia. This minimal model setup is used to explore the potential of machine learning to replace the computationally expensive root-finding algorithm used in computing the optimal ratio of intercellular to ambient CO2 concentration, and hence stomatal conductance.

Machine learning can also be used for better process representation. The recently developed neural or universal differential equations offer a particularly promising methodological framework for learning the dynamics of carbon allocation to different vegetation pools using observations. The dynamics of carbon allocation to different plant components can be effectively modeled using a neural ODE approach, which utilizes observations of observable variables (e.g., Above Ground Biomass (AGB), Leaf Area Index (LAI)) to learn the dynamics of unobservable variables such as vegetation carbon pools.

How to cite: Badri, M., Hess, P., Lin, Y., Bathiany, S., Gelbrecht, M., and Boers, N.: Towards a Hybrid Vegetation Model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17958, https://doi.org/10.5194/egusphere-egu25-17958, 2025.

X5.139
|
EGU25-18139
|
ECS
Darrell Tang and Xiaomei Yang

Modelling microplastic transport through porous media, such as soils and aquifers, is an emerging research topic, where existing hydrogeological models for (reactive) solute and colloid transport have shown limited effectiveness thus far. This perspective article draws upon recent literature to provide a brief overview of key microplastic transport processes, with emphases on less well-understood processes, to propose potential research directions for efficiently modeling microplastic transport through the porous environment. Microplastics are particulate matter with distinct physicochemical properties. Biogeochemical processes and physical interactions with the surrounding environment cause microplastic properties such as material density, geometry, chemical composition, and DLVO interaction parameters to change dynamically, through complex webs of interactions and feedbacks that dynamically affect transport behavior. Furthermore, microplastic material densities, which cluster around that of water, distinguish microplastics from other colloids, with impactful consequences that are often underappreciated. For example, (near-)neutral material densities cause microplastic transport behavior to be highly sensitive to spatio-temporally varying environmental conditions. The dynamic nature of microplastic properties implies that at environmentally relevant large spatio-temporal scales, the complex transport behavior may be effectively intractable to direct physical modeling. Therefore, efficient modeling may require integrating reduced-complexity physics-constrained models, with stochastic or statistical analyses, supported by extensive environmental data. This is a sub-project (focusing on microplastics in the environment) of the Digital Waters Flagship funded by the Research Council of Finland, where we aim to create a digital ecosystem for machine learning aided hydrological modelling of various hydrosphere processes across all environmental compartments, focusing particularly on the critical zone.

How to cite: Tang, D. and Yang, X.: Modeling microplastic transport through porous media: challenges arising from dynamic transport behavior, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18139, https://doi.org/10.5194/egusphere-egu25-18139, 2025.

X5.140
|
EGU25-18810
|
ECS
|
|
Kyrillos Ghattas and Tamás Buday

Permeability should be dispersed conveniently to control the aquifer's type and quality. Permeability in a variety of porous media can be determined using different methods depending on the environment and the scope of the porosity media. These days, permeability of core samples and well logging data with greater aquifer heterogeneity, artificial intelligence algorithms are well-known for estimating permeability. Machine learning and artificial intelligence have gained popularity and credibility across all scientific fields. To address the dearth of resources in geosciences generally and hydrology specifically.

As soft computing techniques, Artificial Neural Networks (ANNs) have demonstrated the capacity to estimate acceptable outputs with tolerable outcomes. The ANN model uses basic processing units, which are networks of interconnected neurons. The simplest approach is the Feed-Forward Artificial Neural Network (FF-ANN). The Middle Jurassic Hugin Formation may have been deposited as a mouth bar setting during the period of general transgression, as evidenced by fluctuating permeability values brought on by changes in the sediment supply, which varying porosity values brought on by variations in the amount of clay and size of grains.

Keywords: Artificial Neural Network, Feed-Forward Artificial Neural Network, Volve oilfield, Hugin Formation, Permeability estimation.

How to cite: Ghattas, K. and Buday, T.: Artificial Neural Network Approaches for Permeability Estimation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18810, https://doi.org/10.5194/egusphere-egu25-18810, 2025.

X5.141
|
EGU25-19319
|
ECS
Peter Ukkonen, Laura Mansfield, and Hannah Christensen

Machine learning (ML) has the potential to reduce systematic uncertainties in Earth System Models by replacing or complementing existing physics-based parameterizations of sub-grid processes. However, after decades of research, ensuring generalization and stability of ML-based parameterizations remains a major challenge.  We aim to minimize both epistemic and aleatoric sources of uncertainty via physically inspired, vertically recurrent neural networks (RNN) which offer key benefits such as parametric sparsity and efficient modeling of non-locality in a column. To address aleatoric uncertainty, we furthermore incorporate stochasticity and convective memory into the ML architecture. We present preliminary results using the ClimSim framework, where the physically inspired ML framework replaces a superparameterization in a low-resolution climate model.

How to cite: Ukkonen, P., Mansfield, L., and Christensen, H.: Emulation of sub-grid physics using stochastic, vertically recurrent neural networks, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19319, https://doi.org/10.5194/egusphere-egu25-19319, 2025.

Posters virtual: Fri, 2 May, 14:00–15:45 | vPoster spot 2

Display time: Fri, 2 May, 08:30–18:00
Chairperson: Viktor J. Bruckman

EGU25-7566 | Posters virtual | VPS30

Machine Learning-Based Prediction of Extreme Temperature Events in Texas: Understanding the Role of Large-Scale Climate Modes 

Andy Chen and Jian Zhao
Fri, 02 May, 14:00–15:45 (CEST)   vPoster spot 2 | vP2.8

Extreme warming events in Texas have far-reaching environmental, economic, and societal consequences, including impacts on agriculture, energy demand, public health, and infrastructure. These events underscore the urgent need for reliable prediction systems that can anticipate their occurrence and inform mitigation and adaptation strategies. In this study, we develop machine-learning-based models to predict extreme temperature events across Texas by identifying and modeling the key drivers of these phenomena. The predictive framework incorporates the influences of large-scale climate modes and processes from both the Pacific and North Atlantic Oceans, including the El Niño–Southern Oscillation (ENSO), Pacific Decadal Oscillation (PDO), Warm Pool (WP), and Atlantic Multidecadal Oscillation (AMO). By integrating these climate indices with regional atmospheric and surface data, the model captures the complex interactions between large-scale climate variability and regional temperature extremes. The contributions of each climate mode are quantified and analyzed to determine their relative importance in driving warming events across different temporal and spatial scales. To ensure the robustness of the predictions, the model outputs are further validated against physical mechanisms linking large-scale climate modes to atmospheric circulation patterns. This validation process provides a mechanistic understanding of the statistical relationships uncovered by the machine-learning models, ensuring that the predictions align with established climate dynamics. The findings from this study enhance our understanding of regional climate dynamics in Texas and demonstrate the potential of machine-learning approaches for improving the predictability of extreme temperature events.

How to cite: Chen, A. and Zhao, J.: Machine Learning-Based Prediction of Extreme Temperature Events in Texas: Understanding the Role of Large-Scale Climate Modes, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7566, https://doi.org/10.5194/egusphere-egu25-7566, 2025.