Abstract Subgrid‐scale processes, such as atmospheric gravity waves (GWs), play a pivotal role in shaping the Earth's climate but cannot be explicitly resolved in climate models due to limitations on resolution. Instead, subgrid‐scale parameterizations are used to capture their effects. Recently, machine learning (ML) has emerged as a promising approach to learn parameterizations. In this study, we explore uncertainties associated with a ML parameterization for atmospheric GWs. Focusing on the uncertainties in the training process (parametric uncertainty), we use an ensemble of neural networks to emulate an existing GW parameterization. We estimate both offline uncertainties in raw NN output and online uncertainties in climate model output, after the neural networks are coupled. We find that online parametric uncertainty contributes a significant source of uncertainty in climate model output that must be considered when introducing NN parameterizations. This uncertainty quantification provides valuable insights into the reliability and robustness of ML‐based GW parameterizations, thus advancing our understanding of their potential applications in climate modeling.
more »
« less
Finetuning AI Foundation Models to Develop Subgrid‐Scale Parameterizations: A Case Study on Atmospheric Gravity Waves
Abstract Global climate models parameterize a range of atmospheric‐oceanic processes, including gravity waves (GWs), clouds, moist convection, and turbulence, that cannot be sufficiently resolved. These subgrid‐scale closures for unresolved processes are a substantial source of model uncertainty. Here, we present a new approach to developing machine learning (ML) parameterizations of small‐scale climate processes by fine‐tuning a pre‐trained AI foundation model (FM). FMs are largely unexplored in climate research. A pre‐trained encoder‐decoder from a 2.3 billion parameter FM (NASA and IBM Research's Prithvi WxC)—which contains a latent probabilistic representation of atmospheric evolution—is fine‐tuned (or reused) to create a deep learning parameterization for atmospheric gravity waves (GWs); a process unseen during pre‐training. The parameterization captures GW effects for a coarse‐resolution climate model by learning the fluxes from an atmospheric reanalysis with 10 times finer resolution. A comparison of monthly averages and instantaneous evolution with a machine learning model baseline (an Attention U‐Net) reveals superior predictive performance of the FM parameterization throughout the atmosphere, even in regions excluded during pre‐training. This performance boost is quantified using the Hellinger distance, which is 0.11 for the baseline and 0.06 for the fine‐tuned model. Our findings emphasize the versatility and reusability of FMs, which could be used to accomplish a range of atmosphere‐ and climate‐related applications, leading the way for the creation of observations‐driven and physically accurate parameterizations for more earth system processes.
more »
« less
- Award ID(s):
- 2004492
- PAR ID:
- 10650062
- Publisher / Repository:
- DOI PREFIX: 10.1029
- Date Published:
- Journal Name:
- Journal of Advances in Modeling Earth Systems
- Volume:
- 17
- Issue:
- 11
- ISSN:
- 1942-2466
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Atmospheric gravity waves are produced when gravity attempts to restore disturbances through stable layers in the atmosphere. They have a visible effect on many atmospheric phenomena such as global circulation and air turbulence. Despite their importance, however, little research has been conducted on how to detect gravity waves using machine learning algorithms. We faced two major challenges in our research: our raw data had a lot of noise and the labeled dataset was extremely small. In this study, we explored various methods of preprocessing and transfer learning in order to address those challenges. We pre-trained an autoencoder on unlabeled data before training it to classify labeled data. We also created a custom CNN by combining certain pre-trained layers from the InceptionV3 Model trained on ImageNet with custom layers and a custom learning rate scheduler. Experiments show that our best model outperformed the best performing baseline model by 6.36% in terms of test accuracy.more » « less
-
Abstract Atmospheric gravity waves (GWs) span a broad range of length scales. As a result, the un‐resolved and under‐resolved GWs have to be represented using a sub‐grid scale (SGS) parameterization in general circulation models (GCMs). In recent years, machine learning (ML) techniques have emerged as novel methods for SGS modeling of climate processes. In the widely used approach of supervised (offline) learning, the true representation of the SGS terms have to be properly extracted from high‐fidelity data (e.g., GW‐resolving simulations). However, this is a non‐trivial task, and the quality of the ML‐based parameterization significantly hinges on the quality of these SGS terms. Here, we compare three methods to extract 3D GW fluxes and the resulting drag (Gravity Wave Drag [GWD]) from high‐resolution simulations: Helmholtz decomposition, and spatial filtering to compute the Reynolds stress and the full SGS stress. In addition to previous studies that focused only on vertical fluxes by GWs, we also quantify the SGS GWD due to lateral momentum fluxes. We build and utilize a library of tropical high‐resolution (Δx = 3 km) simulations using weather research and forecasting model. Results show that the SGS lateral momentum fluxes could have a significant contribution to the total GWD. Moreover, when estimating GWD due to lateral effects, interactions between the SGS and the resolved large‐scale flow need to be considered. The sensitivity of the results to different filter type and length scale (dependent on GCM resolution) is also explored to inform the scale‐awareness in the development of data‐driven parameterizations.more » « less
-
Abstract Gravity waves (GWs) and their associated multi‐scale dynamics are known to play fundamental roles in energy and momentum transport and deposition processes throughout the atmosphere. We describe an initial machine learning model—the Compressible Atmosphere Model Network (CAM‐Net). CAM‐Net is trained on high‐resolution simulations by the state‐of‐the‐art model Complex Geometry Compressible Atmosphere Model (CGCAM). Two initial applications to a Kelvin‐Helmholtz instability source and mountain wave generation, propagation, breaking, and Secondary GW (SGW) generation in two wind environments are described here. Results show that CAM‐Net can capture the key 2‐D dynamics modeled by CGCAM with high precision. Spectral characteristics of primary and SGWs estimated by CAM‐Net agree well with those from CGCAM. Our results show that CAM‐Net can achieve a several order‐of‐magnitude acceleration relative to CGCAM without sacrificing accuracy and suggests a potential for machine learning to enable efficient and accurate descriptions of primary and secondary GWs in global atmospheric models.more » « less
-
Abstract Gravity waves (GWs) make crucial contributions to the middle atmospheric circulation. Yet, their climate model representation remains inaccurate, leading to key circulation biases. This study introduces a set of three neural networks (NNs) that learn to predict GW fluxes (GWFs) from multiple years of high‐resolution ERA5 reanalysis. The three NNs: a ANN, a ANN‐CNN, and an Attention UNet embed different levels of horizontal nonlocality in their architecture and are capable of representing nonlocal GW effects that are missing from current operational GW parameterizations. The NNs are evaluated offline on both time‐averaged statistics and time‐evolving flux variability. All NNs, especially the Attention UNet, accurately recreate the global GWF distribution in both the troposphere and the stratosphere. Moreover, the Attention UNet most skillfully predicts the transient evolution of GWFs over prominent orographic and nonorographic hotspots, with the model being a close second. Since even ERA5 does not resolve a substantial portion of GWFs, this deficiency is compensated by subsequently applying transfer learning on the ERA5‐trained ML models for GWFs from a 1.4 km global climate model. It is found that the re‐trained models both (a) preserve their learning from ERA5, and (b) learn to appropriately scale the predicted fluxes to account for ERA5's limited resolution. Our results highlight the importance of embedding nonlocal information for a more accurate GWF prediction and establish strategies to complement abundant reanalysis data with limited high‐resolution data to develop machine learning‐driven parameterizations for missing mesoscale processes in climate models.more » « less
An official website of the United States government
