skip to main content


The NSF Public Access Repository (NSF-PAR) system and access will be unavailable from 11:00 PM ET on Friday, May 17 until 8:00 AM ET on Saturday, May 18 due to maintenance. We apologize for the inconvenience.

Title: Calibration and Uncertainty Quantification of Convective Parameters in an Idealized GCM

Parameters in climate models are usually calibrated manually, exploiting only small subsets of the available data. This precludes both optimal calibration and quantification of uncertainties. Traditional Bayesian calibration methods that allow uncertainty quantification are too expensive for climate models; they are also not robust in the presence of internal climate variability. For example, Markov chain Monte Carlo (MCMC) methods typically requiremodel runs and are sensitive to internal variability noise, rendering them infeasible for climate models. Here we demonstrate an approach to model calibration and uncertainty quantification that requires onlymodel runs and can accommodate internal climate variability. The approach consists of three stages: (a) a calibration stage uses variants of ensemble Kalman inversion to calibrate a model by minimizing mismatches between model and data statistics; (b) an emulation stage emulates the parameter‐to‐data map with Gaussian processes (GP), using the model runs in the calibration stage for training; (c) a sampling stage approximates the Bayesian posterior distributions by sampling the GP emulator with MCMC. We demonstrate the feasibility and computational efficiency of this calibrate‐emulate‐sample (CES) approach in a perfect‐model setting. Using an idealized general circulation model, we estimate parameters in a simple convection scheme from synthetic data generated with the model. The CES approach generates probability distributions of the parameters that are good approximations of the Bayesian posteriors, at a fraction of the computational cost usually required to obtain them. Sampling from this approximate posterior allows the generation of climate predictions with quantified parametric uncertainties.

more » « less
Award ID(s):
Author(s) / Creator(s):
 ;  ;  ;  
Publisher / Repository:
DOI PREFIX: 10.1029
Date Published:
Journal Name:
Journal of Advances in Modeling Earth Systems
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract

    Data required to calibrate uncertain general circulation model (GCM) parameterizations are often only available in limited regions or time periods, for example, observational data from field campaigns, or data generated in local high‐resolution simulations. This raises the question of where and when to acquire additional data to be maximally informative about parameterizations in a GCM. Here we construct a new ensemble‐based parallel algorithm to automatically target data acquisition to regions and times that maximize the uncertainty reduction, or information gain, about GCM parameters. The algorithm uses a Bayesian framework that exploits a quantified distribution of GCM parameters as a measure of uncertainty. This distribution is informed by time‐averaged climate statistics restricted to local regions and times. The algorithm is embedded in the recently developed calibrate‐emulate‐sample framework, which performs efficient model calibration and uncertainty quantification with onlymodel evaluations, compared withevaluations typically needed for traditional approaches to Bayesian calibration. We demonstrate the algorithm with an idealized GCM, with which we generate surrogates of local data. In this perfect‐model setting, we calibrate parameters and quantify uncertainties in a quasi‐equilibrium convection scheme in the GCM. We consider targeted data that are (a) localized in space for statistically stationary simulations, and (b) localized in space and time for seasonally varying simulations. In these proof‐of‐concept applications, the calculated information gain reflects the reduction in parametric uncertainty obtained from Bayesian inference when harnessing a targeted sample of data. The largest information gain typically, but not always, results from regions near the intertropical convergence zone.

    more » « less
  2. Abstract

    Many chemical processes depend non‐linearly on temperature. Gravity‐wave‐induced temperature perturbations have been shown to affect atmospheric chemistry, but accounting for this process in chemistry‐climate models has been a challenge because many gravity waves have scales smaller than the typical model resolution. Here, we present a method to account for subgrid‐scale orographic gravity‐wave‐induced temperature perturbations on the global scale for the Whole Atmosphere Community Climate Model. Temperature perturbation amplitudesconsistent with the model's subgrid‐scale gravity wave parameterization are derived and then used as a sinusoidal temperature perturbation in the model's chemistry solver. Because of limitations in the parameterization, we explore scaling ofbetween 0.6 and 1 based on comparisons to altitude‐dependentdistributions of satellite and reanalysis data, where we discuss uncertainties. We probe the impact on the chemistry from the grid‐point to global scales, and show that the parameterization is able to represent mountain wave events as reported by previous literature. The gravity waves for example, lead to increased surface area densities of stratospheric aerosols. This increases chlorine activation, with impacts on the associated chemical composition. We obtain large local changes in some chemical species (e.g., active chlorine, NOx, N2O5) which are likely to be important for comparisons to airborne or satellite observations, but the changes to ozone loss are more modest. This approach enables the chemistry‐climate modeling community to account for subgrid‐scale gravity wave temperature perturbations interactively, consistent with the internal parameterizations and are expected to yield more realistic interactions and better representation of the chemistry.

    more » « less
  3. Abstract

    The ocean mixed layer plays an important role in the coupling between the upper ocean and atmosphere across a wide range of time scales. Estimation of the variability of the ocean mixed layer is therefore important for atmosphere‐ocean prediction and analysis. The increasing coverage of in situ Argo profile data allows for an increasingly accurate analysis of the mixed layer depth (MLD) variability associated with deviations from the seasonal climatology. However, sampling rates are not sufficient to fully resolve subseasonal (day) MLD variability. Yet, many multivariate observations‐based analyses include implicit modeled subseasonal MLD variability. One analysis method is optimal interpolation of in situ data, but the interior analysis can be improved by leveraging surface data with regression or variational approaches. Here, we demonstrate how machine learning methods and satellite sea surface temperature, salinity, and height facilitate MLD estimation in a pilot study of two regions: the mid‐latitude southern Indian and the eastern equatorial Pacific Oceans. We construct multiple machine learning architectures to produce weekly 1/2° gridded MLD anomaly fields (relative to a monthly climatology) with uncertainty estimates. We test multiple traditional and probabilistic machine learning techniques to compare both accuracy and probabilistic calibration. We validate our methodology by applying it to ocean model simulations. We find that incorporating sea surface data through a machine learning model improves the performance of spatiotemporal MLD variability estimation compared to optimal interpolation of Argo observations alone. These preliminary results are a promising first step for the application of machine learning to MLD prediction.

    more » « less
  4. Abstract

    A reference or “no‐feedback” radiative response to warming is fundamental to understanding how much global warming will occur for a given change in greenhouse gases or solar radiation incident on the Earth. The simplest estimate of this radiative response is given by the Stefan‐Boltzmann law as W m−2 K−1for Earth's present climate, whereis a global effective emission temperature. The comparable radiative response in climate models, widely called the “Planck feedback,” averages −3.3 W m−2 K−1. This difference of 0.5 W m−2 K−1is large compared to the uncertainty in the net climate feedback, yet it has not been studied carefully. We use radiative transfer models to analyze these two radiative feedbacks to warming, and find that the difference arises primarily from the lack of stratospheric warming assumed in calculations of the Planck feedback (traditionally justified by differing constraints on and time scales of stratospheric adjustment relative to surface and tropospheric warming). The Planck feedback is thus masked for wavelengths with non‐negligible stratospheric opacity, and this effect implicitly acts to amplify warming in current feedback analysis of climate change. Other differences between Planck and Stefan‐Boltzmann feedbacks arise from temperature‐dependent gas opacities, and several artifacts of nonlinear averaging across wavelengths, heights, and different locations; these effects partly cancel but as a whole slightly destabilize the Planck feedback. Our results point to an important role played by stratospheric opacity in Earth's climate sensitivity, and clarify a long‐overlooked but notable gap in our understanding of Earth's reference radiative response to warming.

    more » « less
  5. Abstract

    The oxygen isotopic composition of planktic foraminiferal calcite () is one of the most prevalent proxies used in the paleoceanographic community. The relationship between, temperature, and seawater oxygen isotopic composition () is firmly rooted in thermodynamics, and experimental constraints are commonly used for sea surface temperature (SST) reconstructions. However, in marine sedimentary applications, additional sources of uncertainty emerge, and these uncertainty constraints have not as of yet been included in global calibration models. Here, we compile a global data set of over 2,600 marine sediment core top samples for five planktic species:Globigerinoides ruber,Trilobatus sacculifer,Globigerina bulloides,Neogloboquadrina incompta, andNeogloboquadrina pachyderma. We developed a suite of Bayesian regression models to calibrate the relationship betweenand SST. Spanning SSTs from 0.0 to 29.5 °C, our annual model with species pooled together has a mean standard error of approximately 0.54‰. Accounting for seasonality and species‐specific differences improves model validation, reducing the mean standard error to 0.47‰. Example applications spanning the Late Quaternary show good agreement with independent alkenone‐based estimates. Our pooled calibration model may also be used for reconstruction in the deeper geological past, using modern planktic foraminifera as an analog for non‐extant species. Our core top‐based models provide a robust assessment of uncertainty in thepaleothermometer that can be used in statistical assessments of interproxy and model‐proxy comparisons. The suite of models is publicly available as the Open Source software librarybayfox, for Python, R, and MATLAB/Octave.

    more » « less