skip to main content


Title: DASH: a MATLAB toolbox for paleoclimate data assimilation

Abstract. Paleoclimate data assimilation (DA) is a tool for reconstructing past climates that directly integrates proxy records with climate model output. Despite the potential for DA to expand the scope of quantitative paleoclimatology, these methods remain difficult to implement in practice due to the multi-faceted requirements and data handling necessary for DA reconstructions, the diversity of DA methods, and the need for computationally efficient algorithms. Here, we present DASH, a MATLAB toolbox designed to facilitate paleoclimate DA analyses. DASH provides command line and scripting tools that implement common tasks in DA workflows. The toolbox is highly modular and is not built around any specific analysis, and thus DASH supports paleoclimate DA for a wide variety of time periods, spatial regions, proxy networks, and algorithms. DASH includes tools for integrating and cataloguing data stored in disparate formats, building state vector ensembles, and running proxy (system) forward models. The toolbox also provides optimized algorithms for implementing ensemble Kalman filters, particle filters, and optimal sensor analyses with variable and modular parameters. This paper reviews the key components of the DASH toolbox and presents examples illustrating DASH's use for paleoclimate DA applications.

 
more » « less
Award ID(s):
1803946
PAR ID:
10510130
Author(s) / Creator(s):
; ; ; ;
Publisher / Repository:
Copernicus Publications on behalf of the European Geosciences Union
Date Published:
Journal Name:
Geoscientific Model Development
Volume:
16
Issue:
19
ISSN:
1991-9603
Page Range / eLocation ID:
5653 to 5683
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract

    Paleoclimate reconstructions are increasingly central to climate assessments, placing recent and future variability in a broader historical context. Paleoclimate reconstructions are increasingly central to climate assessments, placing recent and future variability in a broader historical context. Several estimation methods produce plumes of climate trajectories that practitioners often want to compare to other reconstruction ensembles, or to deterministic trajectories produced by other means, such as global climate models. Of particular interest are “offline” data assimilation (DA) methods, which have recently been adapted to paleoclimatology. Offline DA lacks an explicit model connecting time instants, so its ensemble members are not true system trajectories. This obscures quantitative comparisons, particularly when considering the ensemble mean in isolation. We propose several resampling methods to introduce a priori constraints on temporal behavior, as well as a general notion, called plume distance, to carry out quantitative comparisons between collections of climate trajectories (“plumes”). The plume distance provides a norm in the same physical units as the variable of interest (e.g. °C for temperature), and lends itself to assessments of statistical significance. We apply these tools to four paleoclimate comparisons: (1) global mean surface temperature (GMST) in the online and offline versions of the Last Millennium Reanalysis (v2.1); (2) GMST from these two ensembles to simulations of the Paleoclimate Model Intercomparison Project past1000 ensemble; (3) LMRv2.1 to the PAGES 2k (2019) ensemble of GMST and (4) northern hemisphere mean surface temperature from LMR v2.1 to the Büntgen et al. (2021) ensemble. Results generally show more compatibility between these ensembles than is visually apparent. The proposed methodology is implemented in an open-source Python package, and we discuss possible applications of the plume distance framework beyond paleoclimatology.

     
    more » « less
  2. In our increasingly data-driven society, it is critical for high school students to learn to integrate computational thinking with other disciplines in solving real world problems. To address this need for the life sciences in particular, we have developed the Bio-CS Bridge, a modular computational system coupled with curriculum integrating biology and computer science. Our transdisciplinary team comprises university and high school faculty and students with expertise in biology, computer

    science, and education.

    Our approach engages students and teachers in scientific practices using biological data that they can collect themselves, and computational tools that they help to design and implement, to address the real-world problem of

    pollinator decline.

    Our modular approach to high school curriculum design provides

    teachers with the educational flexibility to address national and statewide biology and computer science standards for a wide range of learner types. We are using a teacher- leader model to disseminate the Bio-CS Bridge, whose components will be freely

    available online.

     
    more » « less
  3. Abstract. Climate field reconstruction (CFR) refers to the estimation of spatiotemporal climate fields (such as surface temperature) from a collection of pointwise paleoclimate proxy datasets. Such reconstructions can provide rich information on climate dynamics and provide an out-of-sample validation of climate models. However, most CFR workflows are complex and time-consuming, as they involve (i) preprocessing of the proxy records, climate model simulations, and instrumental observations; (ii) application of one or more statistical methods; and (iii) analysis and visualization of the reconstruction results. Historically, this process has lacked transparency and accessibility, limiting reproducibility and experimentation by non-specialists. This article presents an open-source and object-oriented Python package called cfr that aims to make CFR workflows easy to understand and conduct, saving climatologists from technical details and facilitating efficient and reproducible research. cfr provides user-friendly utilities for common CFR tasks such as proxy and climate data analysis and visualization, proxy system modeling, and modularized workflows for multiple reconstruction methods, enabling methodological intercomparisons within the same framework. The package is supported with extensive documentation of the application programming interface (API) and a growing number of tutorial notebooks illustrating its usage. As an example, we present two cfr-driven reconstruction experiments using the PAGES 2k temperature database applying the last millennium reanalysis (LMR) paleoclimate data assimilation (PDA) framework and the graphical expectation–maximization (GraphEM) algorithm, respectively. 
    more » « less
  4. null (Ed.)
    This paper describes the derivation, analysis and implementation of a new data association method that provides a tight bound on the risk of incorrect association for LiDAR feature-based localization. Data association (DA) is the process of assigning currently-sensed features with ones that were previously observed. Most DA methods use a nearest-neighbor criterion based on the normalized innovation squared (NIS). They require complex algorithms to evaluate the risk of incorrect association because sensor state prediction, prior observations, and current measurements are uncertain. In contrast, in this work, we derive a new DA criterion using projections of the extended Kalman filter's innovation vector. The paper shows that innovation projections (IP) are signed quantities that not only capture the impact of an incorrect association in terms of its magnitude, but also of its direction. The IP-based DA criterion also leverages the fact that incorrect associations are known and well-defined fault modes. Thus, as compared to NIS, IPs provide a much tighter bound on the predicted risk of incorrect association. We analyze and evaluate the new IP method using simulated and experimental data for autonomous inertial-aided LiDAR localization in a structured lab environment. 
    more » « less
  5. null (Ed.)
    Video streaming commonly uses Dynamic Adaptive Streaming over HTTP (DASH) to deliver good Quality of Experience (QoE) to users. Videos used in DASH are predominantly encoded by single-layered video coding such as H.264/AVC. In comparison, multi-layered video coding such as H.264/SVC provides more flexibility for up- grading the quality of buffered video segments and has the potential to further improve QoE. However, there are two challenges for us- ing SVC in DASH: (i) the complexity in designing ABR algorithms; and (ii) the negative impact of SVC’s coding overhead. In this work, we propose a deep reinforcement learning method called Grad for designing ABR algorithms that take advantage of the quality up- grade mechanism of SVC. Additionally, we quantify the impact of coding overhead on the achievable QoE of SVC in DASH, and propose jump-enabled hybrid coding (HYBJ) to mitigate the impact. Through emulation, we demonstrate that Grad-HYBJ, an ABR algo- rithm for HYBJ learned by Grad, outperforms the best performing state-of-the-art ABR algorithm by 17% in QoE. 
    more » « less