skip to main content


Title: A Statistical Interpolation Code for Ocean Analysis and Forecasting
Abstract

We present a data assimilation package for use with ocean circulation models in analysis, forecasting, and system evaluation applications. The basic functionality of the package is centered on a multivariate linear statistical estimation for a given predicted/background ocean state, observations, and error statistics. Novel features of the package include support for multiple covariance models, and the solution of the least squares normal equations either using the covariance matrix or its inverse—the information matrix. The main focus of this paper, however, is on the solution of the analysis equations using the information matrix, which offers several advantages for solving large problems efficiently. Details of the parameterization of the inverse covariance using Markov random fields are provided and its relationship to finite-difference discretizations of diffusion equations are pointed out. The package can assimilate a variety of observation types from both remote sensing and in situ platforms. The performance of the data assimilation methodology implemented in the package is demonstrated with a yearlong global ocean hindcast with a 1/4° ocean model. The code is implemented in modern Fortran, supports distributed memory, shared memory, multicore architectures, and uses climate and forecasts compliant Network Common Data Form for input/output. The package is freely available with an open source license fromwww.tendral.com/tsis/.

 
more » « less
NSF-PAR ID:
10366493
Author(s) / Creator(s):
 ;  ;  ;  ;  
Publisher / Repository:
American Meteorological Society
Date Published:
Journal Name:
Journal of Atmospheric and Oceanic Technology
Volume:
39
Issue:
3
ISSN:
0739-0572
Page Range / eLocation ID:
p. 367-386
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract

    The Transiting Exoplanet Survey Satellite (TESS) has an exceptionally large plate scale of 21″ px−1, causing most TESS light curves to record the blended light of multiple stars. This creates a danger of misattributing variability observed by TESS to the wrong source, which would invalidate any analysis. We developed a method that can localize the origin of variability on the sky to better than one fifth of a pixel. Given measured frequencies of variability (e.g., from periodogram analysis), we show that the best-fit sinusoid amplitudes to raw light curves extracted from each pixel are distributed in the same way as light from the variable source. The primary assumption of this method is that other nearby stars are not variable at the same frequencies. Essentially, we are using the high frequency resolution of TESS to overcome limitations from its low spatial resolution. We have implemented our method in an open-source Python package,TESS_localize(github.com/Higgins00/TESS-Localize), that determines the location of a variable source on the sky and the most likely Gaia source given TESS pixel data and a set of observed frequencies of variability. Our method utilizes models of the TESS pixel response function, and we characterize systematics in the residuals of fitting these models to data. We find that even stars more than three pixels outside a photometric aperture can produce significant contaminant signals in the extracted light curves. Given the ubiquity of source blending in TESS light curves, verifying the source of observed variability should be a standard step in TESS analyses.

     
    more » « less
  2. null (Ed.)
    We develop online graph learning algorithms from streaming network data. Our goal is to track the (possibly) time-varying network topology, and affect memory and computational savings by processing the data on-the-fly as they are acquired. The setup entails observations modeled as stationary graph signals generated by local diffusion dynamics on the unknown network. Moreover, we may have a priori information on the presence or absence of a few edges as in the link prediction problem. The stationarity assumption implies that the observations’ covariance matrix and the so-called graph shift operator (GSO—a matrix encoding the graph topology) commute under mild requirements. This motivates formulating the topology inference task as an inverse problem, whereby one searches for a sparse GSO that is structurally admissible and approximately commutes with the observations’ empirical covariance matrix. For streaming data, said covariance can be updated recursively, and we show online proximal gradient iterations can be brought to bear to efficiently track the time-varying solution of the inverse problem with quantifiable guarantees. Specifically, we derive conditions under which the GSO recovery cost is strongly convex and use this property to prove that the online algorithm converges to within a neighborhood of the optimal time-varying batch solution. Numerical tests illustrate the effectiveness of the proposed graph learning approach in adapting to streaming information and tracking changes in the sought dynamic network. 
    more » « less
  3. null (Ed.)
    We leverage proximal gradient iterations to develop an online graph learning algorithm from streaming network data. Our goal is to track the (possibly) time-varying network topology, and effect memory and computational savings by processing the data on-the-fly as they are acquired. The setup entails observations modeled as stationary graph signals generated by local diffusion dynamics on the unknown network. Moreover, we may have a priori information on the presence or absence of a few edges as in the link prediction problem. The stationarity assumption implies that the observations' covariance matrix and the so-called graph shift operator (GSO - a matrix encoding the graph topology) commute under mild requirements. This motivates formulating the topology inference task as an inverse problem, whereby one searches for a (e.g., sparse) GSO that is structurally admissible and approximately commutes with the observations' empirical covariance matrix. For streaming data said covariance can be updated recursively, and we show online proximal gradient iterations can be brought to bear to efficiently track the time-varying solution of the inverse problem with quantifiable guarantees. Specifically, we derive conditions under which the GSO recovery cost is strongly convex and use this property to prove that the online algorithm converges to within a neighborhood of the optimal time-varying batch solution. Preliminary numerical tests illustrate the effectiveness of the proposed graph learning approach in adapting to streaming information and tracking changes in the sought dynamic network. 
    more » « less
  4. Abstract Background

    Low-depth sequencing allows researchers to increase sample size at the expense of lower accuracy. To incorporate uncertainties while maintaining statistical power, we introduce to analyze population structure of low-depth sequencing data.

    Results

    The method optimizes the choice of nonlinear transformations of dosages to maximize the Ky Fan norm of the covariance matrix. The transformation incorporates the uncertainty in calling between heterozygotes and the common homozygotes for loci having a rare allele and is more linear when both variants are common.

    Conclusions

    We apply to samples from two indigenous Siberian populations and reveal hidden population structure accurately using only a single chromosome. The package is available onhttps://github.com/yiwenstat/MCPCA_PopGen.

     
    more » « less
  5. Abstract

    Seven three‐component ocean bottom seismometers (OBS) of the Ocean Observatories Initiative (OOI) Cabled Array on top of Axial Seamount are continuously streaming data in real time to the Incorporated Research Institutions for Seismology (IRIS). The OBS array records earthquakes from the submarine volcano which last erupted on 24 April 2015, about 4 months after the array came online. The OBS data have proven crucial in providing insight into the volcano structure and dynamics (Wilcock et al., 2016,https://doi.org/10.1126/science.aah5563). We implemented a real‐time double‐difference (RT‐DD) monitoring system that automatically computes high‐precision (tens of meters) locations of new earthquakes. The system's underlying double‐difference base catalog includes nearly 100,000 earthquakes and was computed using kurtosis phase onset picks, cross‐correlation phase delay times, and 3‐DPandSvelocity models to predict the data. The relocations reveal the fine‐scale structures of long‐lived, narrow (<200 m wide), outward dipping, convex faults on the east and west walls of the caldera that appear to form a figure 8‐shaped ring fault system. These faults accommodate stresses caused by the inflation of magma prior to and deflation during eruptions. The east fault is segmented and pulled apart in east‐west direction due to its interaction with the Juan de Fuca Ridge, which at this location forms an overlapping spreading center. The RT‐DD system enables the monitoring and rapid analysis of variations in fine‐scale seismic and fault properties and has the potential to improve prediction of timing and location of the next Axial eruption expected to occur in the 2022–2023 time frame.

     
    more » « less