skip to main content

Title: Data‐driven Evolution Equation Reconstruction for Parameter‐Dependent Nonlinear Dynamical Systems

When studying observations of chemical reaction dynamics, closed form equations based on a putative mechanism may not be available. Yet when sufficient data from experimental observations can be obtained, even without knowing what exactly the physical meaning of the parameter settings or recorded variables are, data‐driven methods can be used to construct minimal (and in a sense, robust) realizations of the system. The approach attempts, in a sense, to circumvent physical understanding, by building intrinsic “information geometries” of the observed data, and thus enabling prediction without physical/chemical knowledge. Here we use such an approach to obtain evolution equationsfor a data‐driven realization of the original system– in effect, allowing prediction based on the informed interrogation of the agnostically organized observation database. We illustrate the approach on observations of (a) the normal form for the cusp singularity, (b) a cusp singularity for the nonisothermal CSTR, and (c) a random invertible transformation of the nonisothermal CSTR, showing that one can predict even when the observables are not “simply explainable” physical quantities. We discuss current limitations and possible extensions of the procedure.

more » « less
Author(s) / Creator(s):
 ;  ;  ;  
Publisher / Repository:
Wiley Blackwell (John Wiley & Sons)
Date Published:
Journal Name:
Israel Journal of Chemistry
Page Range / eLocation ID:
p. 787-794
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract

    Monod and Logistic growth models have been widely used as basic equations to describe cell growth in bioprocess engineering. In the case of the Monod equation, the specific growth rate is governed by a limiting nutrient, with the mathematical form similar to the Michaelis–Menten equation. In the case of the Logistic equation, the specific growth rate is determined by the carrying capacity of the system, which could be growth‐inhibiting factors (i.e., toxic chemical accumulation) other than the nutrient level. Both equations have been found valuable to guide us build unstructured kinetic models to analyze the fermentation process and understand cell physiology. In this work, we present a hybrid Logistic‐Monod growth model, which accounts for multiple growth‐dependent factors including both the limiting nutrient and the carrying capacity of the system. Coupled with substrate consumption and yield coefficient, we present the analytical solutions for this hybrid Logistic‐Monod model in both batch and continuous stirred tank reactor (CSTR) culture. Under high biomass yield (Yx/s) conditions, the analytical solution for this hybrid model is approaching to the Logistic equation; under low biomass yield condition, the analytical solution for this hybrid model converges to the Monod equation. This hybrid Logistic‐Monod equation represents the cell growth transition from substrate‐limiting condition to growth‐inhibiting condition, which could be adopted to accurately describe the multi‐phases of cell growth and may facilitate kinetic model construction, bioprocess optimization, and scale‐up in industrial biotechnology.

    more » « less
  2. Data from the cellular network have been proved as one of the most promising way to understand large-scale human mobility for various ubiquitous computing applications due to the high penetration of cellphones and low collection cost. Existing mobility models driven by cellular network data suffer from sparse spatial-temporal observations because user locations are recorded with cellphone activities, e.g., calls, text, or internet access. In this paper, we design a human mobility recovery system called CellSense to take the sparse cellular billing data (CBR) as input and outputs dense continuous records to recover the sensing gap when using cellular networks as sensing systems to sense the human mobility. There is limited work on this kind of recovery systems at large scale because even though it is straightforward to design a recovery system based on regression models, it is very challenging to evaluate these models at large scale due to the lack of the ground truth data. In this paper, we explore a new opportunity based on the upgrade of cellular infrastructures to obtain cellular network signaling data as the ground truth data, which log the interaction between cellphones and cellular towers at signal levels (e.g., attaching, detaching, paging) even without billable activities. Based on the signaling data, we design a system CellSense for human mobility recovery by integrating collective mobility patterns with individual mobility modeling, which achieves the 35.3% improvement over the state-of-the-art models. The key application of our recovery model is to take regular sparse CBR data that a researcher already has, and to recover the missing data due to sensing gaps of CBR data to produce a dense cellular data for them to train a machine learning model for their use cases, e.g., next location prediction. 
    more » « less
  3. Abstract

    “Supermodeling” climate by allowing different models to assimilate data from one another in run time has been shown to give results superior to those of any one model and superior to any weighted average of model outputs. The only free parameters, connection strengths between corresponding variables in each pair of models, are determined using some form of machine learning. It is demonstrated that supermodeling succeeds because near critical states, interscale interactions are important but unresolved processes cannot be effectively represented diagnostically in any single parameterization scheme. In two examples, a pair of toy quasigeostrophic (QG) channel models of the midlatitudes and a pair of ECHAM5 models of the tropical Pacific atmosphere with a common ocean, supermodels dynamically combine parameterization schemes so as to capture criticality, associated critical structures, and the supporting scale interactions. The QG supermodeling scheme extends a previous configuration in which two such models synchronize with intermodel connections only between medium-scale components of the flow; here the connections are trained against a third “real” model. Intermittent blocking patterns characterize the critical behavior thus obtained, even where such patterns are missing in the constituent models. In the ECHAM-based climate supermodel, the corresponding critical structure is the single ITCZ pattern, a pattern that occurs in neither of the constituent models. For supermodels of both types, power spectra indicate enhanced interscale interactions in frequency or energy ranges of physical interest, in agreement with observed data, and supporting a generalized form of the self-organized criticality hypothesis.

    Significance Statement

    In a “supermodel” of Earth’s climate, alternative models (climate simulations), which differ in the way they represent processes on the smallest scales, are trained to exchange information as they run, adjusting to one another much as weather prediction models adjust to new observations. They form a consensus, capturing atmospheric behaviors that have eluded all the separate models. We demonstrate that simplified supermodels succeed, where no single approach can, by correctly representingcritical phenomenainvolving sudden qualitative transitions, such as occur in El Niño events, that depend on interactions among atmospheric processes on many different scales in space and time. The correct reproduction of critical phenomena is vital both for predicting weather and for projecting the effects of climate change.

    more » « less
  4. Abstract

    Predictions of hydrologic variables across the entire water cycle have significant value for water resources management as well as downstream applications such as ecosystem and water quality modeling. Recently, purely data‐driven deep learning models like long short‐term memory (LSTM) showed seemingly insurmountable performance in modeling rainfall runoff and other geoscientific variables, yet they cannot predict untrained physical variables and remain challenging to interpret. Here, we show that differentiable, learnable, process‐based models (calledδmodels here) can approach the performance level of LSTM for the intensively observed variable (streamflow) with regionalized parameterization. We use a simple hydrologic model HBV as the backbone and use embedded neural networks, which can only be trained in a differentiable programming framework, to parameterize, enhance, or replace the process‐based model's modules. Without using an ensemble or post‐processor,δmodels can obtain a median Nash‐Sutcliffe efficiency of 0.732 for 671 basins across the USA for the Daymet forcing data set, compared to 0.748 from a state‐of‐the‐art LSTM model with the same setup. For another forcing data set, the difference is even smaller: 0.715 versus 0.722. Meanwhile, the resulting learnable process‐based models can output a full set of untrained variables, for example, soil and groundwater storage, snowpack, evapotranspiration, and baseflow, and can later be constrained by their observations. Both simulated evapotranspiration and fraction of discharge from baseflow agreed decently with alternative estimates. The general framework can work with models with various process complexity and opens up the path for learning physics from big data.

    more » « less

    A key initial step in geophysical imaging is to devise an effective means of mapping the sensitivity of an observation to the model parameters, that is to compute its Fréchet derivatives or sensitivity kernel. In the absence of any simplifying assumptions and when faced with a large number of free parameters, the adjoint method can be an effective and efficient approach to calculating Fréchet derivatives and requires just two numerical simulations. In the Glacial Isostatic Adjustment problem, these consist of a forward simulation driven by changes in ice mass and an adjoint simulation driven by fictitious loads that are applied at the observation sites. The theoretical basis for this approach has seen considerable development over the last decade. Here, we present the final elements needed to image 3-D mantle viscosity using a dataset of palaeo sea-level observations. Developments include the calculation of viscosity Fréchet derivatives (i.e. sensitivity kernels) for relative sea-level observations, a modification to the numerical implementation of the forward and adjoint problem that permits application to 3-D viscosity structure, and a recalibration of initial sea level that ensures the forward simulation honours present-day topography. In the process of addressing these items, we build intuition concerning how absolute sea-level and relative sea-level observations sense Earth’s viscosity structure and the physical processes involved. We discuss examples for potential observations located in the near field (Andenes, Norway), far field (Seychelles), and edge of the forebulge of the Laurentide ice sheet (Barbados). Examination of these kernels: (1) reveals why 1-D estimates of mantle viscosity from far-field relative sea-level observations can be biased; (2) hints at why an appropriate differential relative sea-level observation can provide a better constraint on local mantle viscosity and (3) demonstrates that sea-level observations have non-negligible 3-D sensitivity to deep mantle viscosity structure, which is counter to the intuition gained from 1-D radial viscosity Fréchet derivatives. Finally, we explore the influence of lateral variations in viscosity on relative sea-level observations in the Amundsen Sea Embayment and at Barbados. These predictions are based on a new global 3-D viscosity inference derived from the shear-wave speeds of GLAD-M25 and an inverse calibration scheme that ensures compatibility with certain fundamental geophysical observations. Use of the 3-D viscosity inference leads to: (1) generally greater complexity within the kernel; (2) an increase in sensitivity and presence of shorter length-scale features within lower viscosity regions; (3) a zeroing out of the sensitivity kernel within high-viscosity regions where elastic deformation dominates and (4) shifting of sensitivity at a given depth towards distal regions of weaker viscosity. The tools and intuition built here provide the necessary framework to explore inversions for 3-D mantle viscosity based on palaeo sea-level data.

    more » « less