skip to main content


Title: Dynamics Concentration of Large-Scale Tightly-Connected Networks
The ability to achieve coordinated behavior --engineered or emergent-- on networked systems has attracted widespread interest over several fields. This has led to remarkable advances on the development of a theoretical understanding of the conditions under which agents within a network can reach agreement (consensus) or develop coordinated behaviors such as synchronization. However, fewer advances have been made toward explaining another commonly observed phenomena in tightly-connected networks systems: output responses of nodes in the networks are almost identical to each other despite heterogeneity in their individual dynamics. In this work, we leverage tools from high-dimensional probability to provide an initial answer to this phenomena. More precisely, we show that for linear networks of nodal random transfer functions, as the networks size and connectivity grows, every node in the network follows the same response to an input or disturbance irrespectively of the source of this input. We term this behavior as dynamics concentration as it stems from the fact that the network transfer matrix uniformly converges in probability to a unique dynamic response --i.e., it concentrates-- determined by the distribution of the random transfer function of each node. We further discuss the implications of our analysis in the context of model reduction and robustness, and provide numerical evidence that similar phenomena occur in small deterministic networks over a properly defined frequency band.  more » « less
Award ID(s):
1752362 1736448 1711188
NSF-PAR ID:
10132912
Author(s) / Creator(s):
;
Date Published:
Journal Name:
Conference on Decision and Control
Page Range / eLocation ID:
758 to 763
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract Despite the advances in discovering new nuclei, modeling microscopic nuclear structure, nuclear reactors, and stellar nucleosynthesis, we still lack a systemic tool, such as a network approach, to understand the structure and dynamics of over 70 thousands reactions compiled in JINA REACLIB. To this end, we develop an analysis framework, under which it is simple to know which reactions generally are possible and which are not, by counting neutrons and protons incoming to and outgoing from any target nucleus. Specifically, we assemble here a nuclear reaction network in which a node represents a nuclide, and a link represents a direct reaction between nuclides. Interestingly, the degree distribution of nuclear network exhibits a bimodal distribution that significantly deviates from the common power-law distribution of scale-free networks and Poisson distribution of random networks. Based on the dynamics from the cross section parameterizations in REACLIB, we surprisingly find that the distribution is universal for reactions with a rate below the threshold, λ < e − T γ , where T is the temperature and γ ≈ 1.05. Moreover, we discover three rules that govern the structure pattern of nuclear reaction network: (i) reaction-type is determined by linking choices, (ii) network distances between the reacting nuclides on 2D grid of Z vs N of nuclides are short, and (iii) each node in- and out-degrees are close to each other. By incorporating these three rules, our model universally unveils the underlying nuclear reaction patterns hidden in a large and dense nuclear reaction network regardless of nuclide chart expansions. It enables us to predict missing links that represent possible new nuclear reactions not yet discovered. 
    more » « less
  2. Fu, Feng (Ed.)
    Network science has increasingly become central to the field of epidemiology and our ability to respond to infectious disease threats. However, many networks derived from modern datasets are not just large, but dense, with a high ratio of edges to nodes. This includes human mobility networks where most locations have a large number of links to many other locations. Simulating large-scale epidemics requires substantial computational resources and in many cases is practically infeasible. One way to reduce the computational cost of simulating epidemics on these networks is sparsification , where a representative subset of edges is selected based on some measure of their importance. We test several sparsification strategies, ranging from naive thresholding to random sampling of edges, on mobility data from the U.S. Following recent work in computer science, we find that the most accurate approach uses the effective resistances of edges, which prioritizes edges that are the only efficient way to travel between their endpoints. The resulting sparse network preserves many aspects of the behavior of an SIR model, including both global quantities, like the epidemic size, and local details of stochastic events, including the probability each node becomes infected and its distribution of arrival times. This holds even when the sparse network preserves fewer than 10% of the edges of the original network. In addition to its practical utility, this method helps illuminate which links of a weighted, undirected network are most important to disease spread. 
    more » « less
  3. Abstract. Advances in ambient environmental monitoring technologies are enabling concerned communities and citizens to collect data to better understand their local environment and potential exposures. These mobile, low-cost tools make it possible to collect data with increased temporal and spatial resolution, providing data on a large scale with unprecedented levels of detail. This type of data has the potential to empower people to make personal decisions about their exposure and support the development of local strategies for reducing pollution and improving health outcomes. However, calibration of these low-cost instruments has been a challenge. Often, a sensor package is calibrated via field calibration. This involves colocating the sensor package with a high-quality reference instrument for an extended period and then applying machine learning or other model fitting technique such as multiple linear regression to develop a calibration model for converting raw sensor signals to pollutant concentrations. Although this method helps to correct for the effects of ambient conditions (e.g., temperature) and cross sensitivities with nontarget pollutants, there is a growing body of evidence that calibration models can overfit to a given location or set of environmental conditions on account of the incidental correlation between pollutant levels and environmental conditions, including diurnal cycles. As a result, a sensor package trained at a field site may provide less reliable data when moved, or transferred, to a different location. This is a potential concern for applications seeking to perform monitoring away from regulatory monitoring sites, such as personal mobile monitoring or high-resolution monitoring of a neighborhood. We performed experiments confirming that transferability is indeed a problem and show that it can be improved by collecting data from multiple regulatory sites and building a calibration model that leverages data from a more diverse data set. We deployed three sensor packages to each of three sites with reference monitors (nine packages total) and then rotated the sensor packages through the sites over time. Two sites were in San Diego, CA, with a third outside of Bakersfield, CA, offering varying environmental conditions, general air quality composition, and pollutant concentrations. When compared to prior single-site calibration, the multisite approach exhibits better model transferability for a range of modeling approaches. Our experiments also reveal that random forest is especially prone to overfitting and confirm prior results that transfer is a significant source of both bias and standard error. Linear regression, on the other hand, although it exhibits relatively high error, does not degrade much in transfer. Bias dominated in our experiments, suggesting that transferability might be easily increased by detecting and correcting for bias. Also, given that many monitoring applications involve the deployment of many sensor packages based on the same sensing technology, there is an opportunity to leverage the availability of multiple sensors at multiple sites during calibration to lower the cost of training and better tolerate transfer. We contribute a new neural network architecture model termed split-NN that splits the model into two stages, in which the first stage corrects for sensor-to-sensor variation and the second stage uses the combined data of all the sensors to build a model for a single sensor package. The split-NN modeling approach outperforms multiple linear regression, traditional two- and four-layer neural networks, and random forest models. Depending on the training configuration, compared to random forest the split-NN method reduced error 0 %–11 % for NO2 and 6 %–13 % for O3. 
    more » « less
  4. Abstract

    With the rapid development of metro systems, it has become increasingly important to study phenomena such as passenger flow distribution and passenger boarding behavior. It is difficult for existing methods to accurately describe actual situations and to extend to the whole metro system due to the limitations from parameter uncertainties in their mathematical models. In this article, we propose a passenger‐to‐train assignment model to evaluate the probabilities of individual passengers boarding each feasible train for both no‐transfer and one‐transfer situations. This model can be used to understand passenger flows and crowdedness. The input parameters of the model include the probabilities that the passengers take each train and the probability distribution of egress time, which is the time to walk to the tap‐out fare gate after alighting from the train. We present the likelihood method to estimate these parameters based on data from the automatic fare collection and automatic vehicle location systems. This method can construct several nonparametric density estimates without assuming the parametric form of the distribution of egress time. The EM algorithm is used to compute the maximum likelihood estimates. Simulation results indicate that the proposed estimates perform well. By applying our method to real data in Beijing metro system, we can identify different passenger flow patterns between peak and off‐peak hours.

     
    more » « less
  5. Over the last twenty years advances in systems biology have changed our views on microbial communities and promise to revolutionize treatment of human diseases. In almost all scientific breakthroughs since time of Newton, mathematical modeling has played a prominent role. Regulatory networks emerged as preferred descriptors of how abundances of molecular species depend on each other. However, the central question on how cellular phenotypes emerge from dynamics of these network remains elusive. The principal reason is that differential equation models in the field of biology (while so successful in areas of physics and physical chemistry), do not arise from first principles, and these models suffer from lack of proper parameterization. In response to these challenges, discrete time models based on Boolean networks have been developed. In this review, we discuss an emerging modeling paradigm that combines ideas from differential equations and Boolean models, and has been developed independently within dynamical systems and computer science communities. The result is an approach that can associate a range of potential dynamical behaviors to a network, arrange the descriptors of the dynamics in a searchable database, and allows for multi-parameter exploration of the dynamics akin to bifurcation theory. Since this approach is computationally accessible for moderately sized networks, it allows, perhaps for the first time, to rationally compare different network topologies based on their dynamics. 
    more » « less