Title: Nonlinear System Identification via Tensor Completion
Function approximation from input and output data pairs constitutes a fundamental problem in supervised learning. Deep neural networks are currently the most popular method for learning to mimic the input-output relationship of a general nonlinear system, as they have proven to be very effective in approximating complex highly nonlinear functions. In this work, we show that identifying a general nonlinear function y = ƒ(x1,…,xN) from input-output examples can be formulated as a tensor completion problem and under certain conditions provably correct nonlinear system identification is possible. Specifically, we model the interactions between the N input variables and the scalar output of a system by a single N-way tensor, and setup a weighted low-rank tensor completion problem with smoothness regularization which we tackle using a block coordinate descent algorithm. We extend our method to the multi-output setting and the case of partially observed data, which cannot be readily handled by neural networks. Finally, we demonstrate the effectiveness of the approach using several regression tasks including some standard benchmarks and a challenging student grade prediction task. more »« less
Kargas, Nikos; Sidiropoulos, Nicholas D.(
, 2020 Asilomar Conference on on Signals, Systems, and Computers)
null
(Ed.)
Learning nonlinear functions from input-output data pairs is one of the most fundamental problems in machine learning. Recent work has formulated the problem of learning a general nonlinear multivariate function of discrete inputs, as a tensor completion problem with smooth latent factors. We build upon this idea and utilize two ensemble learning techniques to enhance its prediction accuracy. Ensemble methods can be divided into two main groups, parallel and sequential. Bagging also known as bootstrap aggregation is a parallel ensemble method where multiple base models are trained in parallel on different subsets of the data that have been chosen randomly with replacement from the original training data. The output of these models is usually combined and a single prediction is computed using averaging. One of the most popular bagging techniques is random forests. Boosting is a sequential ensemble method where a sequence of base models are fit sequentially to modified versions of the data. Popular boosting algorithms include AdaBoost and Gradient Boosting. We develop two approaches based on these ensemble learning techniques for learning multivariate functions using the Canonical Polyadic Decomposition. We showcase the effectiveness of the proposed ensemble models on several regression tasks and report significant improvements compared to the single model.
A major challenge in many machine learning tasks is that the model expressive power depends on model size. Low-rank tensor methods are an efficient tool for handling the curse of dimensionality in many large-scale machine learning models. The major challenges in training a tensor learning model include how to process the high-volume data, how to determine the tensor rank automatically, and how to estimate the uncertainty of the results. While existing tensor learning focuses on a specific task, this paper proposes a generic Bayesian framework that can be employed to solve a broad class of tensor learning problems such as tensor completion, tensor regression, and tensorized neural networks. We develop a low-rank tensor prior for automatic rank determination in nonlinear problems. Our method is implemented with both stochastic gradient Hamiltonian Monte Carlo (SGHMC) and Stein Variational Gradient Descent (SVGD). We compare the automatic rank determination and uncertainty quantification of these two solvers. We demonstrate that our proposed method can determine the tensor rank automatically and can quantify the uncertainty of the obtained results. We validate our framework on tensor completion tasks and tensorized neural network training tasks.
Table of Contents:
Foreword by the CI 2016 Workshop Chairs …………………………………vi
Foreword by the CI 2016 Steering Committee ..…………………………..…..viii
List of Organizing Committee ………………………….……....x
List of Registered Participants .………………………….……..xi
Acknowledgement of Sponsors ……………………………..…xiv
Hackathon and Workshop Agenda .………………………………..xv
Hackathon Summary .………………………….…..xviii
Invited talks - abstracts and links to presentations ………………………………..xxi
Proceedings: 34 short research papers ……………………………….. 1-135
Papers
1. BAYESIAN MODELS FOR CLIMATE RECONSTRUCTION FROM
POLLEN RECORDS ..................................... 1
Lasse Holmström, Liisa Ilvonen, Heikki Seppä, Siim Veski
2. ON INFORMATION CRITERIA FOR DYNAMIC SPATIO-TEMPORAL
CLUSTERING ..................................... 5
Ethan D. Schaeffer, Jeremy M. Testa, Yulia R. Gel, Vyacheslav Lyubchich
3. DETECTING MULTIVARIATE BIOSPHERE EXTREMES ..................................... 9
Yanira Guanche García, Erik Rodner, Milan Flach, Sebastian Sippel,
Miguel Mahecha, Joachim Denzler
4. SPATIO-TEMPORAL GENERATIVE MODELS FOR RAINFALL OVER
INDIA ..................................... 13
Adway Mitra
5. A NONPARAMETRIC COPULA BASED BIAS CORRECTION METHOD
FOR STATISTICAL DOWNSCALING ..................................... 17
Yi Li, Adam Ding, Jennifer Dy
6. DETECTING AND PREDICTING BEAUTIFUL SUNSETS USING SOCIAL
MEDIA DATA ..................................... 21
Emma Pierson
7. OCEANTEA: EXPLORING OCEAN-DERIVED CLIMATE DATA USING
MICROSERVICES ..................................... 25
Arne N. Johanson, Sascha Flögel, Wolf-Christian Dullo, Wilhelm
Hasselbring
8. IMPROVED ANALYSIS OF EARTH SYSTEM MODELS AND
OBSERVATIONS USING SIMPLE CLIMATE MODELS ..................................... 29
Balu Nadiga, Nathan Urban
9. SYNERGY AND ANALOGY BETWEEN 15 YEARS OF MICROWAVE SST
AND ALONG-TRACK SSH ..................................... 33
Pierre Tandeo, Aitor Atencia, Cristina Gonzalez-Haro
10. PREDICTING EXECUTION TIME OF CLIMATE-DRIVEN ECOLOGICAL
FORECASTING MODELS ..................................... 37
Scott Farley and John W. Williams
11. SPATIOTEMPORAL ANALYSIS OF SEASONAL PRECIPITATION OVER
US USING CO-CLUSTERING ..................................... 41
Mohammad Gorji–Sefidmazgi, Clayton T. Morrison
12. PREDICTION OF EXTREME RAINFALL USING HYBRID
CONVOLUTIONAL-LONG SHORT TERM MEMORY NETWORKS ..................................... 45
Sulagna Gope, Sudeshna Sarkar, Pabitra Mitra
13. SPATIOTEMPORAL PATTERN EXTRACTION WITH DATA-DRIVEN
KOOPMAN OPERATORS FOR CONVECTIVELY COUPLED
EQUATORIAL WAVES ..................................... 49
Joanna Slawinska, Dimitrios Giannakis
14. COVARIANCE STRUCTURE ANALYSIS OF CLIMATE MODEL OUTPUT ..................................... 53
Chintan Dalal, Doug Nychka, Claudia Tebaldi
15. SIMPLE AND EFFICIENT TENSOR REGRESSION FOR SPATIOTEMPORAL
FORECASTING ..................................... 57
Rose Yu, Yan Liu
16. TRACKING OF TROPICAL INTRASEASONAL CONVECTIVE
ANOMALIES ..................................... 61
Bohar Singh, James L. Kinter
17. ANALYSIS OF AMAZON DROUGHTS USING SUPERVISED KERNEL
PRINCIPAL COMPONENT ANALYSIS ..................................... 65
Carlos H. R. Lima, Amir AghaKouchak
18. A BAYESIAN PREDICTIVE ANALYSIS OF DAILY PRECIPITATION DATA ..................................... 69
Sai K. Popuri, Nagaraj K. Neerchal, Amita Mehta
19. INCORPORATING PRIOR KNOWLEDGE IN SPATIO-TEMPORAL
NEURAL NETWORK FOR CLIMATIC DATA ..................................... 73
Arthur Pajot, Ali Ziat, Ludovic Denoyer, Patrick Gallinari
20. DIMENSIONALITY-REDUCTION OF CLIMATE DATA USING DEEP
AUTOENCODERS ..................................... 77
Juan A. Saenz, Nicholas Lubbers, Nathan M. Urban
21. MAPPING PLANTATION IN INDONESIA ..................................... 81
Xiaowei Jia, Ankush Khandelwal, James Gerber, Kimberly Carlson, Paul
West, Vipin Kumar
22. FROM CLIMATE DATA TO A WEIGHTED NETWORK BETWEEN
FUNCTIONAL DOMAINS ..................................... 85
Ilias Fountalis, Annalisa Bracco, Bistra Dilkina, Constantine Dovrolis
23. EMPLOYING SOFTWARE ENGINEERING PRINCIPLES TO ENHANCE
MANAGEMENT OF CLIMATOLOGICAL DATASETS FOR CORAL REEF
ANALYSIS ..................................... 89
Mark Jenne, M.M. Dalkilic, Claudia Johnson
24. Profiler Guided Manual Optimization for Accelerating Cholesky
Decomposition on R Environment ..................................... 93
V.B. Ramakrishnaiah, R.P. Kumar, J. Paige, D. Hammerling, D. Nychka
25. GLOBAL MONITORING OF SURFACE WATER EXTENT DYNAMICS
USING SATELLITE DATA ..................................... 97
Anuj Karpatne, Ankush Khandelwal and Vipin Kumar
26. TOWARD QUANTIFYING TROPICAL CYCLONE RISK USING
DIAGNOSTIC INDICES .................................... 101
Erica M. Staehling and Ryan E. Truchelut
27. OPTIMAL TROPICAL CYCLONE INTENSITY ESTIMATES WITH
UNCERTAINTY FROM BEST TRACK DATA .................................... 105
Suz Tolwinski-Ward
28. EXTREME WEATHER PATTERN DETECTION USING DEEP
CONVOLUTIONAL NEURAL NETWORK .................................... 109
Yunjie Liu, Evan Racah, Prabhat, Amir Khosrowshahi, David Lavers,
Kenneth Kunkel, Michael Wehner, William Collins
29. INFORMATION TRANSFER ACROSS TEMPORAL SCALES IN
ATMOSPHERIC DYNAMICS .................................... 113
Nikola Jajcay and Milan Paluš
30. Identifying precipitation regimes in China using model-based
clustering of spatial functional data .................................... 117
Haozhe Zhang, Zhengyuan Zhu, Shuiqing Yin
31. RELATIONAL RECURRENT NEURAL NETWORKS FOR SPATIOTEMPORAL
INTERPOLATION FROM MULTI-RESOLUTION CLIMATE
DATA .................................... 121
Guangyu Li, Yan Liu
32. OBJECTIVE SELECTION OF ENSEMBLE BOUNDARY CONDITIONS
FOR CLIMATE DOWNSCALING .................................... 124
Andrew Rhines, Naomi Goldenson
33. LONG-LEAD PREDICTION OF EXTREME PRECIPITATION CLUSTER VIA
A SPATIO-TEMPORAL CONVOLUTIONAL NEURAL NETWORK .................................... 128
Yong Zhuang, Wei Ding
34. MULTIPLE INSTANCE LEARNING FOR BURNED AREA MAPPING
USING MULTI –TEMPORAL REFLECTANCE DATA .................................... 132
Guruprasad Nayak, Varun Mithal, Vipin Kumar
Ge, Rong; Kuditipudi, Rohith; Li, Zhize; Wang, Xiang(
, International Conference on Learning Representations)
We give a new algorithm for learning a two-layer neural network under a general
class of input distributions. Assuming there is a ground-truth two-layer network
y = Aσ(Wx) + ξ,
where A,W are weight matrices, ξ represents noise, and the number of neurons in
the hidden layer is no larger than the input or output, our algorithm is guaranteed to
recover the parameters A,W of the ground-truth network. The only requirement
on the input x is that it is symmetric, which still allows highly complicated and
structured input.
Our algorithm is based on the method-of-moments framework and extends several
results in tensor decompositions. We use spectral algorithms to avoid the complicated
non-convex optimization in learning neural networks. Experiments show
that our algorithm can robustly learn the ground-truth neural network with a small
number of samples for many symmetric input distributions.
Abstract. A key challenge for biological oceanography is relating the physiologicalmechanisms controlling phytoplankton growth to the spatial distribution ofthose phytoplankton. Physiological mechanisms are often isolated by varyingone driver of growth, such as nutrient or light, in a controlled laboratorysetting producing what we call “intrinsic relationships”. We contrastthese with the “apparent relationships” which emerge in the environment inclimatological data. Although previous studies have found machine learning(ML) can find apparent relationships, there has yet to be a systematic studyexamining when and why these apparent relationships diverge from theunderlying intrinsic relationships found in the lab and how and why this may depend on the method applied. Here we conduct a proof-of-concept studywith three scenarios in which biomass is by construction a function oftime-averaged phytoplankton growth rate. In the first scenario, the inputsand outputs of the intrinsic and apparent relationships vary over thesame monthly timescales. In the second, the intrinsic relationships relateaverages of drivers that vary on hourly timescales to biomass, but theapparent relationships are sought between monthly averages of these inputsand monthly-averaged output. In the third scenario we apply ML to the outputof an actual Earth system model (ESM). Our results demonstrated that whenintrinsic and apparent relationships operate on the same spatial andtemporal timescale, neural network ensembles (NNEs) were able to extract theintrinsic relationships when only provided information about the apparentrelationships, while colimitation and its inability to extrapolate resulted in random forests (RFs) diverging from the true response. Whenintrinsic and apparent relationships operated on different timescales (aslittle separation as hourly versus daily), NNEs fed with apparentrelationships in time-averaged data produced responses with the right shapebut underestimated the biomass. This was because when the intrinsicrelationship was nonlinear, the response to a time-averaged input differedsystematically from the time-averaged response. Although the limitationsfound by NNEs were overestimated, they were able to produce more realisticshapes of the actual relationships compared to multiple linear regression.Additionally, NNEs were able to model the interactions between predictorsand their effects on biomass, allowing for a qualitative assessment of thecolimitation patterns and the nutrient causing the most limitation. Futureresearch may be able to use this type of analysis for observational datasetsand other ESMs to identify apparent relationships between biogeochemicalvariables (rather than spatiotemporal distributions only) and identifyinteractions and colimitations without having to perform (or at leastperforming fewer) growth experiments in a lab. From our study, it appearsthat ML can extract useful information from ESM output and could likely doso for observational datasets as well.
Kargas, Nikos, and Sidiropoulos, Nicholas D. Nonlinear System Identification via Tensor Completion. Retrieved from https://par.nsf.gov/biblio/10169291. Proceedings of the AAAI Conference on Artificial Intelligence 34.04 Web. doi:10.1609/aaai.v34i04.5868.
Kargas, Nikos, & Sidiropoulos, Nicholas D. Nonlinear System Identification via Tensor Completion. Proceedings of the AAAI Conference on Artificial Intelligence, 34 (04). Retrieved from https://par.nsf.gov/biblio/10169291. https://doi.org/10.1609/aaai.v34i04.5868
@article{osti_10169291,
place = {Country unknown/Code not available},
title = {Nonlinear System Identification via Tensor Completion},
url = {https://par.nsf.gov/biblio/10169291},
DOI = {10.1609/aaai.v34i04.5868},
abstractNote = {Function approximation from input and output data pairs constitutes a fundamental problem in supervised learning. Deep neural networks are currently the most popular method for learning to mimic the input-output relationship of a general nonlinear system, as they have proven to be very effective in approximating complex highly nonlinear functions. In this work, we show that identifying a general nonlinear function y = ƒ(x1,…,xN) from input-output examples can be formulated as a tensor completion problem and under certain conditions provably correct nonlinear system identification is possible. Specifically, we model the interactions between the N input variables and the scalar output of a system by a single N-way tensor, and setup a weighted low-rank tensor completion problem with smoothness regularization which we tackle using a block coordinate descent algorithm. We extend our method to the multi-output setting and the case of partially observed data, which cannot be readily handled by neural networks. Finally, we demonstrate the effectiveness of the approach using several regression tasks including some standard benchmarks and a challenging student grade prediction task.},
journal = {Proceedings of the AAAI Conference on Artificial Intelligence},
volume = {34},
number = {04},
author = {Kargas, Nikos and Sidiropoulos, Nicholas D.},
}
Warning: Leaving National Science Foundation Website
You are now leaving the National Science Foundation website to go to a non-government website.
Website:
NSF takes no responsibility for and exercises no control over the views expressed or the accuracy of
the information contained on this site. Also be aware that NSF's privacy policy does not apply to this site.