skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: PI-LSTM: Physics-informed long short-term memory network for structural response modeling
Deep learning models have achieved remarkable accuracy for structural response modeling. However, these models heavily depend on having a sufficient amount of training data, which can be challenging and time-consuming to collect. Moreover, data-driven models sometimes struggle to adhere to physics constraints. Therefore, in this study, a physics-informed long short-term memory (PI-LSTM) network was applied to structural response modeling by incorporating physics constraints into deep learning. The physics constraints were modified to accommodate the characteristics of both linear and nonlinear cases. The PI-LSTM network, inspired by and compared with existing physics-informed deep learning models (PhyCNN and PhyLSTM), was validated using the numerical simulation results of the single-degree-of-freedom (SDOF) system and the experimental results of the six-story building. Additionally, the PI-LSTM network underwent thorough investigation and validation across the four cases of the SDOF system and numerical simulation results of the six-story building with the comparison of the regular LSTM. The results indicate that the PI-LSTM network outperformed the regular LSTM models in terms of accuracy. Furthermore, the PI-LSTM network exhibited a more concentrated and higher accuracy range when analyzing the results of both the SDOF system and the six-story building. These findings demonstrate that the PI-LSTM network presents a reliable and efficient approach for structural response modeling.  more » « less
Award ID(s):
2308924
PAR ID:
10548515
Author(s) / Creator(s):
; ;
Corporate Creator(s):
Editor(s):
NA
Publisher / Repository:
Elsevier
Date Published:
Journal Name:
Engineering Structures
Edition / Version:
1
Volume:
292
Issue:
C
ISSN:
0141-0296
Page Range / eLocation ID:
116500
Subject(s) / Keyword(s):
Deep learning Physics-informed long short-term memory network Structural response modeling Physics constraint
Format(s):
Medium: X Size: 5 Other: pdf
Size(s):
5
Sponsoring Org:
National Science Foundation
More Like this
  1. In order to evaluate urban earthquake resilience, reliable structural modeling is needed. However, detailed modeling of a large number of structures and carrying out time history analyses for sets of ground motions are not practical at an urban scale. Reduced-order surrogate models can expedite numerical simulations while maintaining necessary engineering accuracy. Neural networks have been shown to be a powerful tool for developing surrogate models, which often outperform classical surrogate models in terms of scalability of complex models. Training a reliable deep learning model, however, requires an immense amount of data that contain a rich input-output relationship, which typically cannot be satisfied in practical applications. In this paper, we propose model-informed symbolic neural networks (MiSNN) that can discover the underlying closed-form formulations (differential equations) for a reduced-order surrogate model. The MiSNN will be trained on datasets obtained from dynamic analyses of detailed reinforced concrete special moment frames designed for San Francisco, California, subject to a series of selected ground motions. Training the MiSNN is equivalent to finding the solution to a sparse optimization problem, which is solved by the Adam optimizer. The earthquake ground acceleration and story displacement, velocity, and acceleration time histories will be used to train 1) an integrated SNN, which takes displacement and velocity states and outputs the absolute acceleration response of the structure; and 2) a distributed SNN, which distills the underlying equation of motion for each story. The results show that the MiSNN can reduce computational cost while maintaining high prediction accuracy of building responses. 
    more » « less
  2. Solar wind modeling is classified into two main types: empirical models and physics-based models, each designed to forecast solar wind properties in various regions of the heliosphere. Empirical models, which are cost-effective, have demonstrated significant accuracy in predicting solar wind at the L1 Lagrange point. On the other hand, physics-based models rely on magnetohydrodynamics (MHD) principles and demand more computational resources. In this research paper, we build upon our recent novel approach that merges empirical and physics-based models. Our recent proposal involves the creation of a new physics-informed neural network that leverages time series data from solar wind predictors to enhance solar wind prediction. This innovative method aims to combine the strengths of both modeling approaches to achieve more accurate and efficient solar wind predictions. In this work, we show the variability of the proposed physics-informed loss across multiple deep learning models. We also study the effect of training the models on different solar cycles on the model’s performance. This work represents the first effort to predict solar wind by integrating deep learning approaches with physics constraints and analyzing the results across three solar cycles. Our findings demonstrate the superiority of our physics-constrained model over other unconstrained deep learning predictive models. 
    more » « less
  3. Operator learning has become a powerful tool in machine learning for modeling complex physical systems governed by partial differential equations (PDEs). Although Deep Operator Networks (DeepONet) show promise, they require extensive data acquisition. Physics-informed DeepONets (PI-DeepONet) mitigate data scarcity but suffer from inefficient training processes. We introduce Separable Operator Networks (SepONet), a novel framework that significantly enhances the efficiency of physics-informed operator learning. SepONet uses independent trunk networks to learn basis functions separately for different coordinate axes, enabling faster and more memory-efficient training via forward-mode automatic differentiation. We provide a universal approximation theorem for SepONet proving the existence of a separable approximation to any nonlinear continuous operator. Then, we comprehensively benchmark its representational capacity and computational performance against PI-DeepONet. Our results demonstrate SepONet's superior performance across various nonlinear and inseparable PDEs, with SepONet's advantages increasing with problem complexity, dimension, and scale. For 1D time-dependent PDEs, SepONet achieves up to 112× faster training and 82× reduction in GPU memory usage compared to PI-DeepONet, while maintaining comparable accuracy. For the 2D time-dependent nonlinear diffusion equation, SepONet efficiently handles the complexity, achieving a 6.44\% mean relative $$\ell_{2}$$ test error, while PI-DeepONet fails due to memory constraints. This work paves the way for extreme-scale learning of continuous mappings between infinite-dimensional function spaces. 
    more » « less
  4. Operator learning has become a powerful tool in machine learning for modeling complex physical systems governed by partial differential equations (PDEs). Although Deep Operator Networks (DeepONet) show promise, they require extensive data acquisition. Physics-informed DeepONets (PI-DeepONet) mitigate data scarcity but suffer from inefficient training processes. We introduce Separable Operator Networks (SepONet), a novel framework that significantly enhances the efficiency of physics-informed operator learning. SepONet uses independent trunk networks to learn basis functions separately for different coordinate axes, enabling faster and more memory-efficient training via forward-mode automatic differentiation. We provide a universal approximation theorem for SepONet proving the existence of a separable approximation to any nonlinear continuous operator. Then, we comprehensively benchmark its representational capacity and computational performance against PI-DeepONet. Our results demonstrate SepONet's superior performance across various nonlinear and inseparable PDEs, with SepONet's advantages increasing with problem complexity, dimension, and scale. For 1D time-dependent PDEs, SepONet achieves up to 112× faster training and 82× reduction in GPU memory usage compared to PI-DeepONet, while maintaining comparable accuracy. For the 2D time-dependent nonlinear diffusion equation, SepONet efficiently handles the complexity, achieving a 6.44\% mean relative $$\el_2$$ test error, while PI-DeepONet fails due to memory constraints. This work paves the way for extreme-scale learning of continuous mappings between infinite-dimensional function spaces. 
    more » « less
  5. Wang, Dong (Ed.)
    Lithium-ion batteries have been extensively used to power portable electronics, electric vehicles, and unmanned aerial vehicles over the past decade. Aging decreases the capacity of Lithium-ion batteries. Therefore, accurate remaining useful life (RUL) prediction is critical to the reliability, safety, and efficiency of the Lithium-ion battery-powered systems. However, battery aging is a complex electrochemical process affected by internal aging mechanisms and operating conditions (e.g., cycle time, environmental temperature, and loading condition). In this paper, a physics-informed machine learning method is proposed to model the degradation trend and predict the RUL of Lithium-ion batteries while accounting for battery health and operating conditions. The proposed physics-informed long short-term memory (PI-LSTM) model combines a physics-based calendar and cycle aging (CCA) model with an LSTM layer. The CCA model measures the aging effect of Lithium-ion batteries by combining five operating stress factor models. The PI-LSTM uses an LSTM layer to learn the relationship between the degradation trend determined by the CCA model and the online monitoring data of different cycles (i.e., voltage, current, and cell temperature). After the degradation pattern of a battery is estimated by the PI-LSTM model, another LSTM model is then used to predict the future degradation and remaining useful life (RUL) of the battery by learning the degradation trend estimated by the PI-LSTM model. Monitoring data of eleven Lithium-ion batteries under different operating conditions was used to demonstrate the proposed method. Experimental results have shown that the proposed method can accurately model the degradation behavior as well as predict the RUL of Lithium-ion batteries under different operating conditions. 
    more » « less