skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Spline-based neural network interatomic potentials: Blending classical and machine learning models
While machine learning (ML) interatomic potentials (IPs) are able to achieve accuracies nearing the level of noise inherent in the first-principles data to which they are trained, it remains to be shown if their increased complexities are strictly necessary for constructing high-quality IPs. In this work, we introduce a new MLIP framework which blends the simplicity of spline-based MEAM (s-MEAM) potentials with the flexibility of a neural network (NN) architecture. The proposed framework, which we call the spline-based neural network potential (s-NNP), is a simplified version of the traditional NNP that can be used to describe complex datasets in a computationally efficient manner. We demonstrate how this framework can be used to probe the boundary between classical and ML IPs, highlighting the benefits of key architectural changes. Furthermore, we show that using spline filters for encoding atomic environments results in a readily interpreted embedding layer which can be coupled with modifications to the NN to incorporate expected physical behaviors and improve overall interpretability. Finally, we test the flexibility of the spline filters, observing that they can be shared across multiple chemical systems in order to provide a convenient reference point from which to begin performing cross-system analyses.  more » « less
Award ID(s):
1940303 1922758
PAR ID:
10532004
Author(s) / Creator(s):
;
Publisher / Repository:
Elsevier
Date Published:
Journal Name:
Computational Materials Science
Volume:
232
Issue:
C
ISSN:
0927-0256
Page Range / eLocation ID:
112655
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Neural network potentials (NNPs) trained against density functional theory (DFT) are capable of reproducing the potential energy surface at a fraction of the computational cost. However, most NNP implementations focus on energy and forces. In this work, we modified the NNP model introduced by Behler and Parrinello to predict Fermi energy, band edges, and partial density of states of Cu 2 O. Our NNP can reproduce the DFT potential energy surface and properties at a fraction of the computational cost. We used our NNP to perform molecular dynamics (MD) simulations and validated the predicted properties against DFT calculations. Our model achieved a root mean squared error of 16 meV for the energy prediction. Furthermore, we show that the standard deviation of the energies predicted by the ensemble of training snapshots can be used to estimate the uncertainty in the predictions. This allows us to switch from the NNP to DFT on-the-fly during the MD simulation to evaluate the forces when the uncertainty is high. 
    more » « less
  2. Ab initio methods offer great promise for materials design, but they come with a hefty computational cost. Recent advances with machine learning interatomic potentials (MLIPs) have revolutionized molecular dynamic simulations by providing high accuracies similar to ab initio models but at much reduced computational cost. Our study evaluates the ultra-fast force fields (UF3) potential, employing linear regression with cubic B-spline basis for assessing effective two- and three-body potentials. On benchmarking, UF3 displays comparable precision to established models like GAP, MTP, NNP (Behler Parrinello), and qSNAP MLIPs, yet is significantly faster by two to three orders of magnitude. A distinct feature of UF3 is its capability to render visual representations of learned two- and three-body potentials, shedding light on potential gaps in the learning model. In refining UF3’s performance, a comprehensive sweep of the hyperparameter space was undertaken. While our current optimizations are concentrated on energies and forces, we are primed to broaden UF3’s evaluation spectrum, focusing on its applicability in critical areas of molecular dynamics simulations. The outcome of these investigations will not only enhance the predictability and usability of UF3 but also pave the way for its broader applications in advanced materials discovery and simulations. 
    more » « less
  3. Abstract Machine learning interatomic potentials (IPs) can provide accuracy close to that of first-principles methods, such as density functional theory (DFT), at a fraction of the computational cost. This greatly extends the scope of accurate molecular simulations, providing opportunities for quantitative design of materials and devices on scales hitherto unreachable by DFT methods. However, machine learning IPs have a basic limitation in that they lack a physical model for the phenomena being predicted and therefore have unknown accuracy when extrapolating outside their training set. In this paper, we propose a class of Dropout Uncertainty Neural Network (DUNN) potentials that provide rigorous uncertainty estimates that can be understood from both Bayesian and frequentist statistics perspectives. As an example, we develop a DUNN potential for carbon and show how it can be used to predict uncertainty for static and dynamical properties, including stress and phonon dispersion in graphene. We demonstrate two approaches to propagate uncertainty in the potential energy and atomic forces to predicted properties. In addition, we show that DUNN uncertainty estimates can be used to detect configurations outside the training set, and in some cases, can serve as a predictor for the accuracy of a calculation. 
    more » « less
  4. Yortsos, Yannis (Ed.)
    Abstract Transfer learning (TL), which enables neural networks (NNs) to generalize out-of-distribution via targeted re-training, is becoming a powerful tool in scientific machine learning (ML) applications such as weather/climate prediction and turbulence modeling. Effective TL requires knowing (1) how to re-train NNs? and (2) what physics are learned during TL? Here, we present novel analyses and a framework addressing (1)–(2) for a broad range of multi-scale, nonlinear, dynamical systems. Our approach combines spectral (e.g. Fourier) analyses of such systems with spectral analyses of convolutional NNs, revealing physical connections between the systems and what the NN learns (a combination of low-, high-, band-pass filters and Gabor filters). Integrating these analyses, we introduce a general framework that identifies the best re-training procedure for a given problem based on physics and NN theory. As test case, we explain the physics of TL in subgrid-scale modeling of several setups of 2D turbulence. Furthermore, these analyses show that in these cases, the shallowest convolution layers are the best to re-train, which is consistent with our physics-guided framework but is against the common wisdom guiding TL in the ML literature. Our work provides a new avenue for optimal and explainable TL, and a step toward fully explainable NNs, for wide-ranging applications in science and engineering, such as climate change modeling. 
    more » « less
  5. In silico property prediction based on density functional theory (DFT) is increasingly performed for crystalline materials. Whether quantitative agreement with experiment can be achieved with current methods is often an unresolved question, and may require detailed examination of physical effects such as electron correlation, reciprocal space sampling, phonon anharmonicity, and nuclear quantum effects (NQE), among others. In this work, we attempt first-principles equation of state prediction for the crystalline materials ScF3 and CaZrF6, which are known to exhibit negative thermal expansion (NTE) over a broad temperature range. We develop neural network (NN) potentials for both ScF3 and CaZrF6 trained to extensive DFT data, and conduct direct molecular dynamics prediction of the equation(s) of state over a broad temperature/pressure range. The NN potentials serve as surrogates of the DFT Hamiltonian with enhanced computational efficiency allowing for simulations with larger supercells and inclusion of NQE utilizing path integral approaches. The conclusion of the study is mixed: while some equation of state behavior is predicted in semiquantitative agreement with experiment, the pressure-induced softening phenomenon observed for ScF3 is not captured in our simulations. We show that NQE have a moderate effect on NTE at low temperature but does not significantly contribute to equation of state predictions at increasing temperature. Overall, while the NN potentials are valuable for property prediction of these NTE (and related) materials, we infer that a higher level of electron correlation, beyond the generalized gradient approximation density functional employed here, is necessary for achieving quantitative agreement with experiment. 
    more » « less