Karst aquifers are important groundwater resources that supply drinking water for approximately 25 % of the world’s population. Their complex hydrogeological structures, dual-flow regimes, and highly heterogeneous flow pose significant challenges for accurate hydrodynamic modeling and sustainable management. Traditional modeling approaches often struggle to capture the intricate spatial dependencies and multi-scale temporal patterns inherent in karst systems, particularly the interactions between rapid conduit flow and slower matrix flow. This study proposes a novel multi-scale dynamic graph attention network integrated with long short-term memory model (GAT-LSTM) to innovatively learn and integrate spatial and temporal dependencies in karst systems for forecasting spring discharge. The model introduces several innovative components: (1) graph-based neural networks with dynamic edge-weighting mechanism are proposed to learn and update spatial dependencies based on both geographic distances and learned hydrological relationships, (2) a multi-head attention mechanism is adopted to capture different aspects of spatial relationships simultaneously, and (3) a hierarchical temporal architecture is incorporated to process hydrological temporal patterns at both monthly and seasonal scales with an adaptive fusion mechanism for final results. These features enable the proposed model to effectively account for the dual-flow dynamics in karst systems, where rapid conduit flow and slower matrix flow coexist. The newly proposed model is applied to the Barton Springs of the Edwards Aquifer in Texas. The results demonstrate that it can obtain more accurate and robust prediction performance across various time steps compared to traditional temporal and spatial deep learning approaches. Based on the multi-scale GAT-LSTM model, a comprehensive ablation analysis and permutation feature important are conducted to analyze the relative contribution of various input variables on the final prediction. These findings highlight the intricate nature of karst systems and demonstrate that effective spring discharge prediction requires comprehensive monitoring networks encompassing both primary recharge contributors and supplementary hydrological features that may serve as valuable indicators of system-wide conditions.
more »
« less
This content will become publicly available on December 1, 2025
Interpretable multi-step hybrid deep learning model for karst spring discharge prediction: Integrating temporal fusion transformers with ensemble empirical mode decomposition
Karst groundwater is a critical freshwater resource for numerous regions worldwide. Monitoring and predicting karst spring discharge is essential for effective groundwater management and the preservation of karst ecosystems. However, the high heterogeneity and karstification pose significant challenges to physics-based models in providing robust predictions of karst spring discharge. In this study, an interpretable multi-step hybrid deep learning model called selective EEMD-TFT is proposed, which adaptively integrates temporal fusion transformers (TFT) with ensemble empirical mode decomposition (EEMD) for predicting karst spring discharge. The selective EEMD-TFT hybrid model leverages the strengths of both EEMD and TFT techniques to learn inherent patterns and temporal dynamics from nonlinear and nonstationary signals, eliminate redundant components, and emphasize useful characteristics of input variables, leading to the improvement of prediction performance and efficiency. It consists of two stages: in the first stage, the daily precipitation data is decomposed into multiple intrinsic mode functions using EEMD to extract valuable information from nonlinear and nonstationary signals. All decomposed components, temperature and categorical date features are then fed into the TFT model, which is an attention- based deep learning model that combines high-performance multi-horizon prediction and interpretable insights into temporal dynamics. The importance of input variables will be quantified and ranked. In the second stage, the decomposed precipitation components with high importance are selected to serve as the TFT model’s input features along with temperature and categorical date variables for the final prediction. Results indicate that the selective EEMD-TFT model outperforms other sequence-to-sequence deep learning models, such as LSTM and single TFT models, delivering reliable and robust prediction performance. Notably, it maintains more consistent prediction performance at longer forecast horizons compared to other sequence-to-sequence models, highlighting its capacity to learn complex patterns from the input data and efficiently extract valuable information for karst spring prediction. An interpretable analysis of the selective EEMD-TFT model is conducted to gain insights into relationships among various hydrological processes and analyze temporal patterns.
more »
« less
- Award ID(s):
- 2407963
- PAR ID:
- 10590604
- Publisher / Repository:
- Elsevier
- Date Published:
- Journal Name:
- Journal of Hydrology
- Volume:
- 645
- Issue:
- PA
- ISSN:
- 0022-1694
- Page Range / eLocation ID:
- 132235
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Abstract Sparse precipitation data in karst catchments challenge hydrologic models to accurately capture the spatial and temporal relationships between precipitation and karst spring discharge, hindering robust predictions. This study addresses this issue by employing a coupled deep learning model that integrates a variation autoencoder (VAE) for augmenting precipitation and a long short‐term memory (LSTM) network for karst spring discharge prediction. The VAE contributes by generating synthetic precipitation data through an encoding‐decoding process. This process generalizes the observed precipitation data by deriving joint latent distributions with improved preservation of temporal and spatial correlations of the data. The combined VAE‐generated precipitation and observation data are used to train and test the LSTM to predict spring discharge. Applied to the Niangziguan spring catchment in northern China, the average performance of NSE, root mean square error, mean absolute error, mean absolute percentage error, and log NSE of our coupled VAE/LSTM model reached 0.93, 0.26, 0.15, 1.8, and 0.92, respectively, yielding 145%, 52%, 63%, 70% and 149% higher than an LSTM model using only observations. We also explored temporal and spatial correlations in the observed data and the impact of different ratios of VAE‐generated precipitation data to actual data on model performances. This study also evaluated the effectiveness of VAE‐augmented data on various deep‐learning models and compared VAE with other data augmentation techniques. We demonstrate that the VAE offers a novel approach to address data scarcity and uncertainty, improving learning generalization and predictive capability of various hydrological models. However, we recognize that innovations to address hydrologic problems at different scales remain to be explored.more » « less
-
Interpreting County-Level COVID-19 Infections using Transformer and Deep Learning Time Series ModelsDeep Learning for Time-series plays a key role in AI for healthcare. To predict the progress of infectious disease outbreaks and demonstrate clear population-level impact, more granular analyses are urgently needed that control for important and potentially confounding county-level socioeconomic and health factors. We forecast US county-level COVID-19 infections using the Temporal Fusion Transformer (TFT). We focus on heterogeneous time-series deep learning model prediction while interpreting the complex spatiotemporal features learned from the data. The significance of the work is grounded in a real-world COVID-19 infection prediction with highly non-stationary, finely granular, and heterogeneous data. 1) Our model can capture the detailed daily changes of temporal and spatial model behaviors and achieves better prediction performance compared to other time-series models. 2) We analyzed the attention patterns from TFT to interpret the temporal and spatial patterns learned by the model. 3) We collected around 2.5 years of socioeconomic and health features for 3142 US counties, such as observed cases, and a number of static (age distribution and health disparity) and dynamic features (vaccination, disease spread, transmissible cases, and social distancing). Using the proposed framework, we have shown that our model can learn complex interactions. Interpreting different impacts at the county level would be crucial for understanding the infection process that can help effective public health decision-making.more » « less
-
Interpreting County-Level COVID-19 Infections using Transformer and Deep Learning Time Series ModelsDeep Learning for Time-series plays a key role in AI for healthcare. To predict the progress of infectious disease outbreaks and demonstrate clear population-level impact, more granular analyses are urgently needed that control for important and potentially confounding county-level socioeconomic and health factors. We forecast US county-level COVID-19 infections using the Temporal Fusion Transformer (TFT). We focus on heterogeneous time-series deep learning model prediction while interpreting the complex spatiotemporal features learned from the data. The significance of the work is grounded in a real-world COVID-19 infection prediction with highly non-stationary, finely granular, and heterogeneous data. 1) Our model can capture the detailed daily changes of temporal and spatial model behaviors and achieves better prediction performance compared to other time-series models. 2) We analyzed the attention patterns from TFT to interpret the temporal and spatial patterns learned by the model. 3) We collected around 2.5 years of socioeconomic and health features for 3142 US counties, such as observed cases, and a number of static (age distribution and health disparity) and dynamic features (vaccination, disease spread, transmissible cases, and social distancing). Using the proposed framework, we have shown that our model can learn complex interactions. Interpreting different impacts at the county level would be crucial for understanding the infection process that can help effective public health decision-making.more » « less
-
Abstract Oxygen and hydrogen isotopes were used in this study to detect a hydraulic connection between a sinkhole lake and a karst spring. In karst areas, surface water that flows to a lake can drain through sinkholes in the lakebed to the underlying aquifer, and then flows in karst conduits and through aquifer matrix. At the study site located in northwest Florida, USA, Lake Miccosukee immediately drains into two sinkholes. Results from a dye tracing experiment indicate that lake water discharges at Natural Bridge Spring, a first‐magnitude spring 32 km downgradient from the lake. By collecting weekly water samples from the lake, the spring, and a groundwater well 10 m away from the lake during the dry period between October 2019 and January 2020, it was found that, when rainfall effects on isotopic signature in spring water are removed, increased isotope ratios of spring water can be explained by mixing of heavy‐isotope‐enriched lake water into groundwater, indicating hydraulic connection between the lake and the spring. Such a detection of hydraulic connection at the scale of tens of kilometers and for a first‐magnitude spring has not been previously reported in the literature. Based on the isotope ratio data, it was estimated that, during the study period, about 8.5% the spring discharge was the lake water that drained into the lake sinkholes.more » « less
An official website of the United States government
