Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Free, publicly-accessible full text available June 1, 2025
-
Abstract There is demand for scalable algorithms capable of clustering and analyzing large time series data. The Kohonen self-organizing map (SOM) is an unsupervised artificial neural network for clustering, visualizing, and reducing the dimensionality of complex data. Like all clustering methods, it requires a measure of similarity between input data (in this work time series). Dynamic time warping (DTW) is one such measure, and a top performer that accommodates distortions when aligning time series. Despite its popularity in clustering, DTW is limited in practice because the runtime complexity is quadratic with the length of the time series. To address this, we present a new a self-organizing map for clustering TIME Series, called SOMTimeS, which uses DTW as the distance measure. The method has similar accuracy compared with other DTW-based clustering algorithms, yet scales better and runs faster. The computational performance stems from the pruning of unnecessary DTW computations during the SOM’s training phase. For comparison, we implement a similar pruning strategy for K-means, and call the latter K-TimeS. SOMTimeS and K-TimeS pruned 43% and 50% of the total DTW computations, respectively. Pruning effectiveness, accuracy, execution time and scalability are evaluated using 112 benchmark time series datasets from the UC Riverside classification archive, and show that for similar accuracy, a 1.8
speed-up on average for SOMTimeS and K-TimeS, respectively with that rates vary between 1$$\times$$ and 18$$\times$$ depending on the dataset. We also apply SOMTimeS to a healthcare study of patient-clinician serious illness conversations to demonstrate the algorithm’s utility with complex, temporally sequenced natural language.$$\times$$ -
Abstract. Large sample datasets are transforming the catchment sciences, but there are few off-the-shelf stream water chemistry datasets with complementary atmospheric deposition, streamflow, meteorology, and catchment physiographic attributes. The existing CAMELS (Catchment Attributes and Meteorology for Large-sample Studies) dataset includes data on topography, climate, streamflow, land cover, soil, and geology across the continental US. With CAMELS-Chem, we pair these existing attribute data for 516 catchments with atmospheric deposition data from the National Atmospheric Deposition Program and water chemistry and instantaneous discharge data from the US Geological Survey over the period from 1980 through 2018 in a relational database and corresponding dataset. The data include 18 common stream water chemistry constituents: Al, Ca, Cl, dissolved organic carbon, total organic carbon, HCO3, K, Mg, Na, total dissolved N, total organic N, NO3, dissolved oxygen, pH (field and lab), Si, SO4, and water temperature. Annual deposition loads and concentrations include hydrogen, NH4, NO3, total inorganic N, Cl, SO4, Ca, K, Mg, and Na. We demonstrate that CAMELS-Chem water chemistry data are sampled effectively across climates, seasons, and discharges for trend analysis and highlight the coincident sampling of stream constituents for process-based understanding. To motivate their use by the larger scientific community across a variety of disciplines, we show examples of how these publicly available datasets can be applied to trend detection and attribution, biogeochemical process understanding, and new hypothesis generation via data-driven techniques.
Free, publicly-accessible full text available January 1, 2025 -
The role of living labs in cultivating inclusive and responsible innovation in precision agricultureFree, publicly-accessible full text available April 1, 2025
-
Abstract Because the manual counting of soybean ( Glycine max ) plants, pods, and seeds/pods is unsuitable for soybean yield predictions, alternative methods are desired. Therefore, the objective was to determine if satellite remote sensing − based artificial intelligence (AI) models could be used to predict soybean yield. In the study, multiple remote sensing − based AI models were developed for soybean growth stage ranging from VE/VC (plant emergence) to R6/R7 (full seed to beginning maturity). The ability of the Deep Neural Network (DNN), Support Vector Machine (SVM), Random Forest (RF), Least Absolute Shrinkage and Selection Operator (LASSO), and AdaBoost to predict soybean yield, based on blue, green, red, and near infrared reflectance data collected by the PlanetScope satellite at 6 growth stages, was determined. Remote sensing and soybean yield monitor data from 3 different fields in two years (2019 and 2021) were aggregated into 24,282 grid cells that had the dimensions of 10 by 10m. A comparison across models showed that the DNN outperformed the other models. Moreover, as crops matured from VE/VC to R4/R5, the R 2 value of the models increased from 0.26 to over 0.70. These findings indicate that remote sensing data collected at different growth stages can be combined for soybean yield predictions. Moreover, additional work needs to be conducted to assess the model's ability to predict soybean yield with vegetation indices (VI) data for fields not used to train the model. This article is protected by copyright. All rights reservedmore » « less
-
Abstract Precision agriculture (PA) has been defined as a “management strategy that gathers, processes and analyzes temporal, spatial and individual data and combines it with other information to support management decisions according to estimated variability for improved resource use efficiency, productivity, quality, profitability and sustainability of agricultural production.” This definition suggests that because PA should simultaneously increase food production and reduce the environmental footprint, the barriers to adoption of PA should be explored. These barriers include (1) the financial constraints associated with adopting decision support system (DSS); (2) the hesitancy of farmers to change from their trusted advisor to a computer program that often behaves as a black box; (3) questions about data ownership and privacy; and (4) the lack of a trained workforce to provide the necessary training to implement DSSs on individual farms. This paper also discusses the lessons learned from successful and unsuccessful efforts to implement DSSs, the importance of communication with end users during DSS development, and potential career opportunities that DSSs are creating in PA.
-
Abstract Artificial intelligence (AI) represents technologies with human‐like cognitive abilities to learn, perform, and make decisions. AI in precision agriculture (PA) enables farmers and farm managers to deploy highly targeted and precise farming practices based on site‐specific agroclimatic field measurements. The foundational and applied development of AI has matured considerably over the last 30 years. The time is now right to engage seriously with the ethics and responsible practice of AI for the well‐being of farmers and farm managers. In this paper, we identify and discuss both challenges and opportunities for improving farmers’ trust in those providing AI solutions for PA. We highlight that farmers’ trust can be moderated by how the benefits and risks of AI are perceived, shared, and distributed. We propose four recommendations for improving farmers’ trust. First, AI developers should improve model transparency and explainability. Second, clear responsibility and accountability should be assigned to AI decisions. Third, concerns about the fairness of AI need to be overcome to improve human‐machine partnerships in agriculture. Finally, regulation and voluntary compliance of data ownership, privacy, and security are needed, if AI systems are to become accepted and used by farmers.