skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: tobac v1.5: introducing fast 3D tracking, splits and mergers, and other enhancements for identifying and analysing meteorological phenomena
Abstract. There is a continuously increasing need for reliable feature detection and tracking tools based on objective analysis principles for use with meteorological data. Many tools have been developed over the previous 2 decades that attempt to address this need but most have limitations on the type of data they can be used with, feature computational and/or memory expenses that make them unwieldy with larger datasets, or require some form of data reduction prior to use that limits the tool's utility. The Tracking and Object-Based Analysis of Clouds (tobac) Python package is a modular, open-source tool that improves on the overall generality and utility of past tools. A number of scientific improvements (three spatial dimensions, splits and mergers of features, an internal spectral filtering tool) and procedural enhancements (increased computational efficiency, internal regridding of data, and treatments for periodic boundary conditions) have been included in tobac as a part of the tobac v1.5 update. These improvements have made tobac one of the most robust, powerful, and flexible identification and tracking tools in our field to date and expand its potential use in other fields. Future plans for tobac v2 are also discussed.  more » « less
Award ID(s):
2019939
PAR ID:
10566810
Author(s) / Creator(s):
; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ;
Publisher / Repository:
Copernicus Publications on behalf of the European Geosciences Union
Date Published:
Journal Name:
Geoscientific Model Development
Volume:
17
Issue:
13
ISSN:
1991-9603
Page Range / eLocation ID:
5309 to 5330
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Computational methods have gained importance and popularity in both academic research and industrial applications in recent years. Since 2014, our team has consistently worked on reforming our Materials Science and Engineering curriculum at the University of Illinois Urbana-Champaign by incorporating computational modules into all mandatory undergraduate courses. Here, we investigate the impact and effectiveness of these computational modules in light of our recent graduates’ feedback. We surveyed alumni who graduated between 2017 – 2021 and asked them about the benefits of the computational curriculum and the significance of computation for their career. “data analysis” was reported to be the most significant computational practice, followed by “programming” and “simulation tools”. Python is the most prevalent programming language, and half of the respondents have reported to use it for their work. Particle based simulation tools are rarely used by our alumni, whereas continuum methods are more relevant, especially for alumni in industry. Graduates who pursued Ph.D. or Master’s degrees benefited more from the existence of computational modules and would also benefit the most from qualitative improvements of the modules. Alumni have reported limited benefits of computational modules during their job search, but note a slightly positive impact on their job performance. Overall, our Alumni think that the current amount of computational material in the curriculum is ideal, but further analysis indicates there is still room for qualitative improvements. We find the perspective provided by alumni to be a valuable tool to evaluate the computational reform of the MatSE curriculum at the University of Illinois Urbana-Champaign and it is a useful guide on how to reshape and improve its effectiveness further. 
    more » « less
  2. Abstract There are over 700 aquatic ecological assessment approaches across the globe that meet specific institutional goals. However, in many cases, multiple assessment tools are designed to meet the same management need, resulting in a confusing array of overlapping options. Here, we look at six riverine wetland assessments currently in use in Montana, USA, and ask which tool (1) best captures the condition across a disturbance gradient and (2) has the most utility to meet the regulatory or management needs. We used descriptive statistics to compare wetland assessments (n = 18) across a disturbance gradient determined by a landscape development intensity. Factor analysis showed that many of the tools had internal metrics that did not correspond well with overall results, hindering the tool’s ability to act as designed. We surveyed regional wetland managers (n = 56) to determine the extent of their use of each of the six tools and how well they trusted the information the assessment tool provided. We found that the Montana Wetland Assessment Methodology best measured the range of disturbance and had the highest utility to meet Clean Water Act (CWA§ 404) needs. Montana Department of Environmental Quality was best for the CWA§ 303(d) & 305(b) needs. The US Natural Resources Conservation Service’s Riparian Assessment Tool was the third most used by managers but was the tool that had the least ability to distinguish across a disturbance, followed by the US Bureau of Land Management’s Proper Functioning Condition. 
    more » « less
  3. Using new technology to provide automated feedback on classroom discourse offers a unique opportunity for educators to engage in self-reflection on their teaching, in particular to ensure that the instructional environment is equitable and productive for all students. More information is needed about how teachers experience automated data tools, including what they perceive as relevant and helpful for their everyday teaching. This mixed-methods study explored the perceptions and engagement of 21 math teachers over two years with a big data tool that analyzes classroom recordings and generates information about their discourse practices in near real-time. Findings revealed that teachers perceived the tool as having utility, yet the specific feedback that teachers perceived as most useful changed over time. In addition, teachers who used the tool throughout both years increased their use of talk moves over time, suggesting that they were making changes due to their review of the personalized feedback. These findings speak to promising directions for the development of AI-based, big data tools that help shape teacher learning and instruction, particularly tools that have strong perceived utility. 
    more » « less
  4. Wearable devices have become commonplace tools for tracking behavioral and physiological parameters in real-world settings. Nonetheless, the practical utility of these data for clinical and research applications, such as sleep analysis, is hindered by their noisy, large-scale, and multidimensional characteristics. Here, we develop a neural network algorithm that predicts sleep stages by tracking topological features (TFs) of wearable data and model-driven clock proxies (CPs) reflecting the circadian propensity for sleep. To evaluate its accuracy, we apply it to motion and heart rate data from the Apple Watch worn by young subjects undergoing polysomnography (PSG) and compare the predicted sleep stages with the corresponding ground truth PSG records. The neural network that includes TFs and CPs along with raw wearable data as inputs shows improved performance in classifying Wake/REM/NREM sleep. For example, it shows significant improvements in identifying REM and NREM sleep (AUROC/AUPRC improvements >13% and REM/NREM accuracy improvement of 12%) compared with the neural network using only raw data inputs. We find that this improvement is mainly attributed to the heart rate TFs. To further validate our algorithm on a different population, we test it on elderly subjects from the Multi-ethnic Study of Atherosclerosis cohort. This confirms that TFs and CPs contribute to the improvements in Wake/REM/NREM classification. We next compare the performance of our algorithm with previous state-of-the-art wearable-based sleep scoring algorithms and find that our algorithm outperforms them within and across different populations. This study demonstrates the benefits of combining topological data analysis and mathematical modeling to extract hidden inputs of neural networks from puzzling wearable data. 
    more » « less
  5. Abstract Estimating cell type composition of blood and tissue samples is a biological challenge relevant in both laboratory studies and clinical care. In recent years, a number of computational tools have been developed to estimate cell type abundance using gene expression data. Although these tools use a variety of approaches, they all leverage expression profiles from purified cell types to evaluate the cell type composition within samples. In this study, we compare 12 cell type quantification tools and evaluate their performance while using each of 10 separate reference profiles. Specifically, we have run each tool on over 4000 samples with known cell type proportions, spanning both immune and stromal cell types. A total of 12 of these represent in vitro synthetic mixtures and 300 represent in silico synthetic mixtures prepared using single-cell data. A final 3728 clinical samples have been collected from the Framingham cohort, for which cell populations have been quantified using electrical impedance cell counting. When tools are applied to the Framingham dataset, the tool Estimating the Proportions of Immune and Cancer cells (EPIC) produces the highest correlation, whereas Gene Expression Deconvolution Interactive Tool (GEDIT) produces the lowest error. The best tool for other datasets is varied, but CIBERSORT and GEDIT most consistently produce accurate results. We find that optimal reference depends on the tool used, and report suggested references to be used with each tool. Most tools return results within minutes, but on large datasets runtimes for CIBERSORT can exceed hours or even days. We conclude that deconvolution methods are capable of returning high-quality results, but that proper reference selection is critical. 
    more » « less