skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Search for: All records

Award ID contains: 2323979

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. Dimension reduction (DR) algorithms have proven to be extremely useful for gaining insight into large-scale high-dimensional datasets, particularly finding clusters in transcriptomic data. The initial phase of these DR methods often involves converting the original high-dimensional data into a graph. In this graph, each edge represents the similarity or dissimilarity between pairs of data points. However, this graph is frequently suboptimal due to unreliable high-dimensional distances and the limited information extracted from the high-dimensional data. This problem is exacerbated as the dataset size increases. If we reduce the size of the dataset by selecting points for a specific sections of the embeddings, the clusters observed through DR are more separable since the extracted subgraphs are more reliable. In this paper, we introduce LocalMAP, a new dimensionality reduction algorithm that dynamically and locally adjusts the graph to address this challenge. By dynamically extracting subgraphs and updating the graph on-the-fly, LocalMAP is capable of identifying and separating real clusters within the data that other DR methods may overlook or combine. We demonstrate the benefits of LocalMAP through a case study on biological datasets, highlighting its utility in helping users more accurately identify clusters for real-world problems. 
    more » « less
    Free, publicly-accessible full text available April 11, 2026
  2. Unaltered biological matter (biomatter) can be harnessed to fabricate cohesive, sustainable bioplastics. However, controlling the material properties of these bioplastics is challenging, as the contributions of different macromolecular building blocks to processability and performance are unknown. To deconvolute the roles of different classes of biomolecules, we developed experimental and computational methods to construct and analyze biomatter analogs composed of carbohydrates, proteins, and lipids. These analogs are intended to improve fundamental understanding of biomatter plastics. Spectroscopic analyses of biomatter analogs suggest that cohesion depends on protein aggregation during thermomechanical processing. Molecular dynamics simulations confirm that alterations to protein conformation and hydrogen bonding are likely the primary mechanisms underlying the formation of a cohesive, proteinaceous matrix. Simulations also corroborate experimental measurements highlighting the importance of hydrogen bonding and self-assembly between specific, small-molecule constituents. These conclusions may enable the engineering of next-generation biomatter plastics with improved performance. 
    more » « less
    Free, publicly-accessible full text available March 1, 2026
  3. Sparsity is a central aspect of interpretability in machine learning. Typically, sparsity is measured in terms of the size of a model globally, such as the number of variables it uses. However, this notion of sparsity is not particularly relevant for decision-making; someone subjected to a decision does not care about variables that do not contribute to the decision. In this work, we dramatically expand a notion of decision sparsity called the Sparse Explanation Value(SEV) so that its explanations are more meaningful. SEV considers movement along a hypercube towards a reference point. By allowing flexibility in that reference and by considering how distances along the hypercube translate to distances in feature space, we can derive sparser and more meaningful explanations for various types of function classes. We present cluster-based SEV and its variant tree-based SEV, introduce a method that improves credibility of explanations, and propose algorithms that optimize decision sparsity in machine learning models. 
    more » « less
    Free, publicly-accessible full text available December 1, 2025