skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Is a Classification Procedure Good Enough?—A Goodness-of-Fit Assessment Tool for Classification Learning
Award ID(s):
2038603
PAR ID:
10347493
Author(s) / Creator(s):
; ;
Date Published:
Journal Name:
Journal of the American Statistical Association
ISSN:
0162-1459
Page Range / eLocation ID:
1 to 11
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract For which choices of$$X,Y,Z\in \{\Sigma ^1_1,\Pi ^1_1\}$$does no sufficiently strongX-sound andY-definable extension theory prove its ownZ-soundness? We give a complete answer, thereby delimiting the generalizations of Gödel’s second incompleteness theorem that hold within second-order arithmetic. 
    more » « less
  2. Hyperdimensional (HD) computing is built upon its unique data type referred to as hypervectors. The dimension of these hypervectors is typically in the range of tens of thousands. Proposed to solve cognitive tasks, HD computing aims at calculating similarity among its data. Data transformation is realized by three operations, including addition, multiplication and permutation. Its ultra-wide data representation introduces redundancy against noise. Since information is evenly distributed over every bit of the hypervectors, HD computing is inherently robust. Additionally, due to the nature of those three operations, HD computing leads to fast learning ability, high energy efficiency and acceptable accuracy in learning and classification tasks. This paper introduces the background of HD computing, and reviews the data representation, data transformation, and similarity measurement. The orthogonality in high dimensions presents opportunities for flexible computing. To balance the tradeoff between accuracy and efficiency, strategies include but are not limited to encoding, retraining, binarization and hardware acceleration. Evaluations indicate that HD computing shows great potential in addressing problems using data in the form of letters, signals and images. HD computing especially shows significant promise to replace machine learning algorithms as a light-weight classifier in the field of internet of things (IoTs). 
    more » « less
  3. In the digital communication age, using social media data to classify first responders presents a new and promising approach to enhancing emergency response strategies. We introduce the First Responder Classification System (FReCS), a framework that annotates and classifies disaster tweets from 26 crisis events. Our annotations cater for first reponders and their sub-layers. Furthermore, we proposed a classifier called RoBERTa-CAFÉ that integrates pre-trained RoBERTa with Cross-Attention and Focused-Entanglement components, improving the precision and reliability of classification tasks. The model is rigorously tested across publicly available disaster datasets. The RoBERTa-CAFÉ model outperformed state-of-the-art models in identifying relevant emergency communications, displaying its generalization, robustness, and adaptability. Our FReCS approach offers a pioneering technique for classifying first responders and enhances emergency management systems’ operational capabilities, leading to more efficient and effective disaster responses. 
    more » « less