Abstract For which choices of$$X,Y,Z\in \{\Sigma ^1_1,\Pi ^1_1\}$$does no sufficiently strongX-sound andY-definable extension theory prove its ownZ-soundness? We give a complete answer, thereby delimiting the generalizations of Gödel’s second incompleteness theorem that hold within second-order arithmetic.
more »
« less
Is a Classification Procedure Good Enough?—A Goodness-of-Fit Assessment Tool for Classification Learning
- Award ID(s):
- 2038603
- PAR ID:
- 10347493
- Date Published:
- Journal Name:
- Journal of the American Statistical Association
- ISSN:
- 0162-1459
- Page Range / eLocation ID:
- 1 to 11
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Hyperdimensional (HD) computing is built upon its unique data type referred to as hypervectors. The dimension of these hypervectors is typically in the range of tens of thousands. Proposed to solve cognitive tasks, HD computing aims at calculating similarity among its data. Data transformation is realized by three operations, including addition, multiplication and permutation. Its ultra-wide data representation introduces redundancy against noise. Since information is evenly distributed over every bit of the hypervectors, HD computing is inherently robust. Additionally, due to the nature of those three operations, HD computing leads to fast learning ability, high energy efficiency and acceptable accuracy in learning and classification tasks. This paper introduces the background of HD computing, and reviews the data representation, data transformation, and similarity measurement. The orthogonality in high dimensions presents opportunities for flexible computing. To balance the tradeoff between accuracy and efficiency, strategies include but are not limited to encoding, retraining, binarization and hardware acceleration. Evaluations indicate that HD computing shows great potential in addressing problems using data in the form of letters, signals and images. HD computing especially shows significant promise to replace machine learning algorithms as a light-weight classifier in the field of internet of things (IoTs).more » « less
-
In the digital communication age, using social media data to classify first responders presents a new and promising approach to enhancing emergency response strategies. We introduce the First Responder Classification System (FReCS), a framework that annotates and classifies disaster tweets from 26 crisis events. Our annotations cater for first reponders and their sub-layers. Furthermore, we proposed a classifier called RoBERTa-CAFÉ that integrates pre-trained RoBERTa with Cross-Attention and Focused-Entanglement components, improving the precision and reliability of classification tasks. The model is rigorously tested across publicly available disaster datasets. The RoBERTa-CAFÉ model outperformed state-of-the-art models in identifying relevant emergency communications, displaying its generalization, robustness, and adaptability. Our FReCS approach offers a pioneering technique for classifying first responders and enhances emergency management systems’ operational capabilities, leading to more efficient and effective disaster responses.more » « less
An official website of the United States government

