Abstract Alfvén eigenmodes (AEs) are an important and complex class of plasma dynamics commonly observed in tokamaks and other plasma devices. In this work, we manually labeled a small database of 26 discharges from the DIII-D tokamak in order to train simple neural-network-based models for classifying AEs. The models provide spatiotemporally local identification of four types of AEs by using an array of 40 electron cyclotron emission (ECE) signals as inputs. Despite the minimal dataset, this strategy performs well at spatiotemporally localized classification of AEs, indicating future opportunities for more sophisticated models and incorporation into real-time control strategies. The trained model is then used to generate spatiotemporally-resolved labels for each of the 40 ECE measurements on a much larger database of 1112 DIII-D discharges. This large set of precision labels can be used in future studies for advanced deep predictors and new physical insights.
more »
« less
Towards Extracting and Understanding the Implicit Rubrics of Transformer Based Automatic Essay Scoring Models
By aligning the functional components derived from the activations of transformer models trained for AES with external knowledge such as human-understandable feature groups, the proposed method improves the interpretability of a Longformer Automatic Essay Scoring (AES) system and provides tools for performing such analyses on further neural AES systems. The analysis focuses on models trained to score essays based on organization, main idea, support, and language. The findings provide insights into the models’ decision-making processes, biases, and limitations, contributing to the development of more transparent and reliable AES systems.
more »
« less
- PAR ID:
- 10510436
- Editor(s):
- Kochmar, Ekaterina; Burstein, Jill; Horbach, Andrea; Laarmann-Quante, Ronja; Madnani, Nitin; Tack, Anaïs; Yaneva, Victoria; Yuan, Zheng; Zesch, Torsten
- Publisher / Repository:
- Association for Computational Linguistics
- Date Published:
- Journal Name:
- Proceedings of the 18th Workshop on Innovative Use of NLP for Building Educational Applications
- Page Range / eLocation ID:
- 232 to 241
- Format(s):
- Medium: X
- Location:
- Toronto, Canada
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Despite many attempts, the state-of-the-art of adversarial machine learning on malware detection systems generally yield unexecutable samples. In this work, we set out to examine the robustness of visualization-based malware detection system against adversarial examples (AEs) that not only are able to fool the model, but also maintain the executability of the original input. As such, we first investigate the application of existing off-the-shelf adversarial attack approaches on malware detection systems through which we found that those approaches do not necessarily maintain the functionality of the original inputs. Therefore, we proposed an approach to generate adversarial examples, COPYCAT, which is specifically designed for malware detection systems considering two main goals; achieving a high misclassification rate and maintaining the executability and functionality of the original input. We designed two main configurations for COPYCAT, namely AE padding and sample injection. While the first configuration results in untargeted misclassification attacks, the sample injection configuration is able to force the model to generate a targeted output, which is highly desirable in the malware attribution setting. We evaluate the performance of COPYCAT through an extensive set of experiments on two malware datasets, and report that we were able to generate adversarial samples that are misclassified at a rate of 98.9% and 96.5% with Windows and IoT binary datasets, respectively, outperforming the misclassification rates in the literature. Most importantly, we report that those AEs were executable unlike AEs generated by off-the-shelf approaches. Our transferability study demonstrates that the generated AEs through our proposed method can be generalized to other models.more » « less
-
Complex events (CEs) play a crucial role in CPS-IoT applications, enabling high-level decision-making in domains such as smart monitoring and autonomous systems. However, most existing models focus on short-span perception tasks, lacking the long-term reasoning required for CE detection. CEs consist of sequences of short-time atomic events (AEs) governed by spatiotemporal dependencies. Detecting them is difficult due to long, noisy sensor data and the challenge of filtering out irrelevant AEs while capturing meaningful patterns. This work explores CE detection as a case study for CPS-IoT foundation models capable of long-term reasoning. We evaluate three approaches: (1) leveraging large language models (LLMs), (2) employing various neural architectures that learn CE rules from data, and (3) adopting a neurosymbolic approach that integrates neural models with symbolic engines embedding human knowledge. Our results show that the state-space model, Mamba, which belongs to the second category, outperforms all methods in accuracy and generalization to longer, unseen sensor traces. These findings suggest that state-space models could be a strong backbone for CPS-IoT foundation models for long-span reasoning tasks.more » « less
-
IntroductionBiomechanical changes due to aging increase the oxygen consumption of walking by over 30%. When this is coupled with reduced oxygen uptake capacity, the ability to sustain walking becomes compromised. This reduced physical activity and mobility can lead to further physical degeneration and mortality. Unfortunately, the underlying reasons for the increased metabolic cost are still inadequately understood. While motion capture systems can measure signals with high temporal resolution, it is impossible to directly characterize the fluctuation of metabolic cost throughout the gait cycle. MethodsTo address this issue, this research focuses on computing the metabolic cost time series from the mean value using two neural-network-based approaches: autoencoders (AEs) and expanders. For the AEs, the encoders are designed to compress the input time series down to their mean value, and the decoder expands those values into the time series. After training, the decoder is extracted and applied to mean metabolic cost values to compute the time series. A second approach leverages an expander to map the mean values to the time series without an encoder. The networks are trained using ten different metabolic cost models generated by a computational walking model that simulates the gait cycle subjected to 35 different robotic perturbations without using experimental input data. The networks are validated using the estimated metabolic costs for the unperturbed gait cycle. ResultsThe investigation found that AEs without tied weights and the expanders performed best using nonlinear activation functions, while the AEs with tied weights performed best with linear activation functions. Unexpectedly, the results show that the expanders outperform the AEs. DiscussionA limitation of this research is the reliance on time series for the initial training. Future efforts will focus on developing methods that overcome this issue. Improved methods for estimating within-stride fluctuations in metabolic cost have the potential of improving rehabilitation and assistive devices by targeting the gait phases with increased metabolic cost. This research could also be applied to expand sparse measurements to locations or times that were not measured explicitly. This application would reduce the number of measurement points required to capture the response of a system.more » « less
-
Recent work in adversarial machine learning started to focus on the visual perception in autonomous driving and studied Adversarial Examples (AEs) for object detection models. However, in such visual perception pipeline the detected objects must also be tracked, in a process called Multiple Object Tracking (MOT), to build the moving trajectories of surrounding obstacles. Since MOT is designed to be robust against errors in object detection, it poses a general challenge to existing attack techniques that blindly target objection detection: we find that a success rate of over 98% is needed for them to actually affect the tracking results, a requirement that no existing attack technique can satisfy. In this paper, we are the first to study adversarial machine learning attacks against the complete visual perception pipeline in autonomous driving, and discover a novel attack technique, tracker hijacking, that can effectively fool MOT using AEs on object detection. Using our technique, successful AEs on as few as one single frame can move an existing object in to or out of the headway of an autonomous vehicle to cause potential safety hazards. We perform evaluation using the Berkeley Deep Drive dataset and find that on average when 3 frames are attacked, our attack can have a nearly 100% success rate while attacks that blindly target object detection only have up to 25%.more » « less
An official website of the United States government

