skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


This content will become publicly available on April 28, 2026

Title: Notification Designs for Influencing Hearing Speakers' Behaviors During Captioned Conversations Among Mixed DHH-Hearing Groups
Award ID(s):
1954284 2125362
PAR ID:
10649216
Author(s) / Creator(s):
; ;
Publisher / Repository:
ACM
Date Published:
Page Range / eLocation ID:
100 to 111
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract topics such as recursion are challenging for many computer science students to understand. In this experience report, we explore function sonification—the addition of sound to a function to communicate information about the function’s behavior in real-time as it runs—as a pedagogical approach for improving students’ understanding of recursion. We present several example iterative and recursive function sonifications, plus spectrograms that illustrate their different sonic behaviors. We also present experimental evidence that using these sonifications significantly improved the understanding of recursion for students who used them, compared to students who used silent (i.e., traditional) versions of the same functions. Based on these experiences, we believe sonification has under-appreciated potential for teaching abstract computing topics. 
    more » « less
  2. Human hearing is robust to noise, but the basis of this robustness is poorly understood. Several lines of evidence are consistent with the idea that the auditory system adapts to sound components that are stable over time, potentially achieving noise robustness by suppressing noise-like signals. Yet background noise often provides behaviorally relevant information about the environment and thus seems unlikely to be completely discarded by the auditory system. Motivated by this observation, we explored whether noise robustness might instead be mediated by internal models of noise structure that could facilitate the separation of background noise from other sounds. We found that detection, recognition, and localization in real-world background noise were better for foreground sounds positioned later in a noise excerpt, with performance improving over the initial second of exposure to a noise. These results are consistent with both adaptation-based and model-based accounts (adaptation increases over time and online noise estimation should benefit from acquiring more samples). However, performance was also robust to interruptions in the background noise and was enhanced for intermittently recurring backgrounds, neither of which would be expected from known forms of adaptation. Additionally, the performance benefit observed for foreground sounds occurring later within a noise excerpt was reduced for recurring noises, suggesting that a noise representation is built up during exposure to a new background noise and then maintained in memory. These findings suggest that noise robustness is supported by internal models—“noise schemas”—that are rapidly estimated, stored over time, and used to estimate other concurrent sounds. 
    more » « less
  3. With SAE Level 3 automated vehicles handling most driving tasks, there are still situations when the driver needs to take over. Multimodal displays have been introduced to inform drivers of the need to take over for critical scenarios (e.g., in construction zones) in instructional or informative formats. However, the effects of multimodal displays on takeover performance for drivers with hearing impairments are still unclear. Therefore, the goal of this study was to investigate how signal type (single tactile (T), single visual (V), and visual and tactile combined (VT)), information type (instructional, informative, and baseline), and hearing impairment (hearing-impaired and non-hearing-impaired drivers) affect drivers’ takeover performance. Findings show that signal type significantly influenced reaction and takeover times, with multimodal signals (VT) resulting in faster reactions compared to single modal signals. Additionally, the baseline condition yielded the faster reaction times compared to both instructional and informative formats. Hearing impairment, however, did not significantly affect reaction and takeover times. Findings may inform the development of future vehicle interfaces to assist drivers with hearing impairments. 
    more » « less