skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Examining Limits of Small Multiples: Frame Quantity Impacts Judgments with Line Graphs
Small multiples are a popular visualization method, displaying different views of a dataset using multiple frames, often with the same scale and axes. However, there is a need to address their potential constraints, especially in the context of human cognitive capacity limits. These limits dictate the maximum information our mind can process at once. We explore the issue of capacity limitation by testing competing theories that describe how the number of frames shown in a display, the scale of the frames, and time constraints impact user performance with small multiples of line charts in an energy grid scenario. In two online studies (Experiment 1 n = 141 and Experiment 2 n = 360) and a follow-up eye-tracking analysis (n = 5), we found a linear decline in accuracy with increasing frames across seven tasks, which was not fully explained by differences in frame size, suggesting visual search challenges. Moreover, the studies demonstrate that highlighting specific frames can mitigate some visual search difficulties but, surprisingly, not eliminate them. This research offers insights into optimizing the utility of small multiples by aligning them with human limitations.  more » « less
Award ID(s):
2238175
PAR ID:
10503942
Author(s) / Creator(s):
; ; ; ;
Publisher / Repository:
IEEE Transactions on Visualization and Computer Graphics
Date Published:
Journal Name:
IEEE Transactions on Visualization and Computer Graphics
ISSN:
1077-2626
Page Range / eLocation ID:
1 to 12
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Humans and other animals are able to perceive and represent a number of objects present in a scene, a core cognitive ability thought to underlie the development of mathematics. However, the perceptual mechanisms that underpin this capacity remain poorly understood. Here, we show that our visual sense of number derives from a visual system designed to efficiently encode the location of objects in scenes. Using a mathematical model, we demonstrate that an efficient but information-limited encoding of objects’ locations can explain many key aspects of number psychophysics, including subitizing, Weber’s law, underestimation, and effects of exposure time. In two experiments (N = 100 each), we find that this model of visual encoding captures human performance in both a change-localization task and a number estimation task. In a third experiment (N = 100), we find that individual differences in change-localization performance are highly predictive of differences in number estimation, both in terms of overall performance and inferred model parameters, with participants having numerically indistinguishable inferred information capacities across tasks. Our results therefore indicate that key psychophysical features of numerical cognition do not arise from separate modules or capacities specific to number, but rather as by-products of lower level constraints on perception. 
    more » « less
  2. Language-guided smart systems can help to design next-generation human-machine interactive applications. The dense text description is one of the research areas where systems learn the semantic knowledge and visual features of each video frame and map them to describe the video's most relevant subjects and events. In this paper, we consider untrimmed sports videos as our case study. Generating dense descriptions in the sports domain to supplement journalistic works without relying on commentators and experts requires more investigation. Motivated by this, we propose an end-to-end automated text-generator, SpecTextor, that learns the semantic features from untrimmed videos of sports games and generates associated descriptive texts. The proposed approach considers the video as a sequence of frames and sequentially generates words. After splitting videos into frames, we use a pre-trained VGG-16 model for feature extraction and encoding the video frames. With these encoded frames, we posit a Long Short-Term Memory (LSTM) based attention-decoder pipeline that leverages soft-attention mechanism to map the semantic features with relevant textual descriptions to generate the explanation of the game. Because developing a comprehensive description of the game warrants training on a set of dense time-stamped captions, we leverage two available public datasets: ActivityNet Captions and Microsoft Video Description. In addition, we utilized two different decoding algorithms: beam search and greedy search and computed two evaluation metrics: BLEU and METEOR scores. 
    more » « less
  3. Nowadays, to assess and document construction and building performance, large amount of visual data are captured and stored through camera equipped platforms such as wearable cameras, unmanned aerial/ground vehicles, and smart phones. However, due to the nonstop fashion in recording such visual data, not all of the frames in captured consecutive footages are intentionally taken, and thus not every frame is worthy of being processed for construction and building performance analysis. Since many frames will simply have non-construction related contents, before processing the visual data, the content of each recorded frame should be manually investigated depending on the association with the goal of the visual assessment. To address such challenges, this paper aims to automatically filter construction big visual data that requires no human annotations. To overcome challenges in pure discriminative approach using manually labeled images, we construct a generative model with unlabeled visual dataset, and use it to find construction-related frames in big visual dataset from jobsites. First, through composition-based snap point detection together with domain adaptation, we filter and remove most of accidently recorded frames in the footage. Then, we create discriminative classifier trained with visual data from jobsites to eliminate non-construction related images. To evaluate the reliability of the proposed method, we have obtained the ground truth based on human judgment for each photo in our testing dataset. Despite learning without any explicit labels, the proposed method shows a reasonable practical range of accuracy, which generally outperforms prior snap point detection. Through the case studies, the fidelity of the algorithm is discussed in detail. By being able to focus on selective visual data, practitioners will spend less time on browsing large amounts of visual data; rather spend more time on looking at how to leverage the visual data to facilitate decision-makings in built environments. 
    more » « less
  4. Understanding how to scale up effects of biological diversity on ecosystem functioning and services remains challenging. There is a general consensus that biodiversity loss alters ecosystem processes underpinning the goods and services upon which humanity depends. Yet, most of that consensus stems from experiments performed at small spatial-scales for short time-frames, which limits transferability of conclusions to longer-term, landscape-scale conservation policies and management. Here we develop quantitative scaling relationships linking 374 experiments that tested plant diversity effects on biomass production across a range of scales. We show that biodiversity effects increase by factors of 1.68 and 1.10 for each 10-fold increase in experiment temporal and spatial scales, respectively. Contrary to prior studies, our analyses suggest that the time scale of experiments, rather than their spatial scale, is the primary source of variation in biodiversity effects. But consistent with earlier research, our analyses reveal that complementarity effects, rather than selection effects, drive the positive space-time interactions for plant diversity effects. Importantly, we also demonstrate complex space-time interactions and nonlinear responses that emphasize how simple extrapolations from small-scale experiments are likely to underestimate biodiversity effects in real-world ecosystems. Quantitative scaling relationships from this research are a crucial step towards bridging controlled experiments that identify biological mechanisms across a range of scales. Predictions from scaling relationships like these could then be compared with observations for fine-tuning the relationships and ultimately improving their capacities to predict consequences of biodiversity loss for ecosystem functioning and services over longer time-frames across real-world landscapes. 
    more » « less
  5. With the increasing reliance on small Unmanned Aerial Systems (sUAS) for Emergency Response Scenarios, such as Search and Rescue, the integration of computer vision capabilities has become a key factor in mission success. Nevertheless, computer vision performance for detecting humans severely degrades when shifting from ground to aerial views. Several aerial datasets have been created to mitigate this problem, however, none of them has specifically addressed the issue of occlusion, a critical component in Emergency Response Scenarios. Natural, Occluded, Multi-scale Aerial Dataset (NOMAD) presents a benchmark for human detection under occluded aerial views, with five different aerial distances and rich imagery variance. NOMAD is composed of 100 different Actors, all performing sequences of walking, laying and hiding. It includes 42,825 frames, extracted from 5.4k resolution videos, and manually annotated with a bounding box and a label describing 10 different visibility levels, categorized according to the percentage of the human body visible inside the bounding box. This allows computer vision models to be evaluated on their detection performance across different ranges of occlusion. NOMAD is designed to improve the effectiveness of aerial search and rescue and to enhance collaboration between sUAS and humans, by providing a new benchmark dataset for human detection under occluded aerial views. 
    more » « less