skip to main content

Attention:

The NSF Public Access Repository (PAR) system and access will be unavailable from 11:00 PM ET on Thursday, January 16 until 2:00 AM ET on Friday, January 17 due to maintenance. We apologize for the inconvenience.


Title: Computational modeling of human reasoning processes for interpretable visual knowledge: a case study with radiographers
Abstract

Visual reasoning is critical in many complex visual tasks in medicine such as radiology or pathology. It is challenging to explicitly explain reasoning processes due to the dynamic nature of real-time human cognition. A deeper understanding of such reasoning processes is necessary for improving diagnostic accuracy and computational tools. Most computational analysis methods for visual attention utilize black-box algorithms which lack explainability and are therefore limited in understanding the visual reasoning processes. In this paper, we propose a computational method to quantify and dissect visual reasoning. The method characterizes spatial and temporal features and identifies common and contrast visual reasoning patterns to extract significant gaze activities. The visual reasoning patterns are explainable and can be compared among different groups to discover strategy differences. Experiments with radiographers of varied levels of expertise on 10 levels of visual tasks were conducted. Our empirical observations show that the method can capture the temporal and spatial features of human visual attention and distinguish expertise level. The extracted patterns are further examined and interpreted to showcase key differences between expertise levels in the visual reasoning processes. By revealing task-related reasoning processes, this method demonstrates potential for explaining human visual understanding.

 
more » « less
PAR ID:
10360741
Author(s) / Creator(s):
; ; ; ; ;
Publisher / Repository:
Nature Publishing Group
Date Published:
Journal Name:
Scientific Reports
Volume:
10
Issue:
1
ISSN:
2045-2322
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Human visual grouping processes consolidate independent visual objects into grouped visual features on the basis of shared characteristics; these visual features can themselves be grouped, resulting in a hierarchical representation of visual grouping information. This “grouping hierarchy“ promotes ef- ficient attention in the support of goal-directed behavior, but improper grouping of elements of a visual scene can also re- sult in critical behavioral errors. Understanding of how visual object/features characteristics such as size and form influences perception of hierarchical visual groups can further theory of human visual grouping behavior and contribute to effective in- terface design. In the present study, participants provided free- response groupings of a set of stimuli that contained consistent structural relationships between a limited set of visual features. These grouping patterns were evaluated for relationships be- tween specific characteristics of the constituent visual features and the distribution of features across levels of the indicated grouping hierarchy. We observed that while the relative size of the visual features differentiated groupings across levels of the grouping hierarchy, the form of visual objects and features was more likely to distinguish separate groups within a partic- ular level of hierarchy. These consistent relationships between visual feature characteristics and placement within a grouping hierarchy can be leveraged to advance computational theories of human visual grouping behavior, which can in turn be ap- plied to effective design for interfaces such as voter ballots. 
    more » « less
  2. Human visual grouping processes consolidate independent visual objects into grouped visual features on the basis of shared characteristics; these visual features can themselves be grouped, resulting in a hierarchical representation of visual grouping information. This “grouping hierarchy“ promotes ef- ficient attention in the support of goal-directed behavior, but improper grouping of elements of a visual scene can also re- sult in critical behavioral errors. Understanding of how visual object/features characteristics such as size and form influences perception of hierarchical visual groups can further theory of human visual grouping behavior and contribute to effective in- terface design. In the present study, participants provided free- response groupings of a set of stimuli that contained consistent structural relationships between a limited set of visual features. These grouping patterns were evaluated for relationships be- tween specific characteristics of the constituent visual features and the distribution of features across levels of the indicated grouping hierarchy. We observed that while the relative size of the visual features differentiated groupings across levels of the grouping hierarchy, the form of visual objects and features was more likely to distinguish separate groups within a partic- ular level of hierarchy. These consistent relationships between visual feature characteristics and placement within a grouping hierarchy can be leveraged to advance computational theories of human visual grouping behavior, which can in turn be ap- plied to effective design for interfaces such as voter ballots. 
    more » « less
  3. Human navigation simulation is critical to many civil engineering tasks and is of central interest to the simulation community. Most human navigation simulation approaches focus on the classic psychology evidence, or assumptions that still need further proofs. The overly simplified and generalized assumption of navigation behaviors does not highlight the need of capturing individual differences in spatial cognition and navigation decision-making, or the impacts of diverse ways of spatial information display. This study proposes the visual attention patterns in floorplan review to be a stronger predictor of human navigation behaviors. To set the theoretical foundation, a Virtual Reality (VR) experiment was performed to test if visual attention patterns during spatial information review can predict the quality of spatial memory, and how the relationship is affected by the diverse ways of information display, including 2D, 3D and VR. The results set a basis for future prediction model developments. 
    more » « less
  4. Human navigation simulation is critical to many civil engineering tasks and is of central interest to the simulation community. Most human navigation simulation approaches focus on the classic psychology evidence, or assumptions that still need further proofs. The overly simplified and generalized assumption of navigation behaviors does not highlight the need of capturing individual differences in spatial cognition and navigation decision-making, or the impacts of diverse ways of spatial information display. This study proposes the visual attention patterns in floorplan review to be a stronger predictor of human navigation behaviors. To set the theoretical foundation, a Virtual Reality (VR) experiment was performed to test if visual attention patterns during spatial information review can predict the quality of spatial memory, and how the relationship is affected by the diverse ways of information display, including 2D, 3D and VR. The results set a basis for future prediction model developments. 
    more » « less
  5. Abstract

    Qualitative spatial/temporal reasoning (QSR/QTR) plays a key role in research on human cognition, e.g., as it relates to navigation, as well as in work on robotics and artificial intelligence. Although previous work has mainly focused on various spatial and temporal calculi, more recently representation learning techniques such as embedding have been applied to reasoning and inference tasks such as query answering and knowledge base completion. These subsymbolic and learnable representations are well suited for handling noise and efficiency problems that plagued prior work. However, applying embedding techniques to spatial and temporal reasoning has received little attention to date. In this paper, we explore two research questions: (1) How do embedding-based methods perform empirically compared to traditional reasoning methods on QSR/QTR problems? (2) If the embedding-based methods are better, what causes this superiority? In order to answer these questions, we first propose a hyperbolic embedding model, called HyperQuaternionE, to capture varying properties of relations (such as symmetry and anti-symmetry), to learn inversion relations and relation compositions (i.e., composition tables), and to model hierarchical structures over entities induced by transitive relations. We conduct various experiments on two synthetic datasets to demonstrate the advantages of our proposed embedding-based method against existing embedding models as well as traditional reasoners with respect to entity inference and relation inference. Additionally, our qualitative analysis reveals that our method is able to learn conceptual neighborhoods implicitly. We conclude that the success of our method is attributed to its ability to model composition tables and learn conceptual neighbors, which are among the core building blocks of QSR/QTR.

     
    more » « less