Conference Paper

The effects of semantic grouping on visual search

DOI: 10.1145/1358628.1358876 Conference: Extended Abstracts Proceedings of the 2008 Conference on Human Factors in Computing Systems, CHI 2008, Florence, Italy, April 5-10, 2008
Source: DBLP


This paper reports on work-in-progress to better understand how users visually interact with hierarchically organized semantic information. Experimental reaction time and eye movement data are reported that give insight into strategies people employ while searching visual layouts containing words that are either grouped by category (i.e. semantically cohesive) or randomly grouped. Additionally, sometimes the category labels of the cohesive groups are displayed as part of the group. Preliminary results suggest that: (a) When groups are cohesive, people tend to search labeled and unlabeled layouts similarly. (b) People seem to trust the categorical information of labels more than non-labels. This work will be used to extend current computational models of visual search to better predict users visual interaction with interfaces.

Download full-text


Available from: Tim Halverson, Jul 03, 2015
  • Source
    • "Also, observations made from eye-gaze statistics in [5] suggest that humans are attentive to interesting objects in semantically rich photographs. Eye gaze measurements have been employed for modeling user attention in a number of applications including visual search for Human-Computer Interaction (HCI) [7] and open signed video analysis [3]. [9] employs low-level image features (contrast, intensity etc.) for computing a saliency map to predict human gaze. "
    [Show abstract] [Hide abstract]
    ABSTRACT: ABSTRACT We propose a novel framework to localize and label affec- tive objects and actions in images through a combination of text, visual and gaze-based analysis. Human gaze pro- vides useful cues to infer locations and interactions of af- fective objects. While concepts (labels) associated with an image can be determined from its caption, we demonstrate localization of these concepts upon learning from a statisti- cal affect model for world concepts. The affect model is derived from non-invasively acquired fixation patterns on labeled images, and guides localization of affective objects (faces, reptiles ) and actions (look, read ) from fixations in un- labeled images. Experimental results obtained on a database of 500 images confirm the effectiveness and promise of the proposed approach. Categories and Subject Descriptors: H.4 [Information Systems Applications]: Multimedia Application General Terms: Human Factors, Algorithms. Keywords: Automated localization and labeling, caption text-cum-eye gaze analysis, affect model for world concepts, statistical model.
    Full-text · Conference Paper · Oct 2009
  • [Show abstract] [Hide abstract]
    ABSTRACT: How do people efficiently locate content in a display? We investigate the effect of text layout on how people decide which area of a display to search first. Using a visual search paradigm, participants were required to locate a known target within a two-column display, in which items were grouped into semantic clusters, and the physical distance between items varied. For 'mixed' trials, the distance between items in each column was varied. Results showed that participants preferred to search the sparser of the two columns first, even though they were faster at locating the target in the denser column. This finding suggests that participants were adopting an inefficient search strategy for locating the target item. Discussion focuses on the implications for models that assume people rationally adapt their search strategy to maximize the gain of task-relevant information over time. Copyright 2010 by Human Factors and Ergonomics Society, Inc. All rights reserved.
    No preview · Article · Sep 2010 · Human Factors and Ergonomics Society Annual Meeting Proceedings
  • [Show abstract] [Hide abstract]
    ABSTRACT: Current models of web navigation focus only on the influence of textual information and ignore the role of graphical information. We studied the differential role of text and graphics in identifying web page widgets classified into two kinds: textual and graphical. Four different versions of web pages were created by systematically removing textual and graphical information from each page. The task of the participants was to locate either textual or graphical widgets on the displayed web page. Results show that for any widget, the task-completion time and the number of clicks were significantly less in web pages with graphics than in those with no graphics. This demonstrates the importance of graphical information. However, textual information is also important because performance in locating graphical widgets under no-graphics conditions was better when text was present than with no text. Since, for identifying graphical widgets, text and graphics interact and complement each other, we conclude that cognitive models on web navigation should include the role of graphical information next to textual information.
    No preview · Article · Jan 2012 · Behaviour and Information Technology
Show more