Conference Paper

Pick-by-Vision: A First Stress Test

Tech. Univ. Munchen, Munich, Germany
DOI: 10.1109/ISMAR.2009.5336484 Conference: Mixed and Augmented Reality, 2009. ISMAR 2009. 8th IEEE International Symposium on
Source: DBLP

ABSTRACT In this paper we report on our ongoing studies around the application of augmented reality methods to support the order picking process of logistics applications. Order picking is the gathering of goods out of a prepared range of items following some customer orders. We named the visual support of this order picking process using head-mounted displays ldquopick-by-visionrdquo. This work presents the case study of bringing our previously developed pick-by-vision system from the lab to an experimental factory hall to evaluate it under more realistic conditions. This includes the execution of two user studies. In the first one we compared our pick-by-vision system with and without tracking to picking using a paper list to check picking performance and quality in general. In a second test we had subjects using the pick-by-vision system continuously for two hours to gain in-depth insight into the longer use of our system, checking user strain besides the general performance. Furthermore, we report on the general obstacles of trying to use HMD-based AR in an industrial setup and discuss our observations of user behaviour.

  • [Show abstract] [Hide abstract]
    ABSTRACT: There has been a rapid increase in research evaluating usability of Augmented Reality (AR) systems in recent years. Although many different styles of evaluation are used, there is no clear consensus on the most relevant approaches. We report a review of papers published in International Symposium of Mixed and Augmented Reality (ISMAR) proceedings in the past decade, building on the previous work of Swan and Gabbard (2005). Firstly, we investigate the evaluation goal, measurement and method of ISMAR papers according to their usability research in four categories: performance, perception and cognition, collaboration and User Experience (UX). Secondly, we consider the balance of evaluation approaches with regard to empirical-analytical, quantitative-qualitative and participant demographics. Finally we identify potential emphases for usability study of AR systems in the future. These analyses provide a reference point for current evaluation techniques, trends and challenges, which benefit researchers intending to design, conduct and interpret usability evaluations for future AR systems.
    Interacting with Computers 11/2012; 24(6):450-460. DOI:10.1016/j.intcom.2012.07.004 · 0.73 Impact Factor
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: This paper presents the design and evaluation of ProFi, a PROduct FInding assistant in a supermarket scenario. We explore the idea of micro-navigation in supermarkets and aim at enhancing visual search processes in front of a shelf. In order to assess the concept, a prototype is built combining visual recognition techniques with an Augmented Reality interface. Two AR patterns (circle and spotlight) are designed to highlight target products. The prototype is formally evaluated in a controlled environment. Quantitative and qualitative data is collected to evaluate the usability and user preference. The results show that ProFi significantly improves the users' product finding performance, especially when using the circle pattern, and that ProFi is well accepted by users.
    Proceedings of the 2013 ACM conference on Pervasive and ubiquitous computing adjunct publication; 09/2013
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: In this paper, a mobile assistance-system is described which supports users in performing manual working tasks in the context of assembling complex products. The assistance system contains a head-worn display for the visualization of information relevant for the workflow as well as a video camera to acquire the scene. This paper is focused on the interaction of the user with this system and describes work in progress and initial results from an industrial application scenario. We present image-based methods for robust recognition of static and dynamic hand gestures in realtime. These methods are used for an intuitive interaction with the assistance-system. The segmentation of the hand based on color information builds the basis of feature extraction for static and dynamic gestures. For the static gestures, the activation of particular sensitive regions in the camera image by the user’s hand is used for interaction. An HMM classifier is used to extract dynamic gestures depending on motion parameters determined based on the optical flow in the camera image.
    Journal of Intelligent Learning Systems and Applications 08/2014; 6(3):141-152. DOI:10.4236/jilsa.2014.63011

Full-text (2 Sources)

Available from
May 23, 2014