Conference Paper

Accuracy Improvement and Objective Evaluation of Annotation Extraction from Printed Documents

DOI: 10.1109/DAS.2008.80 Conference: The Eighth IAPR International Workshop on Document Analysis Systems, DAS 2008, September 16-19, 2008, Nara, Japan
Source: DBLP


There is an approach of annotation extraction from printed documents in which annotations are extracted by comparing the image of an annotated document and its original document image. In one of the previous methods, the image of an original document is actually printed and scanned in order to reproduce image degradations of the image of the annotated document. However such a method lacks convenience since users have to use the same printer and scanner to obtain images of an annotated document and its original document. In this paper, we propose an improved annotation extraction method in which the image degradations are compensated by image processing. In the proposed method, the difference between original and annotated document images due to image degradations is reduced by not only removal of the degradations in the annotated document images but also reproduction of the degradations in the original document images. The proposed method consists of three steps of processing which are for dithering, for color change, and for local displacement. We also propose an objective evaluation of extracted annotations to compare the experimental results accurately. Experimental results of the proposed method have shown that the recall of extracted annotations was 80.94% and the precision was 85.59%.

Full-text preview

Available from:
  • Source
    • "Nakai et al. have proposed another approach called locally likely arrangement hashing (LLAH) [11], which descriptor is composed by the combination of keypoint arrangement. LLAH is applied to several applications such as annotation extraction [6] and document mosaicing [12] because it works in real-time. We have modified LLAH in order to handle a large range of viewpoint by on-line learning of the descriptors [7]. "
    [Show abstract] [Hide abstract]
    ABSTRACT: This paper presents an Augmented Reality (AR) system for physical text documents that enable users to click a document. In the system, we track the relative pose between a camera and a document to overlay some virtual contents on the document continuously. In addition, we compute the trajectory of a fingertip based on skin color detection for clicking interaction. By merging a document tracking and an interaction technique, we have developed a novel tangible document system. As an application, we develop an AR dictionary system that overlays the meaning and explanation of words by clicking on a document. In the experiment part, we present the accuracy of the clicking interaction and the robustness of our document tracking method against the occlusion.
    Full-text · Conference Paper · Nov 2010
  • Source
    • "Uchiyama and Saito extended the framework of LLAH for more wide range tracking [15]. LLAH is applied to annotation extraction written in a physical documents [9] and extended to a framework for augmented reality [15]. Since LLAH can achieve real-time processing thanks to hashing scheme [11] [15], we develop the system based on LLAH as described in Section 4. "
    [Show abstract] [Hide abstract]
    ABSTRACT: We propose a system that registers and retrieves text documents to annotate them on-line. The user registers a text document captured from a nearly top view and adds virtual annotations. When the user thereafter captures the document again, the system retrieves and displays the appropriate annotations, in real-time and at the correct location. Registering and deleting documents is done by user interaction. Our approach relies on LLAH, a hashing based method for document image retrieval. At the on-line registering stage, our system extracts keypoints from the input image and stores their descriptors computed from their neighbors. After registration, our system can quickly find the stored document corresponding to an input view by matching keypoints. From the matches, our system estimates the geometrical relationship between the camera and the document for accurately overlaying the annotations. In the experimental results, we show that our system can achieve on-line and real-time performances.
    Full-text · Conference Paper · Jan 2010
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: This article describes a method for extracting and classifying handwritten annotations on printed documents using a simple camera integrated in a lamp or a mobile phone. The ambition of such a research is to offer a seamless integration of notes taken on printed paper in our daily interactions with digital documents. Existing studies propose a classification of annotations based on their form and function. We demonstrate a method for automating such a classification and report experimental results showing the classification accuracy.
    Full-text · Article ·