Article

DocMIR: An automatic document-based indexing system for meeting retrieval

University of Leeds School of Computing Leeds LS2 9JT United Kingdom
Multimedia Tools and Applications (Impact Factor: 1.06). 04/2008; 37(2):135-167. DOI: 10.1007/s11042-007-0137-4
Source: DBLP

ABSTRACT This paper describes the DocMIR system which captures, analyzes and indexes automatically meetings, conferences, lectures, etc. by taking advantage of the
documents projected (e.g. slideshows, budget tables, figures, etc.) during the events. For instance, the system can automatically
apply the above-mentioned procedures to a lecture and automatically index the event according to the presented slides and
their contents. For indexing, the system requires neither specific software installed on the presenter’s computer nor any
conscious intervention of the speaker throughout the presentation. The only material required by the system is the electronic
presentation file of the speaker. Even if not provided, the system would temporally segment the presentation and offer a simple
storyboard-like browsing interface. The system runs on several capture boxes connected to cameras and microphones that records
events, synchronously. Once the recording is over, indexing is automatically performed by analyzing the content of the captured
video containing projected documents and detects the scene changes, identifies the documents, computes their duration and
extracts their textual content. Each of the captured images is identified from a repository containing all original electronic
documents, captured audio–visual data and metadata created during post-production. The identification is based on documents’
signatures, which hierarchically structure features from both layout structure and color distributions of the document images.
Video segments are finally enriched with textual content of the identified original documents, which further facilitate the
query and retrieval without using OCR. The signature-based indexing method proposed in this article is robust and works with
low-resolution images and can be applied to several other applications including real-time document recognition, multimedia
IR and augmented reality systems.

Download full-text

Full-text

Available from: Denis Lalanne, Apr 16, 2014
0 Followers
 · 
99 Views
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: In some scenarios, it is paramount that collaborative syn-chronous sessions be recorded for later review. Particularly in the case of webconferencing tools, the approach usually adopted for recording a meeting is to generate a linear video with the content of the exchanged media. Such approach limits the review of the meeting to the user watching a video using the traditional timeline-based video controls. In this work we advocate that synchronous communication tools generate interactive multimedia documents as a result of a session. For this purpose we detail the generation of an in-teractive multimedia document by means of user-media op-erators (e.g. slide changes, chat messages, ink-based interac-tions, audio events) that we call Interactors, enabling users to browse the generated document by navigating points of interest in the captured media. We define the Interactors approach and demonstrate it in the context of a tool in use.
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: In the context of animated movie characterization, we present an information fusion approach mixing very different types of data related to the activity within a movie. These data are the features extracted from images, words extracted from the synopses and expert knowledge. The difficulty of this fusion is due to the very different semantic level of these data. The aim of this work is to get a movie activity characterization in order to help the constitution of automatic summary, content based video retrieval system, etc. Two strategies are proposed : a first one aiming at giving a global description of the activity within the movie, and a second one providing a local description of activity. Tests and results are proposed on animated movies from the Annecy International Animation Film Festival.
    Information Fusion, 2009. FUSION '09. 12th International Conference on; 08/2009
  • [Show abstract] [Hide abstract]
    ABSTRACT: Synchronous communication tools allow remote users to collaborate by exchanging text, audio, images or video messages in synchronous sessions. In some scenarios, it is paramount that collaborative synchronous sessions be recorded for later review. In particular in the case of web conferencing tools, the approach usually adopted for recording a meeting is to generate a linear video with the content of the exchanged media. Such approach limits the review of a meeting to users watching a video using traditional timeline-based video controls. In this work we advocate that interactive multimedia documents can be generated automatically as a result of capturing a synchronous session. We outline our approach presenting a case of study involving remote communication, and detail the generation of a multimedia document by means of operators focusing on the interaction among the collaborating users.
    Proceedings of the 2010 ACM Symposium on Document Engineering, Manchester, United Kingdom, September 21-24, 2010; 01/2010