Conference Paper

Using Self-Organizing Maps to Support Video Navigation.

DOI: 10.1007/11840817_42 Conference: ICANN (1)


Content-based video navigation is an efficient method for browsing video information. A common approach is to cluster shots
into groups and visualize them afterwards. In this paper, we present a prototype that follows in general this approach. Unlike
existing systems, the clustering is based on a growing self-organizing map algorithm. We focus on studying the applicability
of SOMs for video navigation support. We ignore the temporal aspect completely during the clustering, but we project the grouped
data on an original time bar control afterwards. This complements our interface by providing – at the same time – an integrated
view of time and content based information. The aim is to supply the user with as much information as possible on one single
screen, without overwhelming him. Special attention is also given to the interaction possibilities which are hierarchically

3 Reads
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: In this paper we present a variety of browsing interfaces for digital video information. The six interfaces are implemented on top of Físchlár, an operational recording, indexing, browsing and playback system for broadcast TV programmes. In developing the six browsing interfaces, we have been informed by the various dimensions which can be used to distinguish one interface from another. For this we include layeredness (the number of “layers” of abstraction which can be used in browsing a programme), the provision or omission of temporal information (varying from full timestamp information to nothing at all on time) and visualisation of spatial vs. temporal aspects of the video. After introducing and defining these dimensions we then locate some common browsing interfaces from the literature in this 3-dimensional “space” and then we locate our own six interfaces in this same space. We then present an outline of the interfaces and include some user feedback.
    12/1999: pages 206-218;
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: eo from a single camera running from camera on to camera off. Using one keyframe per shot means that representing a one-hour video usually requires hundreds of keyframes. In contrast, our approach for video indexing and summarization selects fewer keyframes that represent the entire video and index the interesting parts. The user can select the number of keyframes or the application can select the optimal number of keyframes based on display size, but a one-hour video typically will have between 10 and 40 keyframes. We use several techniques to present the automatically selected keyframes. A video directory listing shows one keyframe for each video and provides a slider that lets the user change the keyframes dynamically. The visual summary of a single video presents images in a compact, visually pleasing display. To deal with the large number of keyframes that represent clips in a video editing system, we group keyframes into piles based on their visual similarity. In all three inter
    Computer 10/2001; 34(9):61-67. DOI:10.1109/2.947093 · 1.44 Impact Factor
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: This article describes the implementation of a system that is able to organize vast document collections according to textual similarities. It is based on the self-organizing map (SOM) algorithm. As the feature vectors for the documents statistical representations of their vocabularies are used. The main goal in our work has been to scale up the SOM algorithm to be able to deal with large amounts of high-dimensional data. In a practical experiment we mapped 6,840,568 patent abstracts onto a 1,002,240-node SOM. As the feature vectors we used 500-dimensional vectors of stochastic figures obtained as random projections of weighted word histograms.
    IEEE Transactions on Neural Networks 02/2000; 11(3):574-85. DOI:10.1109/72.846729 · 2.95 Impact Factor
Show more


3 Reads