To read the full-text of this research, you can request a copy directly from the authors.
... We provide example images to guide users through their search of desired targets. The concept of query-by-example is based on the observation that many users have only vaguely defined information needs so that they may be able to recognize what they are looking for rather than describing or sketching it [15]. For such an application, pictorial examples and an interactive and cooperative human-machine interface can be of great help. ...
... To effectively access information in an image database, there are several design criteria for user interface [15]: 1) integration of various query mechanisms, 2) a visual or graphical user interface (GUI), 3) incorporation of user's relevant feedback, 4) support of user-guided navigation. To meet parts of these criteria, we implemented the Wave-Guide interface on the World Wide Web (WWW) by using C, HTML, UNIX, and Common-Gateway-Interface (CGI). ...
... Just as QBIC, we do not attempt to build up complicated data model with WaveGuide. The system concentrates on the signal processing techniques such as indexing, low-level feature extraction, similarity matching, and compression rather than techniques like data models, semantic world representation or annotation, which are usually applied in the systems developed by database researchers [15]. A complete data model for a natural image database can be a pretty difficult task. ...
Data representation and content description are two basic
components required by the management of any image database. A wavelet
based system, called the WaveGuide, which integrates these two
components in a unified framework, is proposed in this work. In the
WaveGuide system, images are compressed with the state-of-the-art
wavelet coding technique and indexed with color, texture, and object
shape descriptors generated in the wavelet domain during the encoding
process. All the content descriptors are extracted by machines
automatically with a low computational complexity and stored with a low
memory space. Extensive experiments are performed to demonstrate the
performance of the new approach
... The elusive goal that we should strive for would be a generalized notion of guaranteed service quality. The notion of service quality (also known as quality of service or QoS for short) has come up in the context of multimedia communication (see, e.g., [22]), but has so far been limited to low-level issues like guaranteed packet delivery rates. What we need is a well-founded generalization of this notion to the application level, covering building blocks such as database access or middleware components like web application and workflow servers, Web or Intranet search engines, and also comprehensive, value-added information services on the Internet. ...
... For given server resources (e.g., number of disks, amount of memory) as well as data and workload profiles, the server needs to know the maximum number, N, of admissable concurrent streams. This limit can be derived in a conservative manner by assuming a maximum fragment size and worst-case disk access delays for all streams in all scheduling rounds (see, e.g., [22]). However, this crude assumption does not take into account the variability of the fragment size and the disks' performance behavior, and therefore tends to end up with a substantially underutilized server. ...
... A nice property of this classical approach is that it can be generalized to searching on multimedia objects such as images, videos, or music [14,22]. Once an appropriate feature space has been defined, for example, based on color distribution or contours in images, the principles of similarity ranking apply more or less directly. ...
The impressive advances in global networking and information technology provide great opportunities for all kinds of Web-based information services, ranging from digital libraries and information discovery to virtual-enterprise workflows and electronic commerce. However, many of these services still exhibit rather poor quality in terms of unacceptable performance during load peaks, frequent and long outages, and unsatisfactory search results. For the next decade, the overriding goal of database research should be to provide means for building zero-administration, self-tuning information services with predictable response time, virtually continuous availability, and, ultimately, "money-back" servicequality guarantees. A particularly challenging aspect of this theme is the quality of search results in digital libraries, scientific data repositories, and on the Web. To aim for more intelligent search that can truly find needles in haystacks, classical information retrieval methods should ...
... One of the reasons is the lack of reliable methods for content analysis of the different media types, thus basic mechanisms and technologies for the management of multimedia data are not available yet. Many existing relational and objectoriented databases handle multimedia objects as BLOBs (Binary Large Objects) [1,6] and describe their content by a manually compiled and limited set of keywords. The retrieval is then realised by a full text search in the assigned set of keywords. ...
... The state-of-the-art approach [1] for the creation and retrieval of image databases is based on the extraction and comparison of a priori defined features. These are directly derivable from the raw data and represent properties related to the dominant colours in the image and their distribution, important shapes, textures and consider the global layout. ...
This paper presents an overview over parallel architectures for the efficient realisation of digital libraries by considering image databases as an example. The state of the art approach for image retrieval uses a priori extracted features and limits the applicability of the retrieval techniques, as a detail search for objects and for other important elements can't be performed. Well-suited algorithms for dynamic feature extraction and comparison are not often applied, as they require huge computational and memory resources. Integration of parallel methods and architectures enables the use of these alternative approaches for improved classification and retrieval of documents in digital libraries. Therefore implemented prototypes on a symmetric multiprocessor (SMP) and on cluster architecture are introduced in the paper. Performance measurements with a wavelet-based template matching method resulted into a reasonable speedup.
... Most of the applications and services of the Internet and enterprise networks use stored multimedia information. The trend shows the emergence of a multimedia Web-based technology that will allow users to interactively author, store, and share multimedia documents (MMD's) [1]. User interaction in a multimedia Web environment can be characterized by a two-level browsing graph, with vertices pointing toward MMD's and edges representing the logical links among documents. ...
... An example of temporal composition of an MMD is depicted in Fig. 1(b), where a document-level description of a particular node is expanded on the timeline. A meaningful presentation of multimedia information requires synchronous playout of the component objects according to the specified temporal relations [1]- [3]. ...
We consider the problem of multimedia synchronization in a Web
environment. The workload generated by the multimedia server during a
Web session exhibits variations that are quite different from the
traffic fluctuation offered by a single media stream, e.g., a variable
bit rate (VBR) video. We propose a set of parameters that can be used to
characterize the workload generated by the multimedia server in a
Web-type browsing environment. The workload characterization scheme is
subsequently used in designing a server-based synchronization scheme.
The problem of scheduling multimedia information to ensure media
synchronization in a Web environment is identified as a multicriteria
scheduling problem, which is NP-hard. The ability of fuzzy control to
deal with multivariables makes it a good alternative for the
multicriteria scheduling problem considered. Consequently, we propose a
neuro-fuzzy scheduler (NFS) that makes an intelligent compromise among
multicriteria by properly combining some scheduling heuristics.
Performance of the NFS is compared with several known heuristics and a
branch and bound algorithm. The results show that the proposed NFS ran
dynamically adjust to the varying workload quite well
... One of the reasons is the lack of reliable methods for content analysis of the different media types, thus basic mechanisms and technologies for the management of multimedia data are not available yet. Many existing relational and objectoriented databases handle multimedia objects as BLOBs (Binary Large Objects) [1,6] and describe their content by a manually compiled and limited set of keywords. The retrieval is then realised by a full text search in the assigned set of keywords. ...
... The state-of-the-art approach [1] for the creation and retrieval of image databases is based on the extraction and comparison of a priori defined features. These are directly derivable from the raw data and represent properties related to the dominant colours in the image and their distribution, important shapes, textures and consider the global layout. ...
: This paper presents an overview over parallel architectures for the efficient realisation of digital libraries by considering image databases as an example. The state of the art approach for image retrieval uses a priori extracted features and limits the applicability of the retrieval techniques, as a detail search for objects and for other important elements can't be performed. Well-suited algorithms for dynamic feature extraction and comparison are not often applied, as they require huge computational and memory resources. Integration of parallel methods and architectures enables the use of these alternative approaches for improved classification and retrieval of documents in digital libraries. Therefore implemented prototypes on a symmetric multiprocessor (SMP) and on cluster architecture are introduced in the paper. Performance measurements with a wavelet-based template matching method resulted into a reasonable speedup. 1.
... Digital images databases, however, open the way to content-based searching. There are various technical aspects of current content-based image retrieval systems and a number of other overviews on image database systems, image retrieval, or multimedia information systems have been published [2]. ...
The Fast Development of the image capturing in digital form leads to the availability of large databases of images. The manipulation and management of images within these databases depend mainly on the user interface and the search algorithm used to search these huge databases for images, there are two search methods for searching within image databases: Text-Based and Content-Based. In this paper, we present a method for content-based image retrieval based on most used colors to extract image features. A preprocessing is applied to enhance the extracted features, which are smoothing, quantization and edge detection. Color quantization is applied using RGB (Red, Green, and Blue) Color Space to reduce the range of colors in the image and then extract the most used color from the image. In this approach, Color distance is applied using HSV (Hue, Saturation, Value) color space for comparing a query image with database images because it is the closest color space to the human perspective of colors. This approach provides accurate, efficient, less complex retrieval system.
... Digital images databases, however, open the way to content-based searching. There are various technical aspects of current content-based image retrieval systems and a number of other overviews on image database systems, image retrieval, or multimedia information systems have been published [2]. ...
... Another approach for image representation is the perceptual one [9], based on anisotropic filtration controlled by the Human Visual System (HVS) visual attention model. The knowledge-based models for image representation are used mostly in the systems for Visual Information Retrieval [10][11][12][13]. In these cases the main approach for image representation used is the pyramid model of 4 layers, which contain correspondingly: the primary matrix, the features vectors, the description of the relations between the features and the semantic image structure. ...
The contemporary image representation is based on various techniques, using matrices, vectors, multi-resolution pyramids, R-tree, orthogonal transforms, anisotropic perceptual representations, etc. In this paper is offered one new approach for cognitive image representation based on adaptive spectrum pyramid decomposition controlled by neural networks. This approach corresponds to the hypothesis of the human way for image recognition using consecutive approximations with increasing resolution for the selected regions of interest. Such image representation is suitable for the creation of the objects' learning models, which should be extracted from image databases in accordance with predefined decision rules. Significant element of the new representation is the use of a feedback, which to provide iterative change of the cognitive models' parameters in accordance with the data mining results obtained.
... The Research Server Complex Manager (RSCM) provides a uniform external interface to applications hiding the heterogeneity and making the server complex appear as individual requests. Dan et al. justify the particular functions of the RSCM and explains the design decisions and tradeoffs.Buddhikot et al.[Gros97] suggest the Massively-parallel And Real-time Storage (MARS) architecture for the design and prototype implementation of a largescale video server. MARS exploits some of the well-known techniques in parallel I/O, such as data striping and an innovative ATM based interconnect inside the server to achieve a scalable architecture that transparently connects storage devices to an ATM-based broadband network. ...
... However, spatial segmentation is not common in video retrieval because of its computational complexity [Grosky et al., 1997;Petković and Jonker, 2004] and we do not discuss it here. ...
... The second major idea behind Implict Patching is the use of Composite design pattern [23] as a mechanism for constructing dataflow networks. The extended semantics of Slices require careful manipulation of buffer sizes especially if run-time changes are desired. ...
Programming software for audio analysis and synthesis is challenging. Dataflow-based approaches provide a declar- ative specification of computation and result in efficient code. Most practitioners of computer music are familiar with some form of dataflow programming where audio applications are constructed by connecting components with "wires" that carry data. Examples include networks of unit generators in Music-V style languages and visual patches in Max/Msp or PD. Even though existing dataflow- based audio systems offer a concise conceptual model of signal computation, this model does have limitations. In many cases, these limitations are a consequence of the programmer having to explicitly specify connections be- tween components. Two such limitations are the diffi- culty of handling spectral data and the need for fixed-size buffers between components. In this paper we introduce Implicit Patching(IP), a dataflow-based approach to audio analysis and synthesis that attempts to address these limi- tations. By extending dataflow semantics a large number of connections are automatically created and buffer sizes can be changed dynamically. We describe Marsyas-0.2, a software framework based on IP, and comment on the strengths and limitations of the proposed approach.
... The retrieval effectiveness of our system was measured in terms of precision and recall [15], since they have been widely used to evaluate retrieval effectiveness. Precision is defined as the fraction of the retrieved images that are relevant to the given query, while the recall represents the proportion of relevant images among the retrieved ones. ...
This paper presents a novel approach to fingerprint retrieval for personal identi- fication by joining three image retrieval tasks, namely, feature extraction, similarity measurement, and feature indexing, into a wavelet-based fingerprint retrieval system. We propose the use of different types of Wavelets for representing and describing the textural information present in fingerprint images. For that purposes, the feature vectors used to characterize the fingerprints are obtained by computing the mean and the standard deviation of the decomposed images in the Wavelet domain. These feature vectors are used to retrieve the most similar fingerprints given a query image, while their indexation is used to reduce the search spaces of image candidates. The different types of Wavelets used in our study include: Gabor Wavelets (GWs), Tree-Structured Wavelet Decomposition using both Orthogonal Filter Banks (TOWT) and Bi-orthogonal Filter Banks (TBOWT), as well as the Steerable Wavelets. To evaluate the retrieval accuracy of the proposed approach, a total number of eight different data sets were used. Experiments also evaluated different combinations of Wavelets with six similarity measures. The results show that the Gabor Wavelets com- bined with the Square Chord similarity measure achieves the best retrieval effectiveness.
... Another approach for image representation is the perceptual one [9], based on anisotropic filtration controlled by the Human Visual System (HVS) visual attention model. The knowledge-based models for image representation are used mostly in the systems for Visual Information Retrieval [10][11][12][13]. In these cases the main approach for image representation used is the pyramid model of 4 layers, which contain correspondingly: the primary matrix, the features vectors, the description of the relations between the features and the semantic image structure. ...
The contemporary image representation is based on various techniques, using matrices, vectors, multi-resolution pyramids, R-tree, orthogonal transforms, anisotropic perceptual representations, etc. In this paper is offered one new approach for cognitive image representation based on adaptive spectrum pyramid decomposition controlled by neural networks. This approach corresponds to the hypothesis of the human way for image recognition using consecutive approximations with increasing resolution for the selected regions of interest. Such image representation is suitable for the creation of the objects' learning models, which should be extracted from image databases in accordance with predefined decision rules. Significant element of the new representation is the use of a feedback, which to provide iterative change of the cognitive models' parameters in accordance with the data mining results obtained.
... Multimedia would enable strong first impression, mitigating ambiguity and influencing better decision making [4, 5]. With intelligent organization of multimedia content it is possible to add syntax and semantics which can be represented as metadata of the content, which in turn aids content management and location [6, 7]. ...
In this paper, we propose a novel architecture to enable engineering content exchange in mobile computing environment by using a language which is extended from eXtensible Markup Language (XML). To solve the challenges that we are facing in mobile computing environment where network resource and computational resource are limited, we use a hierarchy content representation and on-demand progressive transmission mechanism for an engineering field service automation application, which features with rich graphical engineering content including drawing, schematics, plans, circuits, maps, etc. Since content is delivered and represented in a progressive fashion based on user request, both problems of limited network resource and computing resource are being minimized.
... A more effective and automatic approach is the so-called content-based image retrieval – CBIR, which consists of using low-level image features to represent, compare and retrieve images. Most CBIR systems [15] follow the two-step approach to search image databases [8]. Firstly (indexing), a feature vector representing certain essential properties of the image is extracted and stored as metadata for each database image. ...
This paper presents the use of relevance feedback to the problem of content-based sub-image retrieval (CBsIR). Relevance feedback is used to improve the accuracy of successive retrievals via a tile re-weighting scheme that assigns penalties to each tile of database images and updates the tile penalties for all relevant images retrieved at each iteration using both the relevant (positive) and irrelevant (negative) images identified by the user. Performance evaluation on a dataset of over 10,000 images shows the effectiveness and efficiency of the proposed framework. Using 64 quantized colors in the RGB color space, the system can achieve a stable average recall value of 70% within the top 20 retrieved (and presented) images after only 5 iterations, with each such iteration taking about 2 seconds.
... Two major factors affect the QoS requirements: synchronization and reliability. Temporal synchronization at the client [2] is one of the major issues in the timely transport of multimedia information over a network. Jitter delay experienced over a data network is a major challenge to synchronizing the multimedia stream. ...
We propose a joint optimization network management framework for quality-of-service (QoS) routing with resource allocation. Our joint optimization framework provides a convenient way of maximizing the reliability or minimizing the jitter delay of paths. Data traffic is sensitive to droppage at buffers, while it can tolerate jitter delay. On the other hand, multimedia traffic can tolerate loss but it is very sensitive to jitter delay. Depending on the type of data, our scheme provides a convenient way of selecting the parameters which result in either reliability maximization or jitter minimization. We solve the optimization problem for a GPS network and provide the optimal solutions. We find the values of control parameters which control the type of optimization performed. We use our analytical results in a multi-objective QoS routing algorithm. Finally, we provide insights into our optimization framework using simulations.
... In this paper, we apply software engineering methodology to the production of multimedia applications introducing a principled approach to specify, verify, validate and prototype such applications. Our approach to multimedia application development is based on a collection of tools which support the creation of Teleaction Objects (TAOs) [5,14]. A TAO is a multimedia object with associated hypergraph structure and knowledge structure. ...
Multimedia systems incorporating hyperlinks and user interaction can be prototyped using TAOML, an extension of HTML. TAOML is used to define a Teleaction Object (TAO) which is a multimedia object with associated hypergraph structure and knowledge structure. The user can create and modify the private knowledge of a TAO so that the TAO will react automatically to certain events. The hypergraph structure supports the effective presentation and efficient communication of multimedia information. TAOs are valuable since they greatly improve the selective access and presentation of relevant multimedia information. In this paper, a formal specification methodology for TAOs using SR grammars is described. An attributed SR grammar is then introduced in order to associate knowledge with the TAO. The limitations necessary to achieve an efficient parser are given. The grammatical formalism allows for validation and verification of the system specification. This methodology thus provides a principled approach to specify, verify, validate and prototype multimedia applications.
... In these systems the user can search, browse and retrieve images based on similarity and various automatic feature extraction methods. This work falls under the general area of Multimedia Information Management Systems [51] with specific applications to audio signals. ...
Thesis (Ph.D.)--Princeton University, 2002. Includes bibliographical references (leaves 174-184). Microfiche. s
... Several studies have focused on data placement and retrieval scheduling of continuous media objects, for collections see [9,13,6,10,4,5]. Among these, only a few studied the random placement of data blocks on continuous media servers [3,12,11,14]. ...
Scalable storage architectures allow for the addition of disks to increase storage capacity and/or bandwidth. In its general form, disk scaling also refers to disk removals when either capacity needs to be conserved or old disk drives are retired. Assuming random placement of blocks on multiple nodes of a continuous media server, our optimization objective is to redistribute a minimum number of media blocks after disk scaling. This objective should be met under two restrictions. First, uniform distribution and hence a balanced load should be ensured after redistribution. Second, the redistributed blocks should be retrieved at the normal mode of operation in one disk access and through low complexity computation. We propose a technique that meets the objective, while we prove that it also satisfies both restrictions. The SCADDAR approach is based on using a series of REMAP functions which can derive the location of a new block using only its original location as a basis
... Other issues in image retrieval, such as high dimensional indexing, visualization and browsing, human interaction and knowledge engineering, are out of the scope of this chapter. For detailed information of those topics please refer to (Baeza-Yates & Ribeiro-Neto, 1999;Del Bimbo, 1999;Grosky, Jain, & Mehrotra, 1997;Rui, Huang, & Chang, 1998;Smeulders, et al., 2000). ...
the gap between them is still a huge barrier in front of researchers. Intuitive and heuristic approaches do not provide us with satisfactory performance. Therefore, there is an urgent need of finding the latent correlation between low-level features and high-level concepts and merging them from a different perspective. How to find this new perspective and bridge the gap between visual features and semantic features has been a major challenge in this research field. Our paper addresses these issues. Bridging the Semantic Gap in Image Retrieval 15 INTRODUCTION The emergence of multimedia technology and the rapidly expanding image and video collections on the Internet have attracted significant research efforts in providing tools for effective retrieval and management of visual data. Image retrieval is based on the availability of a representation scheme of image content. Image content descriptors may be visual features such as color, texture, shape, and spa
... Most notably, for smooth playback of video and audio on client machines the server has to guarantee continuous data streams with just-in-time delivery of the underlying data fragments. This problem is referred to in the multimedia information and networking community as the need for guaranteed quality of service, in short QoS, where the details of such guarantees depend on the data and playback speci cations (e.g., video formats) 1,4,8,9]. ...
Servers for multimedia digital libraries have to manage huge amounts of data and pose challenging performance requirements. Most notably, for smooth playback of video and audio on client machines the server has to guarantee continuous data streams with just-in-time delivery of the underlying data fragments. This problem is referred to in the multimedia information and networking community as the need for guaranteed quality of service, in short QoS, where the details of such guarantees depend on the data and playback specifications (e.g., video formats) [1, 4, 8, 9]. Servers aim to maximize their throughput in terms of the concurrently sustained number of continuous data streams, but at the same time have to ensure that each active stream meets the QoS demands of the clients. Therefore servers (and potentially also network components) need to employ an admission control that limits the number of simultaneously active streams. For given server resources such as number of disks, memory s...
... A number of other overviews on image database systems, image retrieval, or multimedia information systems have been published, see e.g. [82], [28], [30], [36], [29], [21], [68], [10]. The purpose of this survey however, is to provide an overview of the functionality of temporary image retrieval systems in terms of technical aspects: querying, relevance feedback, features, matching measures, indexing data structures, and result presentation. ...
this paper we survey some technical aspects of current content-based image retrieval systems.
... For a complete image database management system, there are other important issues need to be addressed[GJM97,Gong98], such as indexing data structures for search (R-tree variants), querying schemes, and query languages. These issues are studied by the database research community are not considered in our work. ...
learning the query, learning the metric --- and combine these learning methods with color correlograms. Our experiments show that these learning methods are quite effective with even a little effort from users. We also adapt the correlogram to handle the problems of image subregion querying, object localization and tracking. We propose the correlogram intersection for object detection and correlogram correction for object localization. These simple methods perform better than methods based on color histograms. Finally, we propose a method for hierarchical classification of images via supervised learning. This scheme uses correlogram as the low-level feature and performs feature-space reconfiguration using singular value decomposition to reduce noise and dimensionality. We use the training data to obtain a hierarchical classification tree that can be used to categorize new images. Our experimental results suggest that this
This chapter presents an introduction to the area of parallel and distributed multimedia database systems. The first part describes the characteristics of multimedia data and depicts the storage and annotation of such data in conventional and in multimedia databases. The main aim is to explain the process of multimedia retrieval by using images as an example. The related computational, storage, and network requirements create an urgent need for the integration of parallel and distributed computer architectures in modern multimedia information systems. Different hardware and software aspects have to be examined, for example the partitioning of multimedia data and the distribution over multiple nodes have a decisive impact on the performance, efficiency, and the usability of such multimedia databases. Other distributed aspects such as streaming techniques, proxy and client issues, security, etc. are only briefly mentioned and are not in the focus of this chapter. The last section gives an overview over an existing cluster-based prototype for image retrieval named Cairo.
A central step in content based video retrieval is the temporal segmentation of video. An application in-dependent approach to video segmentation is to de-tect temporally contiguous segments without signii-cant content change between successive frames. Each such segment is termed as a shot . A high performance shot boundary detection based video segmentation al-gorithm is proposed. The technique uses unsupervised clustering on a multiple feature input space, followed by a heuristic elimination process to detect with almost perfect accuracy shot boundaries in the video. With an extremely high accuracy coupled with a very small number of false positives, this algorithm outperforms most of the existing techniques.
This paper proposes a consistency algorithm that is able to serve streaming data efficiently in VOD system. The media data is stripping into several pieces of data by the Round Robin method in order to media data service. The barrier mechanism is changed into the minimum data factor(SH. GOP) in this paper. The shared memory is allocated at one host with one fragment size. Data is combined with RTP packet transmission data format using barrier mechanism. I experiment and program the suggested algorithm on the VOD system.
The status in quo of remote sensing (RS) image inquiry is discussed in this paper firstly. Then, on the basis of analysis of content-based image inquiry and image data characteristics, an image content-based retrieval method supported by GIS is proposed. This method makes the fusion of image data and vector data's attribute, so it can assign the semantic information to image data indirectly and realize image inquiry. Finally, supported by "digital Mt. Tai construction" project, the semantic retrieval method of Mt. Tai scenic spots and historical relics based on RS image data content is studied and a well-planned scheme and framework to construct the retrieval system is proposed. It is ensured that the management level of Mt. Tai scenic spots and historical relics and service level of Mt. Tai tourism industry is promoted.
A moving object has a various features that its spatial location, shape, and size are changed as time goes. In addition, the moving object has both temporal feature and spatial feature. It is one of the highly interested feature information in video data. In this paper, we propose an efficient content-based multimedia information retrieval system, so tailed ECoMOT which enables user to retrieve video data by using a trajectory information of moving objects in video data. The ECoMOT includes several novel techniques to achieve content-based retrieval using moving objects' trajectories : (1) Muitiple trajectory modeling technique to model the multiple trajectories composed of several moving objects; (2) Multiple similar trajectory retrieval technique to retrieve more similar trajectories by measuring similarity between a given two trajectories composed of several moving objects; (3) Superimposed signature-based trajectory indexing technique to effectively search corresponding trajectories from a large trajectory databases; (4) convenient trajectory extraction, query generation, and retrieval interface based on graphic user interface
Although there are many applications where an object-oriented data model is a good way of representing and querying data, current object database systems are unable to handle objects whose attributes are uncertain. In this article, we extend previous work by Kornatzky and Shimony to develop an algebra to handle object bases with uncertainty. We propose concepts of consistency for such object bases, together with an NP-completeness result, and classes of probabilistic object bases for which consistency is polynomially checkable. In addition, as certain operations involve conjunctions and disjunctions of events, and as the probability of conjunctive and disjunctive events depends both on the probabilities of the primitive events involved as well as on what is known (if anything) about the relationship between the events, we show how all our algebraic operations may be performed under arbitrary probabilistic conjunction and disjunction strategies. We also develop a host of equivalence results in our algebra, which may be used as rewrite rules for query optimization. Last but not least, we have developed a prototype probabilistic object base server on top of ObjectStore. We describe experiments to assess the efficiency of different possible rewrite rules.
We may say the knowledge is the new global asset – so people have to be continually learning, crafting innovative solutions to changing circumstances, staying informed and responsive. Computer Based Learning (CBL) provides perhaps the best opportunity for person self-guided learning. If we take into consideration the role played by the Internet in everybody's life nowadays and if we make a further step forward to a CBL process using the Internet, we reach a more powerful variant of CBL: Computer On-line Learning (COL). Keeping in mind these two important needs we can easily see that general and efficient organization of multimedia (MM) information is one of the main requirements for developing courseware, which can be accesed by anyone's computer quickly and smoothly. MM DataBase Management Systems (MM DBMSs) offer a good solution for this problem. They allow storing and manipulating such data in an efficient manner with a high level of generalizability. Vast digital libraries of information will soon be available on the Internet as a result of emerging technologies for manipulating MM. These libraries will profoundly impact the educational activities. A new virtual teaching and learning environment is evolving.
In the context of constant development of methods for distance learning, or e-learning, the paper presents a project in which
the software tools specific to this type of education are being applied, namely, the ERASMUS project called „ Filière Euro
Qualité Logistique des Organisations” , acronym EURO-QLIO. The e-learning process offers flexibility and affordability "anywhere,
anytime", so that it can be used in both hybrid and joint learning. The project EURO-QLIO ensures the preparation and conduct
of the partnership for training in the field of Metrology, Quality Control and Organization Logistics. The partnership is
ensured by the involvement of three renowned universities from three different countries: „ University Henri Poincaré” from
Nancy, France (which is the project coordinator) "University Politehnica of Bucharest”, Romania, and "University Angel Kanchev”
from Rousse, Bulgaria. The Romanian partner has co-opted from the industrial area the firm “SC Marco & Alex Instalaţi Frig
SRL”. This partner represents the interface between the Euro-QLIO project and the industrial beneficiaries. The educational
platform is based on two very important components, namely: establishment of a virtual campus and issuance of a diploma with
triple recognition.
The US government and its agencies have a mandate to disseminate information on their operations for public use. Two such agencies are the National Marine Fisheries Service, and the Environmental Protection Agency. These agencies have an e-government directive, requiring them to provide historically collected electronic information to the general public and other agencies, organizations or businesses having legitimate uses for their databases. The three most important concepts for e-government data distribution are data preparation, presentation and data management. Unfortunately many of today's Web based information systems put presentation before content, and content before design. All three aspects are of equal importance and should receive equal attention. This paper discusses four Web based data distribution models, protocols, services and policies, with varying levels of complexity, being developed to allow the general information on water quality assessment, watershed health and fisheries assessment in the New England region.
A central step in content-based video retrieval is the temporal
segmentation of video. An application independent approach to video
segmentation is to detect temporally contiguous segments without
significant content change between successive frames. Each such segment
is termed a shot. A high-performance shot boundary detection-based video
segmentation algorithm is proposed. The technique uses unsupervised
clustering on a multiple feature input space, followed by a heuristic
elimination process to detect, with almost perfect accuracy, shot
boundaries in the video. With an extremely high accuracy coupled with a
very small number of false positives, this algorithm outperforms most of
the existing techniques
In this paper, we present a sender-driven adaptive optimal multimedia network transmission control scheme, which takes into
account the buffer occupancy and network delay to maximize the utilization of network resources. For this purpose, adaptive
network resource optimization with quadratic costs is used in the proposed scheme to provide the minimal allocation of the
bandwidth and achieve the maximal utilization of the client buffer. Simulation results show that the tranmission rate can
be dynamically adjusted at the server according to the changing network delays and buffer packet sizes at the client to avoid
the loss of packets, and at the same time to achieve the minimal bandwidth allocation at maximal utilization of the client
buffer.
In this paper, we propose an approach that discovers automatically visual relations in order to make more powerful the image
access. The visual relationships are discovered automatically from images. They are statistical rules in the form of a a-
b which means: if the visual feature “a” is true in an image then the visual feature “b” is true in the same image with a
precision value. The rules concern symbols that are extracted from image numerical features. The transformation of image numerical
features into image symbolic features needs a visual feature book in which each book feature is the gravity center of similar
features. The approach presents the clustering algorithm that creates the feature book.
A number of studies have focused on the design of continuous media, CM, (e.g., video and audio) servers to support the real-time delivery of CM objects. These systems have been deployed in local environments such as hotels, hospitals and cruise ships to support media-on-demand applications. They typically stream CM objects to the clients with the objective of minimizing the buffer space required at the client site. This objective can now be relaxed due to the availability of inexpensive storage devices at the client side. Therefore, we propose a Super-streaming paradigm that can utilize the client side resources in order to improve the utilization of the CM server. To support super-streaming, we propose a technique to enable the CM servers to deliver CM objects at a rate higher than their display bandwidth requirement. We also propose alternative admission control policies to downgrade super-streams in favor of regular streams when the resources are scarce. We demonstrate the superiority of our paradigm over streaming with both analytical and simulation models.
Moreover, new distributed applications such as distant-learning, digital libraries, and home entertainment require the delivery of CM objects to geographically disbursed clients. For quality purposes, recently many studies proposed dedicated distributed architectures to support these types of applications. We extend our super-streaming paradigm to be applicable in such distributed architectures. We propose a sophisticated resource management policy to support super-streaming in the presence of multiple servers, network links and clients. Due to the complexity involved in modeling these architectures, we only evaluate the performance of super-streaming by a simulation study.
Recently, researchers have mainly been interested only in the search for data content that are globally similar to the query and not in the search for inside data items. This paper presents an algorithm, called a generalized virtual node (GVN) algorithm, to search for data items where parts (subdatatype) are similar to the incoming query. We call this “subdatatype”-based multimedia retrieval. Each multimedia datatype, such as image and audio is represented in this paper as a k-dimensional signal in the spatio-temporal domain. A k-dimensional signal is transformed into characteristic features and these features are stored in a hierarchical multidimensional structure, called the k-tree. Each node on the k-tree contains partial content corresponding to the spatial and/or temporal positions in the data. The k-tree structure allows us to build a unified retrieval model for any types of multimedia data. It also eliminates unnecessary comparisons of cross-media querying. The experimental results of the use of the new GVN algorithm for “subaudio” and “subimage” retrievals show that it takes much less retrieval times than other earlier algorithms such as brute-force and the partial-matching algorithm, while the accuracy is acceptable.
In this paper we are using Fast Fourier Transform to generate the feature vector which considers the mean real and mean imaginary parts of complex numbers of polar coordinates in frequency domain. The method proposed here considers 12 mean values of 6 upper half sectors real and imaginary parts of each R, G and B components of an image. The algorithm proposed uses 36 mean values of real and imaginary parts in total. The proposed work experimented over a database of 249 images spread across 10 classes of images. Euclidian distances between the feature vectors of query image and the database images are considered. Images are retrieved in ascending order of Euclidian distances. The Average precision and Average recall of each class and overall average of all averages of each class are calculated as a performance measure. The cross over point of average recall and precision is 50% and it is 40% or above for all classes.
Many shot boundary detection methods have been proposed in recent years. Comparison studies have revealed that, among the existing methods, the local histogram-based method gives the highest precision rate if the highest shot boundary detection rate is not the first priority. In this paper, we present a new shot boundary detection method that surpasses the local histogram-based method in both the boundary detection rate and the precision rate. We propose a novel method of extracting color distributions from video frames to effectively suppress both the intra-frame and the inter-frame color variations caused by noise, minor illumination changes, as well as camera and object motions. We also provide a systematic approach to combine the color and the spatial features of video frames for more accurate shot boundary detection. An experimental evaluation confirms the superior performance of our proposed method
Multimedia data is generally stored in compressed form in order to
efficiently utilize the available storage facilities. Access to archives
is dependent on our ability to browse compressed multimedia
information-retrieval and tracking from coded video databases. In this
paper, a novel visual search engine for video retrieval and tracking
from compressed multimedia databases is proposed. The goal of the
project is the implementation of a visual browser that operates in a
distributed environment where users initiate video searches and retrieve
relevant video information simultaneously from multiple video archives.
Being presented with a query in the form of template images of objects,
the system operates on the compressed video to find the images or video
sequences where those objects are present and their positions in the
image. Upon user's request, the system will decompress and display only
the video sequences of interest
A novel algorithm for video retrieval and tracking from compressed
multimedia databases is proposed. Our approach exploits the structure of
video compression standards in order to perform object matching directly
on the compressed video data. This is achieved by utilizing motion
compensation-a critical prediction filter embedded in video compression
standards-to estimate and interpolate the desired method for template
matching
Ritz vectors approximate eigenvectors that are a common choice for
primary images in content based indexing. They can be computed
efficiently even when the images are accessed through slow communication
such as the Internet. We develop an algorithm that computes Ritz vectors
in one pass through the images. When iterated, the algorithm can recover
the exact eigenvectors. In applications to image indexing and learning
it may be necessary to compute primary images for indexing many
sub-categories of the image set. The proposed algorithm can compute
these age data. Similar computation by other algorithms is much more
costly even when access to the images is inexpensive
Semantic filtering and retrieval of multimedia content is crucial
for efficient use of the multimedia data repositories. Video query by
semantic keywords is one of the most difficult problems in multimedia
data retrieval. The difficulty lies in the mapping between low-level
video representation and high-level semantics. We therefore formulate
the multimedia content access problem as a multimedia pattern
recognition problem. We propose a probabilistic framework for semantic
video indexing, which call support filtering and retrieval and
facilitate efficient content-based access. To map low-level features to
high-level semantics we propose probabilistic multimedia objects
(multijects). Examples of multijects in movies include explosion,
mountain, beach, outdoor, music etc. Semantic concepts in videos
interact and to model this interaction explicitly, we propose a network
of multijects (multinet). Using probabilistic models for six site
multijects, rocks, sky, snow, water-body forestry/greenery and outdoor
and using a Bayesian belief network as the multinet we demonstrate the
application of this framework to semantic indexing. We demonstrate how
detection performance can be significantly improved using the multinet
to take interconceptual relationships into account. We also show how the
multinet can fuse heterogeneous features to support detection based on
inference and reasoning
A multimedia micro-university supports the education programs and
management operations of a small academic institution. The authors
describe its important components-a virtual library, an intelligent
distance learning system, and visualization and planning tools for
management operations-and discuss an assessment plan designed to
evaluate the micro-university's effectiveness
In multiuser multimedia information systems (e.g.,
movie-on-demand, digital-editing), scheduling the retrievals of
continuous media objects becomes a challenging task. This is because of
both intra and inter lobject time dependencies. Intraobject time
dependency refers to the real-time display requirement of a continuous
media object. Interobject time dependency is the temporal relationships
defined among multiple continuous media objects. In order to compose
tailored multimedia presentations, a user might define complex time
dependencies among multiple continuous media objects with various
lengths and display bandwidths. Scheduling the retrieval tasks
corresponding to the components of such a presentation in order to
respect both inter and intra task time dependencies is the focus of this
study. To tackle this task scheduling problem (CRS), we start with a
simpler scheduling problem (ARS) where there is no inter task time
dependency (e.g., movie-on-demand). Next, we investigate an augmented
version of ARS (termed ARS+) where requests reserve displays
in advance (e.g., reservation-based movie-on-demand). Finally, we extend
our techniques proposed for ARS and ARS+ to address the CRS
problem. We also provide formal definition of these scheduling problems
and proof of their NP-hardness
Multimedia digital libraries involve different types of data objects such as text, audio, video and images. Most of these data objects are very large in size and accessing them in a distributed environment causes a transmission delay due to the vast amount of network traffic. Compressing these data objects before transmission, can reduce the response time, although it would mean a reduction in the quality of the output data. If the application doesn't demand a high quality output, data compression can be an acceptable means of reducing transmission time over the internet or other distributed environments. The loss of quality would be proportional to the amount of compression applied to individual data objects. Therefore different quality levels (browsing levels) can be achieved depending on how the data will be used. A lower quality level could be used for general browsing of data whereas a higher quality level could be used where the output data has to be further process...
Efficient utilization of network resources is essential in the provision of quality multimedia services. Different approaches have been proposed to shape the multimedia streams as a transmission schedule with smoothed traffic burst. In this paper, we concentrate on the problem of efficient use of bandwidth and client buffer. For this purpose, an optimal transmission schedule is proposed to provide the minimal allocation of the bandwidth and maximal utilization of the client buffer. Simulation results show that the shaping results obtained can dynamically adjust the transmission rate according to the buffer packet sizes and playback rates at the client to avoid the loss of packets, and at the same time achieve the minimal bandwidth allocation and maximal utilization of the client buffer. 1.
ResearchGate has not been able to resolve any references for this publication.