Figure 1 - uploaded by Meinard Müller
Content may be subject to copyright.
Overview of the retrieval procedure. The sheet music representation of a musical theme and the audio recordings from the database are transformed to chroma features. A DTW-based technique is used to locate the musical themes in the database which results in a ranked list of candidates.  

Overview of the retrieval procedure. The sheet music representation of a musical theme and the audio recordings from the database are transformed to chroma features. A DTW-based technique is used to locate the musical themes in the database which results in a ranked list of candidates.  

Context in source publication

Context 1
... overview of the retrieval procedure is shown in Figure 1. In this example, we use the famous "Fate Motif" from Beethoven's Symphony No. 5 as query. ...

Similar publications

Article
Full-text available
It implemented an Optical-Digital interface, which generates eight types of Fourier holograms. The digital hologram generation and reconstruction is digital, or optical device using a LCR2500. We developed a software application with Graphical User Interface. We present validation results of the implemented system.
Article
Full-text available
The purpose of this paper is to compare the performance of human listeners against the selected machine learning algorithms in the task of the classification of spatial audio scenes in binaural recordings of music under practical conditions. The three scenes were subject to classification: (1) music ensemble (a group of musical sources) located in...
Preprint
Full-text available
In this report, we introduce the task setup and the baseline system for the sub-task B of the DCASE 2022 Challenge Task 6: language-based audio retrieval subtask. For this subtask, the Clotho v2 dataset is utilized as the development dataset, and an additional dataset consisting of 1,000 audio-caption pairs as the evaluation dataset. We train the b...

Citations

... In this chapter, we report on a systematic study considering a cross-modal retrieval scenario which was originally published in [11,4]. Using a musical theme from the book "A Dictionary of Musical Themes" as a query, the objective is to identify all related music recordings from a given audio collection of Western classical music. ...
... Furthermore, as in our scenario, the project offers a web-based service, which allows the playback of enriched videos. 4 In the research project Freischütz Digital, 5 user interfaces for dealing with critical editions in an opera scenario were developed [161,151]. ...
Thesis
Music constitutes a challenging multimedia scenario. Besides music recordings, there exist a number of other media objects including symbolic music representations, video recordings, scanned sheet music, or textual metadata. Developing tools that allow users to retrieve information from different types of music-related data is central to the research area known as Music Information Retrieval (MIR). This requires techniques from various engineering fields such as digital signal processing, image processing, data management, and machine learning. In this thesis, we develop novel multimedia processing techniques and explore their capabilities and limitations within different complex music scenarios. The thesis consists of three main parts. In the first part, we consider retrieval scenarios within a Western classical music setting. For example, given a short monophonic melodic theme in symbolic notation as a query, retrieve all corresponding documents in a collection of polyphonic music recordings. In another related retrieval scenario, we aim to link the score of musical themes, scanned from book pages, to their symbolic counterparts given in MIDI format. Both scenarios require mid-level feature representations derived from different media types, as well as robust retrieval techniques that can handle extraction errors and variations in the data. The second part of this thesis deals with the extraction of musical parameters such as fundamental frequencies or musical pitches from audio recordings. In this context, a general goal is to reduce the variations in the degree of polyphony between monophonic queries and polyphonic music databases. In our computational approach, we propose a data-driven method based on Deep Neural Networks (DNNs) which aims at enhancing salient parts from jazz music recordings. As an example application, we employ the learned model in a retrieval scenario where we take a jazz solo transcription as a query to identify the corresponding music recording. In the third part, we explore the potential of web-based user interfaces for researchers as well as music lovers. We present several prototypical interfaces that offer various functionalities for enabling access and navigation in musical content. Furthermore, these interfaces allow researchers to show their results in an interactive fashion and reduce technical barriers when cooperating with other researchers from related fields such as musicology.