ArticlePDF Available

Abstract and Figures

Phonetic symbols describe movements of the vocal tract, tongue and lips, and are combined into complex movements forming the words of language. In music, vocables are words that describe musical sounds, by relating vocal movements to articulations of a musical instrument. We posit that vo-cable words allow the composers and listeners to engage closely with dimensions of timbre, and that vocables could see greater use in electronic music interfaces. A preliminary system for controlling percussive physical modelling syn-thesis with textual words is introduced, with particular ap-plication in expressive specification of timbre during com-puter music performances.
Content may be subject to copyright.
Words, Movement and Timbre
Alex McLean and Geraint Wiggins
Center for Cognition, Computation and Culture
Department of Computing
Goldsmiths College
{ma503am,g.wiggins}@gold.ac.uk
Abstract
Phonetic symbols describe movements of the vocal tract,
tongue and lips, and are combined into complex movements
forming the words of language. In music, vocables are words
that describe musical sounds, by relating vocal movements
to articulations of a musical instrument. We posit that vo-
cable words allow the composers and listeners to engage
closely with dimensions of timbre, and that vocables could
see greater use in electronic music interfaces. A preliminary
system for controlling percussive physical modelling syn-
thesis with textual words is introduced, with particular ap-
plication in expressive specification of timbre during com-
puter music performances.
Keywords: Vocable synthesis, timbre
1. Introduction
A human speaker can produce timbre of subtlety and range
at great speed, through movements of their vocal tract,
tongue and lips. Perhaps even more impressive is the abil-
ity of a listener to derive the state of a speaker’s vocal tract
from the sound produced, evident in their reconstruction of
the phonemes, words and phrases that a speaker intends to
communicate [1].
Perception of speech is distinct from auditory perception.
This is made clear by sine wave speech [2], where the sound
signal of speech is reduced to the variance of just four sine
waves. The frequency and amplitude of three sine waves
are mapped from the lowest three formant frequencies, and
the fourth sine wave from a fricative formant. The result is a
bistable illusion, where a human subject perceives the sound
as a kind of formless burbling until they are primed with the
original speech signal they then perceive the sine waves
as intelligible speech. Distinct speech perception is further
demonstrated by the McGurk-MacDonald effect [3], where
for example seeing lip movements for the word ‘ga’ while
hearing the word ‘ba’ causes the listener to experience the
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies
are not made or distributed for profit or commercial advantage and that
copies bear this notice and the full citation on the first page. To copy
otherwise, to republish, to post on servers, or to redistribute to lists
requires prior specific permission and/or a fee.
NIME09, June 3-6, 2009, Pittsburgh, PA
Copyright remains with the author(s).
word ‘da’, produced by an imagined articulation halfway
between ‘ga’ and ‘ba’. Here the phoneme ‘da’ exists only
as a speech percept influenced by both visual and auditory
stimuli.
Here we focus on vocal speech, but it is important to note
that the vocal tract is not unique in its expressive encoding
of symbols with movement. Deaf culture has produced sign
languages where symbols are represented by movement of
the hands and face, yet otherwise exhibit all the features of
a spoken human language, including grammar, pragmatics
and metaphor [4]. It may seem odd to mention languages
for the Deaf in a paper about musical interfaces, but we do
so to support our placement of movement at the heart of the
phonetic categories which support language. On this basis
we argue that the symbolic classification of movement is a
general function of the brain, an ability useful in musical
expression, as we will see in the next section.
2. Vocables musical words
In musical tradition, vocable words are those used to de-
scribe an articulation of a musical instrument. An instruc-
tor may use their voice to describe the sound their student
should try to make on their violin, perhaps by singing the
pitch contour while using a particular consonant-vowel pat-
tern to indicate a particular bowing technique. Over time
the student will be able to perceive the phonetics of their
instructor’s voice as sound categories of their instrument.
A formal system of vocables for use in the scores of west-
ern classical music has been proposed and used by Donald
Martino [5] but has so far not found wider use. However
many of the oldest musical cultures have well-developed
formal systems in widespread use. Indian classical music
has the bol syllables [6] where, for example, t
.erepresents a
non-resonating stroke with the 1st finger on the centre of the
d¯
ahin¯
a (right hand) drum. In the Scottish Highlands we find
canntaireachd of the bagpipes [7], for example hiaradalla
represents an echo of the D note in the McArthur cann-
taireachd dialect. In China the delicate finger techniques of
the guqin (Chinese zither) is notated using the chien-tzˆ
u. For
example ch’¨
uan-fu indicates that the index, middle and ring
finger each pull a different string with a light touch, making
the three strings produce one sound ‘melting’ together.
In her doctoral thesis “Non-lexical vocables in Scottish
traditional music” [8], Chambers terms formalised vocables
NIME 2009276
as culturally jelled and ad-hoc vocables as improvisatory,
acknowledging that the line between the two is sometimes
blurred. Chambers goes on to term onomatopoeic vocables
as imitative and more arbitrarily assigned vocables as asso-
ciative. At this point we hit upon major perceptual issues,
as Chambers reports: “Occasionally a piper will say that
a vocable is imitative (indigenous evaluation) when analy-
sis seems to indicate that it is actually associative (analytic
evaluation) because he has connected the vocable with the
specific musical detail for so long that he can no longer di-
vorce the two in his mind” [8, p. 13]. In other words, a
vocable may appear to mimic an instrumental sound on the
perceptual level, when on the level of the sound signal it
does not. This seems to be true in the general context of
onomatopoeia for example where a native English speaker
hears a hen say “cluck”, his German neighbour may per-
ceive the same sound as “tock”. Research into tabla bols
have however found them to be genuinely imitative, sharing
audio features with the instrumental sounds they represent,
identifiable even by naive listeners [9]. Further correlation
has been found between the words commonly used to de-
scribe vowel-like quality of guitar sounds and the associated
mouth shapes [10].
A third distinction can be made between vocable words
in written and spoken form. A reader of a vocable applies
paralinguistic phrasing not derived directly from the text,
but nonetheless with great musical importance. Conversely
a transcriber may resolve ambiguity in a spoken vocable, by
writing a precise interpretation of what was intended. This
issue is of course common to all symbolic notation systems.
We can say however that to some degree a written vocable
may capture the essence of a sound.
3. Vocables as Musical Interface
Electronic music allows sound synthesis free from the con-
straints of physical instruments. However this freedom
presents the problem of how to create new interfaces to con-
trol the new synthesis parameters. Fruitful research into
‘tangible’ physical interfaces to synthesisers is ongoing, but
vocable words offer an alternative approach, which a few
have explored.
From the early 1980s David Evan Jones has played on
the boundary between auditory and speech perception in his
music, for example by using the CHANT software [11] to
apply vowel like quality to instrumental sounds. In “Speech
Extrapolated” he describes how he leads the listener to per-
ceiving non-speech as speech, and vice-versa [12]. Speech
synthesis has also featured as a source of musical timbre in
electronic dance music, for example the largely unintelligi-
ble singing synthesis software written by Chris Jeffs for use
in his compositions under the ‘cylob’ moniker [13]. Gen-
eral use tools are rare, but pioneering research into voice-
control of synthesisers with vocable words is provided by
Jordi Janer, where syllables are sung into a microphone and
the sound signal analysed and mapped to instrumental pa-
rameters [14].
Of course innovation in vocable expression continues out-
side of electronic music. Luciano Berio’s Sequenza III is a
vocal piece featuring mutterings, clicks and shouts of the
female voice, notated with a unique system of symbols in-
cluding vocables. The manipulation of the voice is taken to
different extremes in human beatboxing where extended vo-
cal techniques are employed to produce convincing imper-
sonations of drum machines and bass lines [15]. Beatbox
rhythms may be notated with a system of vocables called
standard beatbox notation [16].
4. Vocables as metaphor
Our particular interest in vocables is that sounds and words
may be perceptually related through imagined, simulated
movement. This argument is supported by work in other
fields. In neuropsychology, research by V.S. Ramachandran
into synaesthesia has found ‘cross-wiring’ between sensory
and/or conceptual maps of the brain to be an evolutionary
trait, particularly common in artists [17]. Synaesthetic cross-
wiring is characterised as an extreme case of the normal
brain’s ability to draw metaphors. Ramachandran posits that
cross-wiring could even have provided the original evolu-
tionary stepping stone to language itself.
From cognitive science, Peter G¨
ardenfors proposes the
theory of conceptual spaces [18], placing a level of concep-
tual representation between the symbolic level (of compu-
tation) and sub-symbolic level (as commonly modelled by
artificial neural networks). This level represents concepts
in geometry, where distance represents dissimilarity, sets of
dimensions form conceptual domains and shapes represent
properties of concepts. This accords well with Ramachan-
dran’s account, indeed G¨
ardenfors characterises metaphor
in much the same way, working it into a theory of seman-
tics. The application of the theory of conceptual spaces to
music and creativity is covered in greater detail in an earlier
co-written paper [19].
These are powerful ideas that could have great conse-
quences for future artistic practice. Indeed, this cross-wiring
could provide a neural basis for exactly the kind of cross-
domain mapping that our research into vocable synthesis
aims to exploit.
5. Vocable Synthesis
We term vocable synthesis as the process where vocable
words are specified in written form, which are mapped to
articulations of a physical model of a (real or imagined) mu-
sical instrument. The use of physical modelling synthesis
allows us to posit that a listener can perceive time variance
of audio features as physical movement. The musician is
then describing movement with words, which the listener
experiences through the medium of timbre.
277
Vocable synthesis was introduced in an earlier paper [20]
and artwork[21], where Karplus-Strong percussive synthe-
sis is controlled by vocable words. Our current system mod-
els a drum head using a 2D waveguide mesh [22] in a tri-
angular geometry for maximum accuracy [23]. The drum
head is excited through interaction with a drumstick, using
a mass-spring model [24]. This model gives greater con-
trol over a broader range of timbre than our previous work.
The drum head has parameters to control the tension and
dampening of the surface, and the drumstick has parameters
to control its stiffness and mass. The drumstick is thrown
against the drum head with parameters controlling the down-
ward velocity,starting x/y position and the angle and veloc-
ity of travel across the drum skin.
Table 1. Mapping of consonants to mallet property
(columns) and movement relative to drum head (rows).
heavy stiff heavy soft light stiff light soft
across q r y s
inward c m f w
outward k n v z
edge x d t b
middle j g p h/l
Table 2. Mapping of vowels to drum head tension (columns)
and dampening (rows).
tense loose
wet i u
a
dry e o
Vocable words are composed from the 26 letters of the
modern English alphabet. The consonants map to the drum-
stick and movement parameters, and vowels to the drum
head parameters, shown in Tables 1 and 2. While this map-
ping is largely arbitrary, the consonant/vowel organisation is
inspired by the International Phonetic Alphabet [25], where
vowels map to the position of the tongue and pulmonic con-
sonants to the place and manner of articulation.
As an example, the articulation “Hit loose, dampened
drum outwards with heavy stiff mallet, then hit the middle
of the drum with a lighter mallet while tightening the skin
slightly and finally hit the edge of the skin with the same
light mallet while loosening and releasing the dampening”
is expressed with the single vocable word “kopatu”.
5.1. Vocable rhythms
Vocable rhythms are implemented using syntax derived by
the Bol Processor [26]. A sequence of vocables are sepa-
rated with white space, with rests denoted with hyphens:
ba da - bing - -
Once the user types in such a sequence, it is played on
a loop until the next sequence is entered. Sequences can be
grouped together into polyphony, by separating sequences
with commas, and surrounding them with braces:
{ba da bing, pip rrrre}
As the sequences in this example are of different lengths,
rests are automatically inserted to pad them out to the length
of the lowest common multiplier. The resulting polymetric
structure is:
ba - da - bing -
pip - - rrrre - -
Note that ‘ba’ and ‘pip’ co-occur on the same measure,
requiring a polyphonic articulation as described in §5.2.2.
If square brackets are used rather than braces, then the se-
quences are repeated in order to fit, like so:
ba da bing ba da bing
pip rrrre pip rrrre pip rrrre
The sequences may be nested to create complex poly-
rhythms from simple parts.
5.2. Vocable manipulation and analysis
Now we have described the representation and mapping of
vocables in our system, we examine ways of analysing and
manipulating vocable words.
5.2.1. Symbolic level
As written vocables represent sounds in symbolic form, we
have a wide range of techniques from computer science avail-
able to us. For example we may analyse sequences of voca-
bles using Markov models and other statistical techniques.
An approach to modelling structures of vocable rhythms in
order to generate rhythmic continuations is introduced in
earlier work [27].
We may also use standard text manipulation techniques
such as regular expressions (regex). Regexes are written
in concise and flexible language allowing general purpose
rule-based string matching [28]. We have embedded a regex
parser in our system, allowing operations such as the fol-
lowing:
˜%3=0 /[aeiou]/to/ fe be
This replaces the vowels of every third vocable with the
string ‘to’, resulting in the following sequence:
fto be fe bto fe be
5.2.2. Geometrical level
Our vocables are direct mappings from the geometry of a
drum and its articulation. It is therefore straightforward to
move from a symbolic to geometric representation, in order
to perform spatial analyses and manipulations.
278
Combining vocables in polyphonic synthesis is straight-
forward, and implemented in our current system as follows.
As consonants control the movement and mallet material,
we allow two consonants to be synthesised concurrently sim-
ply by using multiple mallets in our model. Currently we
allow up to five active mallets per drum, allowing five con-
sonants to be articulated at the same time. As vowels control
the properties of a single drum head, we combine them sim-
ply by taking the mean average of the values they map to.
We may exploit both symbolic and geometric vocable
representations in one operation. For example we could es-
timate the perceptual similarity of two vocable words of dif-
ferent lengths with an approach similar to the symbolic Lev-
enshtein edit distance [29], with edits weighted by phoneme
similarity on the geometrical level. Producing such an algo-
rithm is left for future work.
6. Conclusion
The work of many of those cited in this paper, in particular
Ramachandran, G¨
ardenfors and Patel, stands out by looking
not for oppositions between geometric and symbolic rep-
resentations, or between language and music sounds, but
in the comparisons and interactions between them. In this
spirit we have shown a musical interface that allows sym-
bolic control of a geometrical model in a manner that we
hope is well matched to human perception and production
of instrumental sounds.
Work is ongoing to greater understand the perception of
vocables through experiment, while exploring vocable syn-
thesis through artistic practice, in particular live coding im-
provisation [30].
Video demonstrations, and GNU public licensed source
code for the software described here is available on-line from;
http://yaxu.org/category/vocable/.
References
[1] A. M. Liberman and I. G. Mattingly, “The motor theory of
speech perception revised, Cognition, vol. 21, pp. 1–36, Oc-
tober 1985.
[2] R. E. Remez, J. S. Pardo, R. L. Piorkowski, and P. E. Ru-
bin, “On the bistability of sine wave analogues of speech,
Psychological Science, vol. 12, no. 1, pp. 24–29, 2001.
[3] H. Mcgurk and J. W. Macdonald, “Hearing lips and seeing
voices, Nature, vol. 264, no. 246-248, 1976.
[4] R. Sutton-Spence and B. Woll, The Linguistics of British
Sign Language: An Introduction. Cambridge University
Press, 1999.
[5] D. Martino, “Notation in general-articulation in particular,”
Perspectives of New Music, vol. 4, no. 2, pp. 47–58, 1966.
[6] J. Kippen, The Tabla of Lucknow - A cultural analysis of a
musical tradiation. Cambridge University Press, 1988.
[7] J. F. Campbell, Canntaireachd : articulate music. Archibald
Sinclair, 1880.
[8] C. K. Chambers, Non-lexical vocables in Scottish traditional
music. PhD thesis, University of Edinburgh, 1980.
[9] A. D. Patel and J. R. Iversen, “Acoustic and perceptual com-
parison of speech and drum sounds in the north indian tabla
tradition: An empirical study of sound symbolism,” in 15th
International Congress of Phonetic Sciences (ICPhS), 2003.
[10] C. Traube and N. D’Alessandro, “Vocal synthesis and graph-
ical representation of the phonetic gestures underlying guitar
timbre description,” in 8th International Conference on Dig-
ital Audio Effects (DAFx’05), pp. 104–109, 2005.
[11] X. Rodet, Y. Potard, and J. B. Barriere, “The chant project:
From the synthesis of the singing voice to synthesis in gen-
eral,” Computer Music Journal, vol. 8, no. 3, 1984.
[12] D. E. Jones, “Speech extrapolated,” Perspectives of New Mu-
sic, vol. 28, no. 1, pp. 112–142, 1990.
[13] C. Jeffs, “Cylob music system.” http://durftal.com/
cms/cylobmusicsystem.html, 2007.
[14] J. Janer, Singing-driven Interfaces for Sound Synthesizers.
PhD thesis, Universitat Pompeu Fabra, Barcelona, 2008.
[15] D. Stowell, “Characteristics of the beatboxing vocal style,”
tech. rep., Queen Mary, University of London, 2008.
[16] Tyte, “Standard beatbox notation.” online; http://www.
humanbeatbox.com/tips/p2_articleid/231,
2008.
[17] V. S. Ramachandran and E. M. Hubbard, “Synaesthesia a
window into perception, thought and language,” Journal of
Consciousness Studies, vol. 8, no. 12, pp. 3–34, 2001.
[18] P. G¨
ardenfors, Conceptual Spaces: The Geometry of
Thought. The MIT Press, March 2000.
[19] J. Forth, A. McLean, and G. Wiggins, “Musical creativity on
the conceptual level, in IJWCC 2008, 2008.
[20] A. McLean and G. Wiggins, “Vocable synthesis, in Pro-
ceedings of International Computer Music Conference 2008,
2008.
[21] A. McLean, “Babble.” online artwork; http:
//project.arnolfini.org.uk/projects/
2008/babble/, 2008.
[22] S. Van Duyne and J. O. Smith, “The 2-d digital waveguide
mesh,” in Applications of Signal Processing to Audio and
Acoustics, pp. 177–180, 1993.
[23] F. Fontana and D. Rocchesso, “Signal-theoretic character-
ization of waveguide mesh geometries for models of two-
dimensional wave propagation in elastic media, Speech and
Audio Processing, vol. 9, no. 2, pp. 152–161, 2001.
[24] J. A. Laird, The Physical Modelling of Drums using Digital
Waveguides. PhD thesis, University of Bristol, 2001.
[25] P. Ladefoged, “The revised international phonetic alphabet,
Language, vol. 66, no. 3, pp. 550–552, 1990.
[26] B. Bel, “Rationalizing musical time: syntactic and symbolic-
numeric approaches,” in The Ratio Book (C. Barlow, ed.),
pp. 86–101, 2001.
[27] A. McLean, “Improvising with synthesised vocables, with
analysis towards computational creativity,” Master’s thesis,
Goldsmiths College, University of London, 2007.
[28] J. Friedl, Mastering Regular Expressions. O’Reilly Media,
Inc., August 2006.
[29] V. I. Levenshtein, “Binary codes capable of correcting dele-
tions, insertions and reversals.,” Soviet Physics Doklady.,
vol. 10, no. 8, pp. 707–710, 1966.
[30] N. Collins, A. McLean, J. Rohrhuber, and A. Ward, “Live
coding in laptop performance,” Organised Sound, vol. 8,
no. 03, pp. 321–330, 2003.
279
... patcher language such as Max/MSP) and 'textual' computer language. Attempts to include geometric and natural language representations in the semantics of computer language: the reacTable �Jordà et al., 2007), piet, befunge, inform7, 'postmodern' Perl and vocable synthesis �McLean and Wiggins, 2009). Processes and the activity of programming� The role of feedback and time in programming. ...
... patcher language such as Max/MSP) and 'textual' computer language. Attempts to include geometric and linguistic representations in the semantics of computer language: the reacTable (Jordà et al., 2007), piet, befunge, z code, 'postmodern' Perl and vocable synthesis (McLean and Wiggins, 2009). Processes and the activity of programming. ...
Article
Full-text available
This chapter focuses on the question of how computation is authored as source code, how computer programmers construct discrete symbol sequences within a world of apparently con- tinuous space. If we assume that creative agents are changed as part of their creative processes then this view becomes important: a self-modifying agent is to some degree programming it- self. Anticipated length: 15,000 words Here we consider that performative 1 description of discrete computation known as source code. How source code is created, how it is structured, the cultural context in which this happens and the resulting impacts on creative processes. 1 Background History. A brief section providing historical context to the discussion. Some pre-history about the origins of programming in human computation and textile factories, and human engagement with symbolic representation in music and literature. Overview of the inuences of logic, mathe- matics and linguistics on computer language. An outline of the early development of programming 'hacker' culture at the MIT computer lab which while both funded by the military and heavily male dominated, had a spirit of freedom and openess which inspired the contemporary free software movement (Levy, 2002).
... Vocal emulation of percussion sounds has also been used pedagogically, and as a means of communicating rhythmic motifs. In north Indian musical traditions bols are used to encode tabla rhythms; changgo drum notation is expressed using vocables in Korean samul nori, and Cuban conga players vocalize drum motifs as guauganco or tumbao patterns (Atherton, 2007; McLean and Wiggins, 2009). In contemporary western popular music, human beatboxing is an element of hip hop culture, performed either as its own form of artistic expression, or as an accompaniment to rapping or singing. ...
Article
Full-text available
Real-time Magnetic Resonance Imaging (rtMRI) was used to examine mechanisms of sound production by an American male beatbox artist. rtMRI was found to be a useful modality with which to study this form of sound production, providing a global dynamic view of the midsagittal vocal tract at frame rates sufficient to observe the movement and coordination of critical articulators. The subject's repertoire included percussion elements generated using a wide range of articulatory and airstream mechanisms. Many of the same mechanisms observed in human speech production were exploited for musical effect, including patterns of articulation that do not occur in the phonologies of the artist's native languages: ejectives and clicks. The data offer insights into the paralinguistic use of phonetic primitives and the ways in which they are coordinated in this style of musical performance. A unified formalism for describing both musical and phonetic dimensions of human vocal percussion performance is proposed. Audio and video data illustrating production and orchestration of beatboxing sound effects are provided in a companion annotated corpus.
Chapter
Full-text available
This paper deals with various problems in quantifying musical time encountered both in the analysis of traditional drumming and in computer-generated musical pieces based on "sound-objects", hereby meaning code sequences controlling a real-time sound processor. In section 1 it is suggested that syntactic approaches may be closer to the intuitions of musicians and musicologists than commonly advocated numeric approaches. Further, symbolic-numeric approaches lead to efficient and elegant solutions of constraint-satisfaction problems relative to symbolic and physical durations, as illustrated in sections 2 and 3 respectively.
Article
Full-text available
We investigated grapheme--colour synaesthesia and found that: (1) The induced colours led to perceptual grouping and pop-out, (2) a grapheme rendered invisible through `crowding' or lateral masking induced synaesthetic colours --- a form of blindsight --- and (3) peripherally presented graphemes did not induce colours even when they were clearly visible. Taken collectively, these and other experiments prove conclusively that synaesthesia is a genuine perceptual phenomenon, not an effect based on memory associations from childhood or on vague metaphorical speech. We identify different subtypes of number--colour synaesthesia and propose that they are caused by hyperconnectivity between colour and number areas at different stages in processing; lower synaesthetes may have cross-wiring (or cross-activation) within the fusiform gyrus, whereas higher synaesthetes may have cross-activation in the angular gyrus. This hyperconnectivity might be caused by a genetic mutation that causes defective pruning of connections between brain maps. The mutation may further be expressed selectively (due to transcription factors) in the fusiform or angular gyri, and this may explain the existence of different forms of synaesthesia. If expressed very diffusely, there may be extensive cross-wiring between brain regions that represent abstract concepts, which would explain the link between creativity, metaphor and synaesthesia (and the higher incidence of synaesthesia among artists and poets). Also, hyperconnectivity between the sensory cortex and amygdala would explain the heightened aversion synaesthetes experience when seeing numbers printed in the `wrong' colour. Lastly, kindling (induced hyperconnectivity in the temporal lobes of temporal lobe epilepsy [TLE] patients) may explain the purp...
Thesis
In the context of the live coding of music and computational creativity, literature examining perceptual relationships between text, speech and instrumental sounds are surveyed, including the use of vocable words in music. A system for improvising polymetric rhythms with vocable words is introduced, together with a working prototype for producing rhythmic continuations within the system. This is shown to be a promising direction for both text based music improvisation and research into creative agents.
Thesis
Together with the sound synthesis engine, the user interface, or controller, is a basic component of any digital music synthesizer and the primary focus of this dissertation. Under the title of singing-driven interfaces, we study the design of systems, that based on the singing voice as input, can control the synthesis of musical sounds. From a number of preliminary experiments and studies, we identify the principal issues involved in voice-driven synthesis. We propose one approach for controlling a singing voice synthesizer and another one for controlling the synthesis of other musical instruments. In the former, input and output signals are of the same nature, and control to signal mappings can be direct. In the latter, mappings become more complex, depending on the phonetics of the input voice and the characteristics of the synthesized instrument sound. For this latter case, we present a study on vocal imitation of instruments showing that these voice signals consist of syllables with musical meaning. Also, we suggest linking the characteristics of voice signals to instrumental gestures, describing these signals as vocal gestures. Within the wide scope of the voice-driven synthesis topic, this dissertation studies the relationship between the human voice and the sound of musical instruments by addressing the automatic description of the voice and the mapping strategies for a meaningful control of the synthesized sounds. The contributions of the thesis include several voice analysis methods for using the voice as a control input: a) a phonetic alignment algorithm based on dynamic programming; b) a segmentation algorithm to isolate vocal gestures; c) a formant tracking algorithm; and d) a breathiness characterization algorithm. We also propose a general framework for defining the mappings from vocal gestures to the synthesizer parameters, which are configured according to the instrumental sound being synthesized. As a way to demonstrate the results obtained, two real-time prototypes are implemented. The first prototype controls the synthesis of a singing voice and the second one is a generic controller for other instrumental sounds.
Book
Regular expressions are a central element of UNIX utilities like egrep and programming languages such as Perl. But whether you're a UNIX user or not, you can benefit from a better understanding of regular expressions since they work with applications ranging from validating data-entry fields to manipulating information in multimegabyte text files. Mastering Regular Expressions quickly covers the basics of regular-expression syntax, then delves into the mechanics of expression-processing, common pitfalls, performance issues, and implementation-specific differences. Written in an engaging style and sprinkled with solutions to complex real-world problems, Mastering Regular Expressions offers a wealth information that you can put to immediate use. Regular expressions are an extremely powerful tool for manipulating text and data. They are now standard features in a wide range of languages and popular tools, including Perl, Python, Ruby, Java, VB.NET and C# (and any language using the .NET Framework), PHP, and MySQL. If you don't use regular expressions yet, you will discover in this book a whole new world of mastery over your data. If you already use them, you'll appreciate this book's unprecedented detail and breadth of coverage. If you think you know all you need to know about regular expressions, this book is a stunning eye-opener. As this book shows, a command of regular expressions is an invaluable skill. Regular expressions allow you to code complex and subtle text processing that you never imagined could be automated. Regular expressions can save you time and aggravation. They can be used to craft elegant solutions to a wide range of problems. Once you've mastered regular expressions, they'll become an invaluable part of your toolkit. You will wonder how you ever got by without them. Yet despite their wide availability, flexibility, and unparalleled power, regular expressions are frequently underutilized. Yet what is power in the hands of an expert can be fraught with peril for the unwary. Mastering Regular Expressions will help you navigate the minefield to becoming an expert and help you optimize your use of regular expressions. Mastering Regular Expressions , Third Edition, now includes a full chapter devoted to PHP and its powerful and expressive suite of regular expression functions, in addition to enhanced PHP coverage in the central "core" chapters. Furthermore, this edition has been updated throughout to reflect advances in other languages, including expanded in-depth coverage of Sun's java.util.regex package, which has emerged as the standard Java regex implementation. Topics include: A comparison of features among different versions of many languages and tools How the regular expression engine works Optimization (major savings available here!) Matching just what you want, but not what you don't want Sections and chapters on individual languages Written in the lucid, entertaining tone that makes a complex, dry topic become crystal-clear to programmers, and sprinkled with solutions to complex real-world problems, Mastering Regular Expressions , Third Edition offers a wealth information that you can put to immediate use. Reviews of this new edition and the second edition: "There isn't a better (or more useful) book available on regular expressions." --Zak Greant, Managing Director, eZ Systems "A real tour-de-force of a book which not only covers the mechanics of regexes in extraordinary detail but also talks about efficiency and the use of regexes in Perl, Java, and .NET...If you use regular expressions as part of your professional work (even if you already have a good book on whatever language you're programming in) I would strongly recommend this book to you." --Dr. Chris Brown, Linux Format "The author does an outstanding job leading the reader from regex novice to master. The book is extremely easy to read and chock full of useful and relevant examples...Regular expressions are valuable tools that every developer should have in their toolbox. Mastering Regular Expressions is the definitive guide to the subject, and an outstanding resource that belongs on every programmer's bookshelf. Ten out of Ten Horseshoes." --Jason Menard, Java Ranch
Thesis
In this thesis the physical modelling of percussive drums was approached using digital waveguides. The constituent components of a drum were considered individually before connecting them together to complete the model. To model the drumskin techniques were created to incorporate smooth curved boundaries, calculate the impedance of a 2D waveguide mesh and include the effect of the bearing edge. The accuracy of the curved boundary model, which utilised 'rimguides', was demonstrated with a good reproduction of the first seven resonant modes of a circular membrane. The impedance was used in a kettledrum model where it correctly controlled the exchange of energy between the drumskin and interior air. Simulations of different bearing edge sizes revealed that a blunt edge takes energy from low frequencies and redistributes it into higher frequencies. These decay faster and so the result is a decrease in sustain. For the interior air it was necessary to correctly model 3D wave propagation and incorporate diffuse reflections, which occur at rough surfaces. Unlike 3D meshes used in previous studies, the new dodecahedral mesh proposed here was found to exhibit near direction independent dispersion error. The effect of diffusion was adequately simulated with a technique that was shown to be controllable, enabling different types of surface to be modelled. To complete the drum model a way of connecting different waveguide meshes together was found and a new procedure for modelling a mallet exciter was proposed. The interfacing method enabled a lossless interconnection between two 2D meshes and also 2D and 3D meshes. The procedure used for the mallet exciter incorporated non-linear stiffness and the mallet's contact area. Its behaviour was shown to be almost identical to that of a real mallet. Finally, a digital waveguide model of a kettledrum was constructed to demonstrate the techniques and the results were promising; the resonant modes were reproduced with good accuracy and their decay was sufficient to give the impression of realism, whilst not exactly matching that found through measurement.
Article
Book
An examination of the tabla, a drum in North Indian classical music. This study aims to increase understanding of the Lucknow tabla-playing tradition by demonstrating its technical and musical distinction, and by analyzing the processes involved in composition and improvization.