ArticlePDF Available

Abstract

The process of transforming data into sounds for auditory display provides unique user experiences and new perspectives for analyzing and interpreting data. A research study for data transformation to sounds based on musical elements, called data-to-music sonification, reveals how musical characteristics can serve analytical purposes with enhanced user engagement. An existing user engagement scale has been applied to measure engagement levels in three conditions within melodic, rhythmic, and chordal contexts. This article reports findings from a user engagement study with musical traits and states the benefits and challenges of using musical characteristics in sonifications. The results can guide the design of future sonifications of multivariable data.
TYPE Original Research
PUBLISHED 10 August 2023
DOI 10.3389/fdata.2023.1206081
OPEN ACCESS
EDITED BY
Feng Chen,
Dallas County, United States
REVIEWED BY
F. Amilcar Cardoso,
University of Coimbra, Portugal
Minglai Shao,
Tianjin University, China
*CORRESPONDENCE
Jonathan Middleton
jmiddleton@ewu.edu
RECEIVED 14 April 2023
ACCEPTED 17 July 2023
PUBLISHED 10 August 2023
CITATION
Middleton J, Hakulinen J, Tiitinen K, Hella J,
Keskinen T, Huuskonen P, Culver J, Linna J,
Turunen M, Ziat M and Raisamo R (2023)
Data-to-music sonification and user
engagement. Front. Big Data 6:1206081.
doi: 10.3389/fdata.2023.1206081
COPYRIGHT
©2023 Middleton, Hakulinen, Tiitinen, Hella,
Keskinen, Huuskonen, Culver, Linna, Turunen,
Ziat and Raisamo. This is an open-access article
distributed under the terms of the Creative
Commons Attribution License (CC BY). The use,
distribution or reproduction in other forums is
permitted, provided the original author(s) and
the copyright owner(s) are credited and that
the original publication in this journal is cited, in
accordance with accepted academic practice.
No use, distribution or reproduction is
permitted which does not comply with these
terms.
Data-to-music sonification and
user engagement
Jonathan Middleton1,2*, Jaakko Hakulinen2, Katariina Tiitinen2,
Juho Hella2, Tuuli Keskinen2, Pertti Huuskonen2, Jerey Culver3,
Juhani Linna2, Markku Turunen2, Mounia Ziat4and
Roope Raisamo2
1Department of Fine and Performing Arts, Eastern Washington University, Cheney, WA, United States,
2Tampere Unit for Computer-Human Interaction (TAUCHI), Tampere University, Tampere, Finland,
3School of Business, Eastern Washington University, Spokane, WA, United States, 4Department of
Information Design and Corporate Communication, Bentley University, Waltham, MA, United States
The process of transforming data into sounds for auditory display provides
unique user experiences and new perspectives for analyzing and interpreting
data. A research study for data transformation to sounds based on musical
elements, called data-to-music sonification, reveals how musical characteristics
can serve analytical purposes with enhanced user engagement. An existing user
engagement scale has been applied to measure engagement levels in three
conditions within melodic, rhythmic, and chordal contexts. This article reports
findings from a user engagement study with musical traits and states the benefits
and challenges of using musical characteristics in sonifications. The results can
guide the design of future sonifications of multivariable data.
KEYWORDS
data-to-music, sonification, user engagement, auditory display, algorithms
1. Introduction
Music can be a highly engaging art form in terms of pure listening entertainment and,
as such, a powerful complement to theater, film, video games, sports, ballet, ceremonies,
and sacred rituals. So it seems reasonable to assume music has the ability to capture the
focus of people who also listen to data in the form of auditory display. In this article,
we will refer to auditory display as sonification, “the transformation of data relations into
perceived relations in an acoustic signal for the purposes of facilitating communication or
interpretation” (Neuhoff, 2011). The following is a research study conducted to establish how
musical characteristics contribute to engagement with data-to-music sonification. Our study
stems from a 3-year investigation of data-to-music for five companies in Finland that were
seeking innovative ways to present data for their employees and customers. The companies
were represented by industries related to power generation, smart electronics for medical
devices and building monitoring, smart watches, construction, and architecture. The project
began with the development of new sonification software for data transformation called
D2M. The D2M software was used to create sonifications for the user engagement study to
determine the reliability of musical characteristics for enhanced engagement. Engagement
offers a key advantage to auditory displays of data if the roles of musical characteristics and
engagement are understood correctly.
Frontiers in Big Data 01 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
2. Background
2.1. Sonification
The use of musical elements in sonification has been formally
explored by members of the International Community for Auditory
Display since the mid-1990s (Kramer et al., 1999;Hermann et al.,
2011). The interest in incorporating musical characteristics in
sonification can be summarized from the study by Brown et al.
(2003): “the use of musical sounds has been recommended because
of the ease with which musical sounds are perceived.” A summary
of music-related research and software development in sonification
can be obtained from the study by Bearman and Brown (2012).
Earlier attempts were made by Pollack and Ficks (1954), yet
according to Vickers (2017), and other researchers, such as Walker
and Nees (2011), “questions of aesthetics and musicality remain
open in the field of sonification, and therefore the path for
music in sonification remains uncertain. This impression shows
how challenging an interdisciplinary area of research can be. As
an example, Vickers, Hogg, and Worrall submit that the dual
nature of music and analysis is hard to achieve: “A major design
challenge is to create sonifications that are not only effective at
communicating information but which are sufficiently engaging
to engender sustained attention” (Vickers et al., 2017). It appears
from this statement that the process of interpreting data with
musical qualities can contribute to enhanced levels of engagement
but also distract from the analysis. As an example, the concern
can be seen in a condition monitoring study by Hildebrandt et al.
(2016), which did not apply musical contexts due to concerns of
continuous monitoring fatigue with music. Their study reveals how
researchers perceive risks in using musical sounds, but the authors
acknowledge there is a lack of attention to musical aesthetics and
“potentially more pleasing designs” for long-term usability and
effectiveness. Meanwhile, Vickers (2017) suggests the path toward
successful sonifications with musical elements can be assisted by
the knowledge and experience of composers by recommending “an
aesthetic perspective space in which practice in various schools of
music composition might be used to improve the aesthetic design
and interest of sonifications.”
It is important to note that the use of musical characteristics
in sonification can serve multiple purposes. To start, musical
expressions that relate to aesthetics can enhance the user
experience, and this can strengthen the perceptual experience
with data. Positive user experiences can translate to more time
spent with data analysis and improve interpretations. Finally, the
mapping possibilities to musical traits are numerous, and this
brings opportunities to associate multivariate data with different
types of musical features, forms, and expressions.
The study reported in this article was carried out in a
project called Data-to-Music, which focused on the development
of custom-made software to map multivariate data with musical
characteristics. The data came mostly from monitoring the
conditions of buildings, machines, weather, fitness tracking, and
athletic experiences. What made the project unique was the focus
on the user experience with music. The project sought to design
the most effective sonifications with only musical elements rather
than any sounds (synthetic, nature, auditory icons, or earcons)
(Brewster, 2009).
In this article, we first outline our sonification approach with
musical characteristics and engagement as priorities, and then
describe the surveys we used to evaluate user engagement. The
auditory features for the surveys relied on D2M software to
transform weather data into musical elements and expressions in
melodic, chordal, and rhythmic contexts.
2.2. Form and function
In this article, we are guided by the research question “Can
musical characteristics contribute to meaningful data perception,
analysis, and interpretation?” The challenge is to make sure the
aesthetic qualities of musical sounds will maintain or contribute to
the integrity of coherent information in the auditory display. We
submit that data-to-music sonification can not only engage users
but also maintain coherence as a foundation for understanding
combinations of musical characteristics.
In addressing this challenge, we venture into the human–
computer interaction (HCI) domain of form and function or
aesthetics and usability (Norman, 2004;Tractinsky, 2006). Various
studies have revealed that this duality is most likely interconnected
to the extent that a positive perception of usability enhances our
perception of aesthetics, and aesthetic appeal can enhance our
perception of usability (Tractinsky et al., 2000;Tuch et al., 2012).
This association suggests an aesthetic appeal from musical traits
could contribute to increased usability and therefore indicates some
potential advantages for enhancing the quality and quantity of
time users could dedicate to experiences and interactions in data-
to-music. The interpreting experiences with the aid of musical
characteristics offer new perspectives for the exploration of data
through a relatively new form of mediation (with musical ideas)
between humans and data (Dillon, 2006). From this interaction
within arcs of exploration and interpretation, one would hope for
a meaningful connection via engagement to help users analyze
information from data. Engagement’s significance for analysis and
its inherent connections to usability and aesthetics will be addressed
later in the paper.
3. D2M software for data-to-music
sonification
To test the possibilities of musical characteristics for data
representation and display, our research team designed, and
programmed the D2M data-to-music mapping software. The
software enables the users to map variables from time series
datasets to musical characteristics in independent tracks we call
“streams.” In each track, users can make selections for instrument
type (or timbre), pitch range, rhythmic complexity, accent chords,
articulation, and loudness. The user can also modify the duration of
the generated sonification, and the data can be scaled accordingly.
The wide selection of mapping options provides users with
opportunities to hear data from many musical attributes. The
musical structures are generated as MIDI output from the data
content in combination with mapping selections for each stream.
Streams can represent one or many datasets from a data bank, e.g.,
Frontiers in Big Data 02 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
one stream might represent cloud cover and another stream might
represent wind speed, and users could hear these combined.
Additional tools include the possibility of setting thresholds to
filter out segments of the data. There are also preprocessing tools
for linear and logarithmic scaling and a tool to invert data values.
The high degree of flexibility of the tools enabled the research team
to create auditory displays that sound composed, even though the
results were determined by the data and algorithms.
While the algorithms for the D2M software were researched
and developed over 3 years at TAUCHI (Middleton et al., 2018),
the designs were informed by 10 years of experiences from users
and composers with the music algorithms Web-based software
(https://musicalgorithms.org/4.1/app/#) (Middleton and Dowd,
2008;Bywater and Middleton, 2016). The data for testing and
developing the D2M software were provided by companies with the
goal of developing proof-of-concepts that would allow users to hear
and interpret data through musical forms of expression. The D2M
development project was funded by a Tekes innovation grant.
4. User engagement
Multiple studies have shown that the task of defining and
measuring engagement is quite complex, and many evaluations are
connected to the field of education to understand how students
are engaged in learning (Lutz Klauda and Guthrie, 2015). In
various studies on engagement, common descriptive terms, such
as motivation, persistence, and effort, emerge. Lutz Klauda and
Guthrie (2015) elegantly separate motivation as goal-oriented,
based on values and beliefs, a mode of “behavioral displays of effort,
time, and persistence in attaining desired outcomes.” Additional
attributes from other studies include focused attention, curiosity,
novelty, and challenge (Webster and Ahuja, 2006;O’Brien and
Toms, 2008).
The most seminal work on engagement evaluation appears to
be by O’Brien and Toms (2008,2010). The O’Brien and Toms
study defines engagement in extensive detail by building on the
premise that engagement is a process in three stages, namely, point
of engagement, sustained engagement, and disengagement, and as
the process unfolds in time, there are multiple layers of experience
called threads. The categorization of threads is derived from the
work of McCarthy and Wright (2004).
The O’Brien and Toms study from 2008 sought to refine the
threads of experience: Compositional (narrative), spatiotemporal,
emotional, and sensual, into six key factors for engagement. Their
research generated questionnaire items for a User Engagement
Scale (UES) consisting of 31 statements associated with the
six engagement factors (O’Brien and Toms, 2010). O’Brien and
Toms (2008,2010) built their engagement model mainly from
visual displays related to Web searching, video games, online
learning, and online shopping. Our Data-to-Music study adopted
the O’Brien and Toms’ engagement instrument from 2008 to
2010 to measure user engagement from musical characteristics
and contexts in auditory displays of data. In O’Brien et al. (2018)
published an updated user engagement scale called a “short-form
framework” with an attempt to consolidate the six engagement
factors down to four factors. They also sought to reduce the
number of questionnaire statements from 31 to 12. Our user
engagement study for data-to-music was already in progress by
this time, so we will report findings based on the long-form UES
with six factors (focused attention,perceived usability,aesthetics,
endurability,novelty, and felt involvement) and 31 statements.
In addition to demonstrating engagement levels from musical
characteristics in sonifications, we hope our study, based on the
UES long form, will provide some observations and comparisons
to inform the UES short form’s objectives. In particular, our results
may provide insights toward the decision to consolidate UES
factors for endurability,novelty, and felt involvement into one factor
called “reward” (O’Brien et al., 2018).
5. Methods
5.1. Three engagement surveys
From 2017 to 2018, three user engagement surveys were
conducted with 72 human subjects. The purpose of the surveys was
to capture the impact of basic musical characteristics, such as pitch,
rhythm, and timbre, in auditory displays of data in the contexts of
melodic, chordal, and rhythmic sonifications.
There were 24 participants per study (N=72). Participants’
ages ranged from 18 to 56 years. Nearly 53% had more than a year
of musical training, but only 29% of all participants considered
themselves semi-professional or professionally trained musicians.
Nearly 78% of participants across the three studies were unfamiliar
with data-to-music sonification based on the question “Have
you heard data sonifications before this experiment.” Additional
background information includes 33 male participants and 38
female participants, and one was of a non-specified gender. All
participants rated how they were feeling as they started the surveys
based on five descriptions, namely, sad, a bit sad, neutral, a bit
happy, and happy. All were within the range of neutral to happy.
5.2. Three musical contexts
Three surveys were designed to place participants in the context
of either melody, chords, or rhythm as defining features for data
mapping. In converting data to melodic contexts, the focus was
on pitch mapping, where high and low numeric data values were
mapped to high and low pitches. In this context, the auditory
display unfolds in a linear, melodic line of sound. In mapping to the
context of chords, pitch mapping is executed in the same manner
as melody, but an array of tones is added vertically to the melodic
line. In D2M, chords are featured as a composite sound of two
to four pitches perceived as one sonority (or sound entity), and
these vertical sonorities flow in a harmonic sequence. The Chord-
based survey used standard usage of dyads, tetrachords, and triadic
combinations of tones that are found in tonal Western music,
but the study did not attempt to apply harmonic principles that
relate to harmonic syntax, semantics, or voice leading (Aldwell
and Schachter, 2011). Instead, the D2M algorithm mapped data to
generic chord structures with some variability of chord tone density
and attack, i.e., the chords could be presented as a simultaneity with
two, three, or four tones all at once or rolled with a quick linear
unfolding of the chord.
Frontiers in Big Data 03 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
FIGURE 1
D2M interface showing how clouds data in okta were inverted so that cloud cover would represent low pitch ranges to express “deep” or “heavy”
impressions and sunshine would represent high pitch ranges for brightness. The top section represents the data, the middle section represents the
inverted data with auditory stream mappings, and the bottom section shows how the auditory stream could have been mapped without the inversion
(along with a box showing the rhythmic selection of the moderate level “2” distribution of rhythms). Each result was 40 s in duration with an
underlying tempo of 60 beats per minute.
The context of Rhythm featured rhythmic mapping, i.e., data
mapped to rhythmic durations assigned to a uniform noise or
pitched sound. A data point could be mapped to a rhythmic value
or a rhythmic motive. In general, a set of low data points would
generate long rhythmic durations (showing slow activity), and by
contrast, a set of high data points would generate short rhythmic
durations (showing accelerated activity).
Although rhythm was a featured context for one survey, it was
also one of three conditions for all three surveys. The musical
context of Rhythm and the use of rhythm as a key element for
a survey condition were not mutually exclusive, so in order to
provide some differentiation, the Melody and Chords surveys used
a moderate scope of rhythmic activity as a condition, and a more
wide scope of rhythmic activity was used for the entire Rhythm
survey. Details are provided in the conditions section.
5.3. Weather data sources and listening
tasks
The sonifications for the user engagement surveys used data
derived from weather forecasting. In the Melody and Chords
surveys, the data were derived from cloud cover (called oktas),
while the Rhythm survey used data from wind speeds. The wind
speed and oktas data came from 24-h periods in three different
months of the year, namely, April, May, and November. The
data came from the Finnish national weather service database
(https://en.ilmatieteenlaitos.fi/statistics-from-1961-onwards). The
data from three separate months showed rather diverse weather
patterns, and this enabled the studies to present mostly active,
moderately active, and mostly static sonification experiences for
each condition. As an example, the data for May was more static
than that for November (compare Figure 1 with Figure A1).
To determine user engagement, the surveys relied on user tasks
and statement responses upon completion of tasks. One of the tasks
we created was a simple listen and click interface with a question
referencing whether you hear clouds or sunshine. We were able to
use the D2M data inversion tool to generate pitched results from
the data that express amounts of sunshine or cloud cover from the
same data source (Figure 1), a technique similar to one described
by Walker and Kramer (2005). High okta values for extreme
cloud cover would sound low, while low okta values representing
sunshine would sound high and bright after the inversion tool
was applied. For mapping details, see Appendix.
Wind speeds were used for the Rhythm survey since the
mapping process and correlations between wind speeds and
rhythm seemed more appropriate than clouds and rhythm.
5.4. Three conditions per survey
Each survey featured sonifications under three conditions,
which were differentiated by the order of complexity among
three fundamental musical characteristics. The simplest condition
Frontiers in Big Data 04 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
FIGURE 2
The flow of the Melody survey with cloud cover data from three separate months shown as Data 1, Data 2, and Data 3. Participants listened to each
condition with two dierent tasks (answering a question after listening and then clicking while listening in real-time). Then participants responded to
the 31 statements in the UES questionnaire. All three surveys followed the same structure. This flow was repeated three times for each condition. The
survey was randomized on two levels, namely, (1) the order of three conditions and (2) the order of three data sonifications within a condition. The
conditions were expressed as PS, PT, and PTR, referencing pitched sine waves, pitched timbre, and pitched timbre with rhythm. The randomizations
followed an ABC_CBA scheme. The survey duration was about 45min.
featured plain sounds; the next level of complexity featured tones
with timbre; and the most complex condition featured tones with
timbre and rhythm.
Three conditions in detail:
(1) Plain sounds, as defined by tones represented by sine waves
or a singular semi-pitched percussive noise (from claves: a
monotone, partially pitched sound with a woody attack noise).
Plain sounds may be referenced as “noise” or “pure tones.” In
the Melody and Chords surveys, the plain sounds were guided
by one steady rhythmic value of quarter notes (or crotchets)
at a moderate tempo. The D2M software refers to this rhythm
setting as “level 1.” In the Rhythm survey, percussive noise was
used in lieu of pure tones, and the condition featured a “level 4”
rhythmic setting (widening the scope of rhythmic options).
(2) Tones with a uniform sound color, called timbre. In this
condition, listeners hear the same pitch mapping results as in the
plain sound condition, but in this case, there is a timbral color
from a marimba added. This condition with a marimba timbre
was present in all three surveys. The Melody and Chords surveys
maintained the same level 1 rhythmic setting, so that the sounds
were guided by a uniform rhythmic value of quarter notes at a
moderate tempo. In the Rhythm survey, with a focus on data
mapped to rhythms, the condition used the level 4 setting. The
pitches for this condition in the Rhythm survey were set to a
monotone (singular pitch)—not informed by the data.
(3) Tones with sound color and rhythm based on an elevated
rhythmic activity relative to the other conditions. Activity level
2 was used in the Melody and Chords surveys, and level
4 was used for the Rhythm survey. The elevated rhythmic
Frontiers in Big Data 05 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
levels increasingly expand the rhythmic results, such as quarter
notes, eighth notes, eighth note triplets, and sixteenth notes in
a moderate tempo. Although the conditions for the Rhythm
survey were quite similar to the Melody and Chords surveys, the
third condition varied the most in how the number of rhythmic
options at level 4 (instead of level 2) allowed the data to generate
a wider array of complex rhythmic results than the other studies.
This expansion of options allowed the survey to adhere to an
enhanced rhythmic context, e.g., high wind speeds generated
very active rhythmic experiences, while slow wind speeds
generated very passive rhythmic experiences. The conditions for
the Rhythm survey used traditional drum set sounds consisting
of cymbal noises and drum tones, representing a degree of
timbral variety in the context of rhythmic variety. One can
listen and compare all data-to-music audio files for May datasets
located in Supplementary material.
5.5. Survey tasks and statements
The study was designed with two primary tasks, namely, (1)
listening tasks with immediate questions (to give users time to
experience the sounds and respond in a user-friendly environment)
and (2) responding to questionnaire items represented by 31 survey
statements (adopted from the O’Brien and Toms, 2008). The
surveys’ structure and the UES statements are given in Figure 2 and
Table 1, respectively.
Users responded to the 31 statements after a series of listening
tasks related to each condition, namely, (a) plain sounds, (b)
tones with sound color, and (c) tones with sound color and
rhythm presented via ABC-CBA counterbalanced scheme. The
31 statements were completed three times, following the listening
tasks with three sonifications (created from data from three
different months). The engagement involved listening and clicking
when impressions of weather were perceived, for sun, clouds,
and wind (See example in Figure 3). The 31 statements cover six
different engagement factors called focused attention, perceived
usability, aesthetics, endurability, novelty, and felt involvement
(O’Brien and Toms, 2010). Responses were measured on a 5-step
Likert scale from “strongly disagree” to “strongly agree, generating
data that are ordinally scaled.
5.6. Statistical methods for analysis
Survey responses from three conditions, namely, (1) plain
sounds, (2) tones with timbre, and (3) tones with timbre with
rhythm) were collected from a non-parametric (distribution-
free) testing method, and results were calculated by Kendall’s
Concordance Coefficient—Kendall’s W for approximate mean
affect, or agreement among raters (to determine low statistical
dependency), and Friedman’s test for statistically significant
differences among median ranked values in paired groups,
covering all combinations of three conditions per survey. There
were 279 median responses drawn from 31 statements from
three conditions across three surveys. Pairwise comparisons were
made from the median-ranked results from three conditions
TABLE 1 Thirty-one statements for the data-to-music user engagement
scale (UES).
# Statement
1 I was so involved in my listening task that I lost
track of time.
2 The time I spent listening just slipped away.
3 My sound experience was rewarding.
4 I felt interested in my listening task.
5 During this sound experience, I let myself go.
6 When I was listening, I lost track of the world
around me.
7 These sounds appealed to my auditory senses.
8 I could not identify some of the things I needed to
identify from these sounds.
9 I liked the beats and rhythms used in these sounds.
10 If made available, I would continue to listen to
these kinds of sounds out of curiosity.
11 I felt frustrated while listening to these sounds.
12 The content of the sounds incited my curiosity.
13 I felt involved in this listening task.
14 Listening to these sounds was worthwhile.
15 I felt in control of my sound experience.
16 I found these sounds confusing to understand.
17 This sound experience was fun.
18 I consider my sound experience a success.
19 These sounds were aesthetically appealing.
20 This sound experience did not work out the way I
had expected.
21 The sound layout of these sounds was auditorily
pleasing.
22 I felt annoyed while listening to these sounds.
23 I was absorbed in my listening task.
24 These sounds were attractive.
25 I lost myself in this sound experience.
26 I blocked out things around me when I was
listening to the sound data.
27 This sound experience was demanding.
28 I felt discouraged while listening to these sounds.
29 I would recommend listening to these kinds of
sounds to my friends and family.
30 I was really drawn into my listening task.
31 Understanding these sounds was mentally taxing.
The UES statements are modified from the O’Brien and Toms study from 2010 (Figure A1)
to address the context of auditory display. The item number indicates the order in which
the statements were presented to the participants. The asterisks indicate that the item was
reverse-coded in the same manner as the O’Brien and Toms (2010).
for each survey. Friedman’s tests were run to compare whether
the ordinally scaled data would show differences between the
conditions. Statistical significance was determined by the P-value
Frontiers in Big Data 06 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
FIGURE 3
Survey engagement activity. Participants were asked to listen to a sonification, mapped from oktas data, and click when they heard clouds. Upon
each click, a cloud would appear at the time point relative to the audio file. The activities of listening and responding to sonifications served as
references for measuring engagement, which was the primary focus of the survey. The accuracy of the participants’ responses was not measured.
being less than the alpha level of 0.05 (p<0.05), which would
reject the null hypothesis of no significant difference. Adjusted
and non-adjusted significance levels for pairwise comparisons of
conditions were explored (the adjusted significance level applies
aBonferroni correction to account for the increased likelihood
of a rare event when testing multiple hypotheses). We cross-
referenced Kendall’s W results with Friedman’s tests for all
three studies to report the most favorable results by order of
engagement factors.
6. Results
6.1. UES results based on fair agreement
and statistically significant dierences
Friedman’s statistical analysis showing adjusted significant
differences (p<0.05) combined with Kendall’s W tests in the
general range of “fair” agreement (0.199–0.383) reveal 14 results
across four different engagement factors, namely, focused attention,
perceived usability, aesthetics, and novelty (see Table 2). The results
below are presented by order of UES engagement factors and
statement number(s) from Table 1.
6.1.1. Engagement factor for focused attention
There is only one result to report for focused attention, and it
relates to the Rhythm survey results from pairwise comparisons for
statement 23, which showed that listeners found they were more
absorbed in their listening tasks when presented with the simplified
condition of a plain sound as opposed to the complexities of pitched
timbre or pitched timbre with noise.
6.1.2. Engagement factor for perceived usability
There is only one result to report for perceived usability, and it
relates to the Melody survey results from pairwise comparisons for
statement 15, which showed that listeners found they were more in
control of the sound experience when presented with pitches with
timbre and rhythms as opposed to plain sounds (pure tones).
6.1.3. Engagement factor for aesthetics
There are 10 results to report for aesthetics:
The Melody survey results from pairwise comparisons for
statement 7 showed that listeners found the sounds were more
pleasing to their auditory senses when presented with (a) pitches
with timbre and rhythm as opposed to plain sounds (pure
tones); and (b) pitches with timbre as opposed to plain sounds
(pure tones).
The Melody survey results from pairwise comparisons for
statement 9 showed that listeners found they liked the beats and
rhythms of the sounds most when presented with pitches with
timbre and rhythms as opposed to plain sounds (pure tones).
The Melody survey results from pairwise comparisons
for statement 19 showed that listeners found the sounds
aesthetically appealing to their auditory senses when presented
with a) pitches with timbre and rhythms as opposed to plain
sounds (pure tones); and b) pitches with timbre as opposed to
plain sounds (pure tones).
The Melody survey results from pairwise comparisons for
statement 21 showed that listeners found the sound layout to
be most pleasing when presented with pitches with timbre and
rhythms as opposed to plain sounds (pure tones).
The Chords survey results from pairwise comparisons for
statement 21 showed two pairs of marginally adjusted significant
differences of 0.052 and 0.052. Listeners found the layout of the
Frontiers in Big Data 07 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
sounds to be auditory pleasing when presented with (a) pitches
with timbre as opposed to plain sounds (pure tones) and (b)
pitches with timbre and rhythms as opposed to plain sounds
(pure tones).
The Melody survey results from pairwise comparisons for
statement 24 showed that (a) listeners found the sounds
more attractive when presented with pitches with timbre
as opposed to plain sounds (pure tones); and (b) listeners
found the sounds more attractive when presented with
pitches with timbre and rhythm as opposed to plain sounds
(pure tones).
6.1.4. Engagement factor for novelty
There are only two results to report for novelty, and they
relate to the Melody survey results from pairwise comparisons
for statements 4 and 10. Statement 4 showed that listeners
found they were more interested in their listening task when
presented with pitches with timbre and rhythms as opposed to
plain sounds (pure tones). Statement 10 showed that listeners
found they would continue listening to the sounds out of curiosity
when presented with pitches with timbre as opposed to plain
sounds (pure tones).
7. Discussion
7.1. Factors and elevated engagement
levels
The broadest engagement impact, with musical characteristics,
appears to reside in the aesthetics factor with the conditions for
pitches with timbre and rhythm in the context of melody. Among
the statistically significant differences, the results show the median
was consistently 4.0 (the average range for engagement from the
same data for aesthetics was 3.54 to 4.04 on a scale of 1 to 5). The
results represent elevated engagement levels in relation to data-to-
music, with musical results based on simple, pure tones. Aesthetics
is clearly a primary factor in engagement in musical contexts.
The two results to report from the novelty factor for
engagement had median values of 4.0 and 3.5, or an average
engagement result of 4.33 for pitches with timbre and rhythm from
statement 4 and an average result of 3.17 for pitches with timbre
from statement 10 (in this Discussion, please refer to Table 1 for the
wording of all statements). While 4.33 is one of the highest average
engagement levels to report among 14 results with statistically
significant differences, the average of 3.17 is one of the lowest
engagement results, barely over the mean threshold of 3.
The remaining results come from two factors, perceived
usability and focused attention. Statement 15 for perceived usability
in the melody study had a median engagement result of 4.0 for
pitched timbre and rhythm. Statement 23 for focused attention in
the Rhythm study also had a median engagement result of 4.0 for
plain sound (noise). This result is the only one reported from the
Rhythm study, and it is a unique case where a plain noise sound
was more effective in boosting engagement than a pitched timbre.
We attribute this result to the challenges users had with hearing
wind data among more complex sounds. A simple sound prevailed
over the rich musical environment.
When looking more closely at median and average results,
what was most interesting was how timbre seemed to make a
significant contribution to engagement, while rhythm only seemed
to mildly boost the results. As an example, in Statement 7 of the
Melody survey, the average engagement level was 3.71 for pitches
with timbre, whereas the average result was 3.79 with rhythm
added. We see a similar pattern for statement 19 from the Melody
survey, where the average for pitches with timbre was 3.67, and
with rhythm added, the average was 3.79. The significance of
timbre’s contribution to engagement can also be observed from the
comparison of results between plain tones and tones with timbre.
As an example, in statement 7 of the Melody study, the plain tone
average result was 2.71, while the pitch with timbre average result
was 3.71. A similar comparison can be made between statement
19, where the plain tone average result was 2.63, and the pitch
with timbre result was 3.67. In these cases, timbre, as a musical
characteristic, added a full-point enhancement in relation to plain
tones. The perceived sound colors provided by timbre (McAdams,
2013) provide enhanced experiences for listeners in the surveys.
In general, the Chords survey had many similar results to the
Melody study, with elevated results for pitches with timbre and
rhythm; however, after running a stringent statistical analysis for
significant differences, we can only report two median results of 4.0
for survey statement 21, with average engagement levels of 3.83 for
pitches with timbre and rhythm and 3.75 for pitches with timbre
(both related to the aesthetics factor). The results from the Chords
survey are based on marginally adjusted significant differences of
0.052 in relation to plain sounds. The Chords study used the same
data as the Melody study, so we could anticipate how the results
might be similar with more parity among the number of results
with statistically significant differences. This did not happen, so we
note that the chordal context for the sonifications contributed to
results that were too uniform. We would need more research to
see why significant differences were harder to achieve than in the
melody study.
7.2. Global view and further studies
From a global view, one might have expected a full complement
of musical characteristics (pitch with timbre and rhythm) to
consistently contribute to elevated engagement levels. Instead, the
results seem mixed. Timbre seems to contribute more than we
expected because the separation in the study’s results between pure
tones and tones with timbre was noticeable, and the integrity of
the assessment was affirmed by pairwise comparisons based on
Friedman’s analysis (significant differences were observed by the
separation between plain sounds and those with timbre).
The Rhythm study had remarkably fewer engagement factors
represented, with only three statements from focused attention and
perceived usability showing results with significant differences and
only one statement, number 23, meeting our stringent criteria.
There was something about the complexities of rhythm and data
representation in this study that may have brought the median and
average responses to their lowest levels across the three studies. The
Frontiers in Big Data 08 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
TABLE 2 Results from the data-to-music user engagement scale (UES) surveys based on statistically significant dierences from pairwise comparisons from the Friedman hypothesis test (P<0.05).
Factor of
engagement
Survey Condition Statement Median
value 1–5
Average
value 1–5
Adjusted significant
dierence
Pvalue (P
<0.05)
Kendall’s W
Focused attention Rhythm Plain sounds 23 4.0 3.75 0.015 0.002 0.260
Perceived usability Melody Pitches with timbre and
rhythm
15 4.0 3.75 0.012 0.001 0.274
Aesthetics Melody Pitches with timbre and
rhythm
7 4.0 3.79 0.006 0.000 0.335
Aesthetics Melody Pitches with timbre 7 4.0 3.71 0.035 0.000 0.335
Aesthetics Melody Pitches with timbre and
rhythm
9 4.0 3.83 0.042 0.006 0.215
Aesthetics Melody Pitches with timbre and
rhythm
19 4.0 3.79 0.006 0.000 0.366
Aesthetics Melody Pitches with timbre 19 4.0 3.67 0.018 0.000 0.366
Aesthetics Melody Pitches with timbre and
rhythm
21 4.0 4.04 0.002 0.000 0.383
Aesthetics Chords Pitches with timbre and
rhythm
21 4.0 3.83 0.052 0.002 0.270
Aesthetics Chords Pitches with timbre 21 4.0 3.75 0.052 0.002 0.270
Aesthetics Melody Pitches with timbre and
rhythm
24 4.0 3.75 0.015 0.000 0.357
Aesthetics Melody Pitches with timbre 24 4.0 3.54 0.007 0.000 0.357
Novelty Melody Pitches wit timbre and rhythm 4 4.0 4.33 0.042 0.003 0.249
Novelty Melody Pitches with timbre 10 3.5 3.17 0.018 0.002 0.258
Frontiers in Big Data 09 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
result from statement 23 in the Rhythm study is intriguing in how
it may say something about focused attention. In this unique case, a
singular noise-based sound with a simple array of varying rhythms,
determined by the data, showed the potential to capture our
listeners’ attention equal to or better than more complex musical
events. One of the challenges in the Rhythm study also relates to the
complexity of drum set sounds (a combination of noise, timbre, and
rhythm) used as one of the conditions—requiring a more holistic
listening experience than many participants were able to grasp.
This musically complex condition received some of the lowest
engagement responses across the three studies, mostly for focused
attention, perceived usability, endurability, and felt involvement. We
suspect the participants may have been confused by the complexity
of details in a saturated sound environment, making it challenging
to interpret the data globally from the speed of rhythmic activity in
relation to the speed of wind. One might have anticipated higher
levels of engagement from a drum set, as one might perceive
from the active listening experience in the introduction to John
Coltrane’s A Love Supreme: Pursuance Part III, but that was not
the case. It should be noted that the context of rhythm in more
structured Western musical settings is often in relation to repeated
beat patterns and meter. In the rhythmic study, there was no such
context for the rhythmic events, as the sonifications placed focus
only on data-derived rhythms outside of metric constraints. The
Rhythm survey results may show that users expect more familiar
musical sounds based on traditional roles of rhythm that would rely
on repeated rhythmic patterns and hierarchical beat structures in a
metric context (Bouwer et al., 2018). Evaluating sonifications in the
context of rhythm appears to encompass some secondary contexts
with expectations of beat pattern repetition and meter (Desain and
Honing, 2003). This would require further study to validate.
The surveys show that musical characteristics in our data
sonifications with D2M software contribute to engagement, but
further studies are required to determine the extent to which
each musical characteristic enhances listening and interpreting
experiences. The most stringent analyses of the data from our
three studies indicate that tones with timbre and rhythm show
promise in elevating engagement, especially in the factors of
aesthetics,novelty, and perceived usability. Plain sounds alone,
as mostly a uniform percussion noise (from claves) with some
rhythm, elevated engagement within the factor of focused attention
(statement 23 with reference to being “absorbed” in the listening
task). This preliminary finding suggests that data sonification
mapping should consider the role of timbre and rhythm to enhance
user experiences.
8. Conclusion
The aural experience of auditory display (in terms of data
with sounds) can exist with or without musical characteristics. A
new path for the inclusion of musical characteristics is opening
up based on the influence of engagement in relation to the user
experience. The design of the D2M conversion software shows
promise in how musical sounds can capture the meaning of
datasets and enhance the aesthetic experience, which represents a
significant goal for usability and added focused time for analysis.
There are also benefits to experiencing data analysis from a unique
and alternative perspective (aural). Research in data-to-music has
generally been tentative because of (1) the challenges of design and
decision-making for musical experiences from data, (2) biases of
musical tastes among users, and (3) the risks of cluttering the data
message with decorative sounds. However, the potential rewards
of engagement can be meaningful. Our user engagement study
provides a preliminary understanding of the relevance timbre,
pitch, and rhythm can have to the experience of data via musically
informed sonifications. The characteristics of timbre and pitch, in
particular, appear to have a significant impact on the aesthetics of
auralization and auditory display. The studies also show that timbre
and pitch, in the context of melody and chords, are influential
factors in the engagement factors of perceived usability, aesthetics,
and novelty, and that plain sounds can be influential for focused
attention in the context of rhythm. Musical attributes remain
inconclusive for the factors endurability and felt involvement.
This preliminary study was simplified to isolate musical traits
and allow for improved observation. For real life use, we believe
that a larger set of sonifications should be tested. Such sonifications
should communicate information clearly and this will likely have
improved results for the engagement factors endurability and felt
involved. The observations presented in this paper set a foundation
to investigate even more elaborate musical representations of data.
Data availability statement
The datasets and audio files presented in this study can be found
in online repositories. The names of the repository/repositories and
accession number(s) can be found below: https://data.mendeley.
com/datasets/srp4rhjkmy/2.
Ethics statement
The studies involving human participants were reviewed and
approved by Institutional Review Board: Exemption was provided
by Eastern Washington University (human subjects protocol HS-
5429) on December 1, 2017. The patients/participants provided
their written informed consent to participate in this study.
Author contributions
Conceptualization: JM, JHa, and RR. Methodology: JM, JHa,
and PH. Software: JHa, KT, JM, and JHe. Validation: JM, KT, JHe,
TK, and MZ. Formal analysis: JM, JC, and TK. Investigation: JM,
KT, TK, JHe, and PH. Resources: MT and JL. Data curation: JHe
and KT. Writing original draft: JM. Writing review and editing:
JM, JC, and RR. Visualization and auralization: PH and JM.
Supervision: RR. Project administration: MT, JL, and RR. Funding
acquisition: RR and JM. All authors have read and agreed to the
published version of this manuscript.
Funding
This study was funded by Tekes, the Finnish agency for
Innovation (decision 40296/14).
Frontiers in Big Data 10 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
Acknowledgments
Special thanks to the Tekes committee and company
representatives in the Tekes sponsored project; EWU graduate
student Tim Gales; professors Nancy Birch and Mari Tervaniemi;
and the EWU Analytics Senior Capstone class of JC (DSCI 490,
spring 2020).
Conflict of interest
The authors declare that the research was conducted in the
absence of any commercial or financial relationships that could be
construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those
of the authors and do not necessarily represent those of
their affiliated organizations, or those of the publisher,
the editors and the reviewers. Any product that may be
evaluated in this article, or claim that may be made by
its manufacturer, is not guaranteed or endorsed by the
publisher.
Supplementary material
The Supplementary Material for this article can be found
online at: https://www.frontiersin.org/articles/10.3389/fdata.2023.
1206081/full#supplementary-material
Supplementary samples of audio files from May
weather datasets:
AUDIO S1
May clouds data melody as pure tones.
AUDIO S2
May clouds data melody as pitches with timbre.
AUDIO S3
May clouds data melody as pitches with timbre and some rhythm.
AUDIO S4
May clouds data chords as pure tones.
AUDIO S5
May clouds data chords as pitches with timbre.
AUDIO S6
May clouds data chords as pitches with timbre and some rhythm.
AUDIO S7
May wind data as rhythms with uniform noise.
AUDIO S8
May wind data as rhythms with uniform pitch with timbre.
AUDIO S9
May wind data as rhythms with pitched timbre and noise.
References
Aldwell, E., and Schachter, C. (2011). Harmony and Voice Leading. 4th ed., Boston,
MA: Schirmer.
Bearman, N., and Brown, E. (2012). “Who’s sonifying data and how are they doing
it? A comparison of ICAD and other venues since 2009, in Proceedings of the 18th
International Conference of Auditory Display (ICAD) (Atlanta, GA).
Bouwer, F. L., Burgoyne, J. A., Odijk, D., Honing, H., and Grahn, J. A. (2018). What
makes a rhythm complex? The influence of musical training and accent type on beat
perception. PLoS ONE. 13, e0190322. doi: 10.1371/journal.pone.0190322
Brewster, S. (2009). “Non-Speech auditory output, in Human Computer Interaction
Handbook: Fundamentals, eds. A., Sears, J.A. Jacko (Boca Raton, FL: CRC
Press) 223–240.
Brown, L. M., Brewster, S. A., Ramloll, S. A., Burton, R., and Riedel, B. (2003).
“Design guidelines for audio presentation of graphs and tables, in Proceedings of the
9th International Conference of Auditory Display (ICAD) (Boston, MA).
Bywater, R. P., and Middleton, J. (2016). Melody discrimination and protein fold
classification. Heliyon 2, e00175. doi: 10.1016/j.heliyon.2016.e00175
Desain, P., and Honing, H. (2003). The formation of rhythmic categories and metric
priming. Perception. 32, 341–365. doi: 10.1068/p3370
Dillon, A. (2006). “Information interactions: Bridging disciplines in the creation of
new technologies, in the Human-Computer Interaction and Management Information
Systems: Foundations, eds. D.F., Galletta, Z., Ping (Oxford: Routledge) 21–31.
Hermann, T., Hunt, A., and Neuhoff, J. G. (2011). The Sonification Handbook.
Berlin: Logos Verlag.
Hildebrandt, T., Hermann, T., and Rinderle-Ma, S. (2016). Continuous sonification
enhances adequacy of interactions in peripheral process monitoring. Int. J. Hum-
Comput. Stud. 95, 54–65. doi: 10.1016/j.ijhcs.2016.06.002
Kramer, G., Walker, B., and Bonebright, T. (1999). International Community for
Auditory Display Conferences, ICAD Conferences. Available online at: https://icad.org/
conferences/ (accessed August 23, 2021).
Lutz Klauda, S., and Guthrie, J. T. (2015). Comparing relations of motivation,
engagement, and achievement among struggling and advance adolescent readers. Read
Writ. 28, 239–269. doi: 10.1007/s11145-014-9523-2
McAdams, S. (2013). “Musical timbre perception, in the Psychology of
Music, 3rd ed., D., Deutsch (London, UK: Elsevier Academic Press) 35–67.
doi: 10.1016/B978-0-12-381460-9.00002-X
McCarthy, J., and Wright, P. (2004). Technology as Experience. MIT Press:
Cambridge, MA. doi: 10.1145/1015530.1015549
Middleton, J., and Dowd, D. (2008). Web-based algorithmic composition from
extramusical resources. Leonardo 41, 128–135. doi: 10.1162/leon.2008.41.2.128
Middleton, J., Hakulinen, J., Tiitinen, K., Hella, J., Keskinen, T., Huuskonen, P., et al.
(2018). “Sonification with musical characteristics: a path guided by user-engagement,
in Proceedings of the 24th International Conference of Auditory Display (Houghton, MI).
doi: 10.21785/icad2018.006
Neuhoff, J. G. (2011). “Perception, cognition and action in auditory displays, in
The Sonification Handbook, eds. T., Hermann, A., Hunt, J. G., Neuhoff (Berlin: Logos
Verlag) 63–85.
Norman, D. A. (2004). Introduction to this special section on beauty, goodness, and
usability. Hum. Comput. Interact. 19, 311–318. doi: 10.1207/s15327051hci1904_1
O’Brien, H. L., Cairns, P., and Hall, M. (2018). A practical approach to measuring
user engagement with the refined user engagement scale (UES) and new UES
short form. Int. J. of Hum- Comput. Stud. 112, 28–39. doi: 10.1016/j.ijhcs.2018.
01.004
O’Brien, H. L., and Toms, E. G. (2008). What is user engagement? A conceptual
framework for defining user engagement with technology. J. Am. Soc. Inf. Sci. Technol.
59, 938–955. doi: 10.1002/asi.20801
O’Brien, H. L., and Toms, E. G. (2010). The development and evaluation of
a survey to measure user engagement. J. Am. Soc. Inf. Sci. Technol. 61, 50–69.
doi: 10.1002/asi.21229
Pollack, I., and Ficks, L. (1954). Information of elementary multidimensional
auditory displays. J. Acoust. Soc. Am. 26, 155. doi: 10.1121/1.1907300
Tractinsky, N. (2006). “Aesthetics in information technology: Motivation
and future research directions, in the Human-Computer Interaction and
Management Information Systems: Foundations eds. D.F., Galletta, Z., Ping (Oxford:
Routledge) 330–347.
Frontiers in Big Data 11 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
Tractinsky, N., Katz, A. S., and Ikar, D. (2000). What is beautiful is usable. Interact.
Comput. 13, 127–145. doi: 10.1016/S0953-5438(00)00031-X
Tuch, A. N., Roth, S. P., Hornbæk, K., Opwis, K., and Bargas-Avila, J. A. (2012). Is
beautiful usable? Toward understanding the relation between usability, aesthetics, and
affect in HCI. Comput. Hum. Behav. 28, 1596–1607. doi: 10.1016/j.chb.2012.03.024
Vickers, P. (2017). “Sonification and music, music and sonification, in The
Routledge Companion to Sounding Art, eds. M., Cobussen, V., Meelberg, B., Truax
(Oxford: Routledge) 135–144.
Vickers, P., Hogg, B., and Worrall, D. (2017). “Aesthetics of sonification:
Taking the subject- position, in the Body, Sound and Space in Music and
Beyond: Multimodal Explorations, ed. C., Wöllner (Oxford: Routledge) 89–109.
doi: 10.4324/9781315569628-6
Walker, B., and Nees, M. A. (2011). “Theory of sonification, in The Sonification
Handbook, eds. T., Hermann, A., Hunt, J. G., Neuhoff (Berlin: Logos Verlag) 9–31.
Walker, B. N., and Kramer, G. (2005). Mappings and metaphors in auditory
displays: An experiment assessment. ACM Trans. Appl. Percept. 2, 407–412.
doi: 10.1145/1101530.1101534
Webster, J., and Ahuja, J. S. (2006). Enhancing the design of Web navigation
systems: The influence of user disorientation on engagement and performance. MIS
Q. 30, 661–678. doi: 10.2307/25148744
Frontiers in Big Data 12 frontiersin.org
Middleton et al. 10.3389/fdata.2023.1206081
Appendix
Data mapping and audio file standards
The oktas data, on a scale of 0–8, were mapped to a range of
a musical octave and a half with a tempo of one beat per second
(See sample score in Figure A2). A total of 40 data points in a
24-h period were displayed in 40 s. A MIDI range of 43–60 was
used as a destination span, and all results adhered to a C major
diatonic collection (Figure A1). A sample musical score of the May
oktas data inverted for sunshine (seen in Figure 1) is provided
below to show the musical range and level 1 rhythmic activity
(Figure A2). The resulting MIDI files were rendered through
software instruments on Logic Pro X and Ableton Live, and the
resulting audio files were exported and converted to mp3 files for
Web delivery with loudness normalized to negative 25 LUFS.
FIGURE A1
D2M software interface mapping November cloud cover data in oktas to a MIDI range (43–60), instrument (marimba), and musical scale (C major).
FIGURE A2
Melody generated by D2M from inverted clouds data in okta from May. Compare the melodic contour of the musical score with Figure 1 and May’s
clouds melodies below. The melodic contour slopes down and then back up. This was represented in the UES study as pitched pure tones, pitched
timbre (marimba), and pitched timbre (marimba) with some rhythm.
Frontiers in Big Data 13 frontiersin.org
... Either of these approaches are completely valid within their discipline [4]. The motivation is to find a middle ground whereby the auditory display can function as a tool for data analysis and have intentional musical attributes This is advantageous to avoid analytical fatigue [5], and enhance user engagement [6] whilst listening. Additionally, intended musical attributes may be of interest to the creative arts and wider community providing an opportunity promote the research more widely [7]. ...
Conference Paper
DNA sequences contain vast amounts of biological data and computer algorithms play an important role in processing these data for human inspection. Here we describe an updated computer-generated auditory display tool to be used as stand-alone audio or as a complement to a visual display for DNA sequence inspection. The auditory display uses musical notes to represent the data in relation to the process of gene expression or DNA replication. Given the use of musical notes in the auditory display raises the possibility these may be considered algorithmic music. To pursue this notion further the auditory displays were used in a music studio setting outside of the science laboratory. Musicians were challenged to play in-sync with the audio and to embellish the melodic and harmonic content of the auditory display. New music compositions featuring the auditory displays were recorded and performed live in outreach events to promote a wider understanding of the processes of gene expression and DNA replication and how gene sequence information affects human health conditions.
... The full terms of the License are available at http://creativecommons.org/licenses/by-nc/4.0/ interpreting scientific data, such as ways to map protein molecules to music [8]. Multimodality allows users to integrate information at an enhanced rate because of the associative connections multiple stimuli provide the brain. ...
Conference Paper
Full-text available
This research investigated audio-visual analytics of geoscientific data in virtual reality (VR)-enhanced implementation, where users interacted with the dataset with a VR controller and a haptic device. Each interface allowed users to explore rock minerals in unimodal and multimodal virtual environments (VE). In the unimodal version, color variations demonstrated differences in minerals. As users navigated the data using different interfaces, visualization options could be switched between the original geographical topology and its color-coded version, signifying underlying minerals. During the multimodal navigation of the dataset, in addition to the visual feedback, an auditory display was performed by playing a musical tone in different timbres. For example, ten underlying minerals in the sample were explored. Among them, anorthite was represented by nylon guitar, the grand piano was used for albite, and so on. Initial findings showed that users preferred the audio-visual exploration of geoscientific data over the visual-only version. Virtual touch enhanced the user experience while interacting with the data.
Conference Paper
Full-text available
In the realm of music we have multiple examples of successful rock stars, composers and producers who describe themselves as self-taught. This suggests there might be a demand in formal music education for learning technologies that support students' self-propelled discovery. In this theoretical work, we explore the design space of educational music composing games that would allow students to explore and learn music theory concepts at their leisure. We designed four unique tile-based music creation games based on popular contemporary video game genres, and evaluated them from the perspectives of learning and musical expression. This study opens up new avenues in music composing game design, and offers examples of some ways in which games can be harnessed as vehicles to learn music theory and composition.
Conference Paper
Full-text available
Sonification with musical characteristics can engage users, and this dynamic carries value as a mediator between data and human perception, analysis, and interpretation. A user engagement study has been designed to measure engagement levels from conditions within primarily melodic, rhythmic, and chordal contexts. This paper reports findings from the melodic portion of the study, and states the challenges of using musical characteristics in sonifications via the perspective of form and function-a long standing debate in Human-Computer Interaction. These results can guide the design of more complex sonifications of multivariable data suitable for real life use.
Article
Full-text available
Perception of a regular beat in music is inferred from different types of accents. For example, increases in loudness cause intensity accents, and the grouping of time intervals in a rhythm creates temporal accents. Accents are expected to occur on the beat: when accents are “missing” on the beat, the beat is more difficult to find. However, it is unclear whether accents occurring off the beat alter beat perception similarly to missing accents on the beat. Moreover, no one has examined whether intensity accents influence beat perception more or less strongly than temporal accents, nor how musical expertise affects sensitivity to each type of accent. In two experiments, we obtained ratings of difficulty in finding the beat in rhythms with either temporal or intensity accents, and which varied in the number of accents on the beat as well as the number of accents off the beat. In both experiments, the occurrence of accents on the beat facilitated beat detection more in musical experts than in musical novices. In addition, the number of accents on the beat affected beat finding more in rhythms with temporal accents than in rhythms with intensity accents. The effect of accents off the beat was much weaker than the effect of accents on the beat and appeared to depend on musical expertise, as well as on the number of accents on the beat: when many accents on the beat are missing, beat perception is quite difficult, and adding accents off the beat may not reduce beat perception further. Overall, the different types of accents were processed qualitatively differently, depending on musical expertise. Therefore, these findings indicate the importance of designing ecologically valid stimuli when testing beat perception in musical novices, who may need different types of accent information than musical experts to be able to find a beat. Furthermore, our findings stress the importance of carefully designing rhythms for social and clinical applications of beat perception, as not all listeners treat all rhythms alike.
Article
Full-text available
User engagement (UE) and its measurement have been of increasing interest in human-computer interaction (HCI). The User Engagement Scale (UES) is one tool developed to measure UE, and has been used in a variety of digital domains. The original UES consisted of 31-items and purported to measure six dimensions of engagement: aesthetic appeal, focused attention, novelty, perceived usability, felt involvement, and endurability. A recent synthesis of the literature questioned the original six-factors. Further, the ways in which the UES has been implemented in studies suggests there may be a need for a briefer version of the questionnaire and more effective documentation to guide its use and analysis. This research investigated and verified a four-factor structure of the UES and proposed a Short Form (SF). We employed contemporary statistical tools that were unavailable during the UES’ development to re-analyze the original data, consisting of 427 and 779 valid responses across two studies, and examined new data (N=344) gathered as part of a three-year digital library project. In this paper we detail our analyses, present a revised long and short form (SF) version of the UES, and offer guidance for researchers interested in adopting the UES and UES-SF in their own studies.
Article
Full-text available
One of the greatest challenges in theoretical biophysics and bioinformatics is the identification of protein folds from sequence data. This can be regarded as a pattern recognition problem. In this paper we report the use of a melody generation software where the inputs are derived from calculations of evolutionary information, secondary structure, flexibility, hydropathy and solvent accessibility from multiple sequence alignment data. The melodies so generated are derived from the sequence, and by inference, of the fold, in ways that give each fold a sound representation that may facilitate analysis, recognition, or comparison with other sequences.
Article
Full-text available
As many users who are charged with process monitoring need to focus mainly on other work while performing monitoring as a secondary task, monitoring systems that purely rely on visual means are often not well suited for this purpose. Sonification, the presentation of data as (non-speech) sound, has proven in several studies that it can help in guiding the user's attention, especially in scenarios where process monitoring is performed in parallel with a different, main task. However, there are several aspects that have not been investigated in this area so far, for example if a continuous soundscape can guide the user's attention better than one that is based on auditory cues. We have developed a system that allows reproducible research to answer such questions. In this system, the participants’ performance both for the main task (simulated by simple arithmetic problems) and for the secondary task (a simulation of a production process) can be measured in a more fine-grained manner than has been the case for existing research in this field. In a within-subject study (n=18), we compared three monitoring conditions - visual only, visual + auditory alerts and a condition combining the visual mode with continuous sonification of process events based on a forest soundscape. Participants showed significantly higher process monitoring performances in the continuous sonification condition, compared to the other two modes. The performance in the main task was at the same time not significantly affected by the continuous sonification.
Article
Full-text available
Chapter
Full-text available
Despite it being more than twenty years since the launch of an international conference series dedicated to its study, there is still much debate over what sonification really is, and especially as regards its relationship to music. A layman’s definition of sonification might be that it is the use of non-speech audio to communicate data, the aural counterpart to visualization. Many researchers have claimed musicality for their sonifications, generally when using data-to-pitch mappings. In 2006 Bennett Hogg and I (Vickers and Hogg 2006) made a rather provocative assertion that bound music and sonification together (q.v., and further developed in Vickers (2006)), not so much to claim an ontological truth but to foreground a debate that has simmered since the first International Conference on Auditory Display (ICAD) in 1992. Since then there has been an increasing number of musical and sonic art compositions driven by the data of natural phenomena, some of which are claimed by their authors to be sonifications. This chapter looks at some of the issues surrounding the relationship between sonification and music and at developments that have the potential to draw sonification and the sonic arts into closer union.
Chapter
Full-text available
Musical timbre encompasses a complex set of auditory attributes and raises a plethora of musical and psychological issues. To discover the underlying perceptual structure of timbre, psychophysical approaches have used multidimensional scaling of (dis)similarity ratings to model relations among sounds differing in timbre as points within a "timbre space." The dimensions of this space are considered to represent salient perceptual aspects of the sounds based on definable acoustical parameters or perceptual categories. Timbre provides numerous perceptual cues for the categorization and identification of sound sources. As the primary perceptual result of orchestration practice, timbre can serve a form-bearing role in the creation of musical structures and musical discourse. Some auditory scene analysis processes create timbre through the blending of concurrent acoustic information, including the blending of several instruments. Others use timbre to organize events into coherent groups of events, the grouping of which can be predicted by the relative distances between sounds in timbre space. Timbre space representations also provide a model for defining timbral intervals and operations on those intervals such as transposition. Timbre can play a significant role in inflecting musical tension, and there is evidence that transition probabilities between timbres in musical sequences can be learned implicitly, opening up the possibility of creating grammatical operations in music based on timbre.