Auditory display is the use of sound to present information to a listener. Sonification is a particular type of auditory display technique in which data is mapped to non-speech sound to communicate information about its source to a listener. Sonification generally aims to leverage the temporal and frequency resolution of the human ear and is a useful technique for representing data that cannot be represented by visual means alone. Taking this perspective as our point of departure, we believe that sonification may benefit from being informed by aesthetic explorations and academic developments within the wider fields of music technology, electronic music and sonic arts. In this paper, we will seek to explore areas of common ground between sonification and electronic music/sonic arts using unifying frameworks derived from musical aesthetics and embodied cognitive science (Kendall, 2014; Lakoff & Johnson, 1999).
Sonification techniques have been applied across a wide range of contexts including the presentation of information to the visually impaired (Yoshida et al., 2011), process monitoring for business and industry (Vickers, 2011), medical applications (Ballora et al., 2004), human computer interfaces (Brewster, 1994), to supplement or replace visual displays (Fitch & Kramer, 1994), exploratory data analysis (Hermann & Ritter, 1999) and, most importantly for the current milieu, to reveal the invisible data flows of smart cities and the internet of things (Rimland et al., 2013; Lockton et al., 2014). The use of sonification as a broad and inclusive aesthetic practice and cultural medium for sharing, using and enjoying information is discussed by Barrass (2012). As networked smart societies grow in size and becomes increasingly complex the ubiquitous invisible data flows upon which these societies run are becoming hard to monitor and understand by visual means alone. Sonification might provide a means by which these invisible data flows can be monitored and understood.
In order to achieve this type of usage, sonification solutions need to be applicable to and intelligible to an audience of general listeners. This requires a universal shared context by which sonifications can be interpreted. Embodied cognition researchers argue that the shared physical features of the human body, and the capacities and actions which our bodies afford us, define and specify mid-level structures of human cognitive processing, providing shared contexts by which people can interpret meaning in and assign meaning to their worlds (Lakoff and Johnson 1980; 1999; Varela et al., 1991). At present, embodied perspectives on cognition are infrequently explored in auditory display research, which tends to focus on either higher level processing in terms of language and semiotics (Vickers, 2012) or lower level processing in terms of psychoacoustics and Auditory Scene Analysis (Carlile, 2011).