Conference PaperPDF Available

Recognition of Car Warnings: An Analysis of Various Alert Types

Authors:

Abstract and Figures

Warnings are integral to ensuring the safe operation of a vehicle. The use of auditory alerts and warnings has the potential to alleviate drivers' workload, increase drivers' situation awareness, and facilitate efficient and safe driving. The present study assessed the ability of individuals to react to auditory warnings. The warnings took the form of text-to-speech (TTS), spearcons at two levels of linear compression, and auditory icons. To assess usability, the NASA Task Load Index and an annoyance question were used. Participants pressed a space bar when they recognized a warning, and then identified auditory warnings by selecting a picture corresponding to the meaning of the warning. The results showed that participants responded to the fastest spearcon warnings more quickly compared to TTS and auditory icons. Responses to auditory icons were slowest compared to all other auditory types. Importantly, responses to spearcon warnings were no less accurate in comparison to TTS warnings.
Content may be subject to copyright.
Recognition of Car Warnings: An
Analysis of Various Alert Types
Abstract
Warnings are integral to ensuring the safe operation of
a vehicle. The use of auditory alerts and warnings has
the potential to alleviate drivers’ workload, increase
drivers’ situation awareness, and facilitate efficient and
safe driving. The present study assessed the ability of
individuals to react to auditory warnings. The warnings
took the form of text-to-speech (TTS), spearcons at
two levels of linear compression, and auditory icons. To
assess usability, the NASA Task Load Index and an
annoyance question were used. Participants pressed a
space bar when they recognized a warning, and then
identified auditory warnings by selecting a picture
corresponding to the meaning of the warning. The
results showed that participants responded to the
fastest spearcon warnings more quickly compared to
TTS and auditory icons. Responses to auditory icons
were slowest compared to all other auditory types.
Importantly, responses to spearcon warnings were no
less accurate in comparison to TTS warnings.
Author Keywords
Warning systems; auditory icons; text-to-speech; car
alerts; spearcons; sonification.
ACM Classification Keywords
H.1.2 [Information Systems]: User/Machine Systems
Permission to make digital or hard copies of part or all of this work for
personal or classroom use is granted without fee provided that copies are
not made or distributed for profit or commercial advantage and that
copies bear this notice and the full citation on the first page. Copyrights
for third-party components of this work must be honored. For all other
uses, contact the Owner/Author.
Copyright is held by the owner/author(s).
CHI'17 Extended Abstracts, May 06-11, 2017, Denver, CO, USA
ACM 978-1-4503-4656-6/17/05.
http://dx.doi.org/10.1145/3027063.3053149
Edin Sabic
Scott Mishler
Jing Chen
Bin Hu
New Mexico State University
Las Cruces, NM 88003, USA
sabic@nmsu.edu
smishler@nmsu.edu
jingchen@nmsu.edu
bhu2@alumni.nd.edu
Late-Breaking Work
CHI 2017, May 6–11, 2017, Denver, CO, USA
2010
Introduction
Car warnings can represent various issues and levels of
urgency but all aim to communicate with the driver in
an unobtrusive and informative manner. Recognition of
the meaning behind a warning is important, as
ambiguous warnings and symbols can be neglected or
even completely ignored. Further, if the warning is hard
to understand it may be misidentified, leading to an
undesired response. As a result, the semantic mapping
of a car warning to its referent is crucial in facilitating a
response behavior [9]. An ambiguous and unclear
relationship between the auditory warning and the
phenomena it aims to represent can lead to less
accurate and slow responses. If car warnings are either
confusing or inefficient, they may impair drivers’
performance in response to potential danger.
Auditory alerts or feedback can take shape in many
forms, including text-to-speech (TTS), auditory icons,
earcons, tones, and spearcons, to name just a few.
Earcons were introduced as brief melodies that do not
semantically map onto the construct which they
represent [3]. In contrast, auditory icons are sounds
that can accurately depict the construct intended
through the use of semantically-congruent
relationships. Auditory icons utilize pre-existing
constructs, such as the connection between the shutter
sound and a camera. While auditory icons have the
benefit of being language-free, and therefore, are more
generalizable cross-culturally, some constructs or
messages are hard to represent without using
language. Lastly, spearcons are sped-up versions of
TTS that have increased tempo but are no different in
pitch compared to TTS [15]. Spearcons are sped up to
the point where they may not be recognizable as
speech, although it is not strictly necessary. Studies
assessing the learnability of sound cues have shown
that auditory icons and spearcons both require less
time to learn than earcons [1,4,11], and that spearcons
are comparable to TTS in terms of learnability [1].
Further, spearcons and TTS outperform auditory icons
and earcons in terms of learnability for longer lists [1].
Previous research investigating the efficacy of various
auditory car warnings has supported the integral role of
semantic congruency in auditory warnings [9], while
others have found that the use of auditory warnings in
a vehicle context was perceived as less effort-intensive
than visual warnings [10]. Research analyzing gaze
movements and interaction with an in-vehicle GPS
interface also supported auditory warnings in vehicles
by demonstrating that sonification of an in-vehicle
device increased time spent looking at the primary
(driving) task as compared with visual warnings [14].
Similarly, other research has demonstrated that
supplementing a vehicle’s infotainment system with
auditory cues compared with no sound at all increased
both driving and menu navigation performance [7].
Another study comparing the implementation of
spearcons, earcons, or no sound at all to an interactive
vehicle interface showed that the use of spearcons
decreased the time spent looking at the visual display
and increased subjective driving performance [8].
Previous studies [12,16] demonstrated that spearcons
could be responded to more quickly than TTS, without
sacrificing response accuracy. Other research has also
demonstrated that manipulating properties of auditory
warnings, such as speed, can convey urgency [2,6,13].
Urgency has been measured by rating and ranking
sounds based on perceived urgency [2,6], or by having
participants manipulate the properties of sounds to
Late-Breaking Work
CHI 2017, May 6–11, 2017, Denver, CO, USA
2011
create their own urgent warnings [13]. With these
findings in mind, the implementation of spearcons into
a car warning system may allow drivers and passengers
to respond more quickly to an aversive event. This
facilitated response is important because in driving
situations, time can be critical.
Experiment
So far, no previous studies have evaluated the efficacy
of spearcons compared to other auditory alert types
with potential usage for car warnings. In addition,
further research in the usability aspect of auditory
warnings is necessary as limitations, such as the
potential for annoyance [9] and interference, are
apparent. Thus, the purpose of the present study was
to determine the efficacy of various forms of auditory
alerts (TTS, auditory icons, and spearcons), while also
measuring workload and annoyance associated with
these forms of auditory alerts. Further, the present
study assessed the learnability of the auditory alert
types across the four experimental blocks in terms of
both RT and accuracy. These auditory types were
included due to their relevance in the field of auditory
displays and alerts. The independent variable was the
auditory alert type. Dependent variables consisted of
reaction time (RT), accuracy, the NASA Task Load
Index (NASA-TLX) [5] and an annoyance measure.
Methods
A between-subjects design was used consisting of one
between-subjects factor, auditory alert type, with four
levels: (a) spearcons at 40% of original audio length
(40% spearcons), (b) spearcons at 60% of original
audio length (60% spearcons), (c) TTS, and (d)
auditory icons. A between-subjects design was chosen
due to the semantic similarity across alert types. There
were 22 participants in each of the groups.
Participants
Eighty-eight students (52 female; mean age = 19.38;
SD = 3.12) from New Mexico State University
participated in the experiment for course credit. All of
them reported normal or corrected-to-normal hearing.
Apparatus
Participants completed the experiment on a computer,
a Dell OptiPlex 7020, running E-Prime software.
Participants wore a pair of headphones, the Audio-
Technica ATH-M30x Professional Studio Monitor
Headphones, to hear all stimuli during the experiment.
Computer volume was kept constant at 30% of full
volume. Practice served as a check to ensure
participants could hear the stimuli at this volume, and
all participants reported as having normal hearing.
Stimuli
TTS was created using NaturalReader 14.0 software,
which reads entered text using a synthesized voice.
Sound was captured using a playback option on
Audacity, a free software used for audio editing. The
WASAPI playback option allows for the recording of
computer playback without the usage of a microphone.
Spearcons were produced by taking each corresponding
TTS alert or warning and increasing the tempo within
Audacity. Tempo was increased through linear
compression to create 40% and 60% spearcons. That
is, if the original sound duration was 1 second, the 40%
spearcon version was 400 ms. All auditory icons were
gathered from the British Broadcasting Corporation
(BBC) Sound Effects Library, from CDs: 1, 5, 12, 13,
and 19. These auditory icons were used in previous
Late-Breaking Work
CHI 2017, May 6–11, 2017, Denver, CO, USA
2012
research [9] to assess signal-referent relationships. All
sound effects were edited through Audacity to ensure
that sounds did not include trailing noise, and were
only as long as needed to convey meaning. All warning
phrases (see Table 1) and illustrations (see Figure 1)
were borrowed from a previous study that compared
TTS to other auditory feedback types [9].
Subjective measures
Participants responded to seven questions at the end of
the experiment. NASA-TLX was used to assess mental
workload [5]. This survey measures workload on six
scales, which include: mental demand, physical
demand, temporal demand, performance, effort, and
frustration. Further, an annoyance question consisting
of the prompt, “How annoying did you find these
warnings?” was used to assess perceived annoyance.
All questions utilized a 21-point scale.
Procedure
Participants first signed the consent form and provided
demographics information. At the beginning of the
experiment, participants were familiarized with the
auditory type that would be used in the study. They
were then presented with a slideshow, which paired the
auditory warning with a picture that would be used to
assess accuracy (see Figure 1). The practice session
then began, which could last between 9 and 45 trials
depending on the participant’s performance. The
practice session was completed when participants
performed at above 85% accuracy. The practice trials
mirrored the experimental blocks completely. Each trial
began with the presentation of a sound, and
participants were required to press the space bar as
soon as they recognized the warning. RT was logged
from stimulus onset until the space bar was pressed.
After this bar press, they were taken to a screen with
nine pictures (see Figure 1) corresponding to all
auditory warnings, and asked to select which warning
that they just heard. After practice was completed,
participants performed 432 experimental trials split up
into four blocks. At the end of each block, participants
answered the NASA-TLX and annoyance questions
based on their experience with the sound warnings.
Participants were given the choice to take a break
lasting up to five minutes between blocks.
Figure 1: Illustrations for car warnings as seen by
participants. Each illustration corresponded to a specific
warning.
Results
A one-way analysis of variance (ANOVA) was conducted
with auditory alert type as a between-subjects factor on
RT. The main effect of alert type on RT was significant,
F(3, 84) = 42.11, p < .001, ηp2 = .60. Mean RTs tended
to decrease across TTS, 60% spearcons, and 40%
spearcons (see Figure 2). Pairwise comparisons (Sidak)
showed that auditory icons were responded to more
slowly compared to all other auditory types, ps < .001.
40% spearcons were responded to more quickly than
TTS
Auditory
Icons
Headway
closing fast
Collision
sound
Drifting off
road
Rumble
strips
Speed limit
exceeded
Car
speeding
past
Car in blind
spot
Honk
Tire
pressure
low
Air
release
Door is
open
Door
shutting
Hand break
on
Creaking
sound
Gas is low
Water
pouring
Oil is low
Steam and
water
Late-Breaking Work
CHI 2017, May 6–11, 2017, Denver, CO, USA
2013
TTS, p = .002. Responses to 60% spearcons were not
significantly different from 40% spearcons, p = .357, or
from TTS, p = .319.
To determine if there were any practice effects for RT,
a mixed-design ANOVA was conducted with block
number being a within-subjects factor and auditory
type a between-subjects factor. There was a significant
effect of block number, F(3, 252) = 98.40, p < .001,
ηp2 = .54, and an interaction between block number
and condition, F(9, 252) = 2.21, p = .022, ηp2 = .07.
Pairwise comparisons (Sidak) showed that there were
significantly faster RTs moving from block 1 to 2, and
from block 2 to 3, ps < .001. However, RTs were not
significantly different in blocks 3 and 4, p = .369
Figure 2: Reaction time (RT) across auditory type.
A one-way ANOVA was conducted with alert type as the
between-subjects factor on accuracy. The overall
analysis of accuracy was significant, F(3, 84) = 3.81, p
= .013, ηp2 = .12. Pairwise comparisons (Sidak)
showed that the only significant difference in accuracy
was between TTS and Auditory icons, p = .015, with
auditory icons exhibiting a lower accuracy (M =
95.57%, SD = .03) than TTS (M = 97.81%, SD = .01).
Across all auditory alert types, there were no significant
learning effects for accuracy.
Subjective measures
To determine whether there was a difference of
temporal demand across auditory types, a one-way
ANOVA was conducted with auditory alert type as the
between-subjects factor. Only the temporal demand
data was used to determine whether the increased
tempo of spearcons would create a sense of urgency.
Responses to other questions from the NASA-TLX were
not analyzed. The main effect of alert type was
significant, F(3, 84) = 3.71, p = .015, ηp2 = .12.
However, pairwise comparisons (Sidak) showed that
40% spearcons were not perceived as being
significantly more temporally demanding than 60%
spearcons, TTS, or auditory icons, ps > .05. To
determine whether there was a difference of perceived
annoyance across auditory alert types, a one-way
ANOVA was conducted with auditory alert type as the
between-subjects factor. The main effect of alert type
was not significant, F(3, 84) = 2.54, p = .062, ηp2 =
.08. Planned comparisons (Sidak) showed that 40%
spearcons were not perceived as more annoying
compared to the other alert types, ps > .05.
Discussion
The present study compared the efficacy and usability
of four auditory alert types in a car warning recognition
task. After only a brief training session, accuracy was
not significantly lower for spearcons compared to TTS.
Auditory icons had the lowest accuracy among all alert
Late-Breaking Work
CHI 2017, May 6–11, 2017, Denver, CO, USA
2014
types. Considering that the auditory icons contained
lower semantic congruency than TTS and spearcons,
this result is consistent with previous research
investigating signal-to-referent relationships [9].
Although there were few differences in accuracy when
comparing the auditory alert types, there were
significant differences in terms of RTs. Analyses showed
that auditory icons were responded to more slowly
when compared to 40% spearcons, 60% spearcons,
and TTS. This may be a result of longer durations of the
warning, or a weaker signal-to-referent relationship for
auditory icons [9]. Importantly, 40% spearcon
warnings were responded to more quickly than TTS
warnings. While responses to 60% spearcon warnings
were not significantly different from TTS, RTs decreased
numerically going from TTS, to 60% spearcons, and to
40% spearcons (see Figure 2), suggesting that further
research should be conducted to analyze whether this
level of linear compression (60% of original audio
duration) can be consistently responded to more
quickly when compared to TTS. Analyses assessing RTs
across all alert types provided evidence for learning
effects, indicating that training may be required to
reach optimal RTs for all alert types.
The finding that 40% spearcons were responded to
most quickly without any significant accuracy loss has
important implications for auditory car warnings. Car
warnings often need to communicate highly critical and
time sensitive information. Facilitating a faster response
by providing a more efficient warning may lead to safer
driving and fewer accidents on the road. The finding
that these spearcons were also not perceived as more
annoying than the other alert types is important in
assessing usability. Further research is needed to
establish which types of auditory alerts should be
designated to which warnings, especially when
considering severity. Considering that annoyance and
fatigue can set in with repetition of one type of auditory
alert, designating low-risk and highly repetitive events
with auditory icons may increase satisfaction. However,
for high-risk situations that will demand quick and
accurate responses, the present findings support that
spearcons may be an appropriate design choice.
Further research is also needed to establish whether
the present findings can be replicated outside of a quiet
environment while using a dual-task paradigm to
assess the true validity of these auditory alert types.
Conclusion
The 40% spearcon warnings were responded to more
quickly when compared to both TTS and auditory icons,
and were not perceived as being more annoying
compared to 60% spearcons, TTS, or auditory icons.
Further, there was no significant difference in accuracy
when comparing spearcons to TTS. The present
research provides evidence that even brief spearcon
alerts may be useful for providing information in time-
critical situations which could potentially increase safety
and driving performance. However, these findings need
to be replicated in a driving task to ensure that
spearcons can provide viable communication during the
operation of a vehicle.
Acknowledgements
The authors thank Jaymison Miller, Steven Archuleta,
Gabrielle Campbell, and Graham Strom for their help in
data collection. Funding for this research was provided
by the National Science Foundation under Grant
1314688.
Late-Breaking Work
CHI 2017, May 6–11, 2017, Denver, CO, USA
2015
References
1. Tilman Dingler, Jeffrey Lindsay, and Bruce N.
Walker. 2008. Learnability of Sound Cues for
Environmental Features: Auditory Icons, Earcons,
Spearcons, and Speech. 14th International
Conference on Auditory Display, 1-6.
2. J. Edworthy and N. Stanton. 1995. A user-centered
approach to the design and evaluation of auditory
warning signals: 1. Methodology. Ergonomics, 38,
11: 2262-2280.
3. William W. Gaver. 1986. Auditory icons: Using
sound in computer interfaces. Human-computer
interaction 2, 2: 167-177.
4. Stavros Garzonis, Simon Jones, Tim Jay, and
Eamonn O'Neill. 2009. Auditory Icon and Earcon
Mobile Service Notifications: Intuitiveness,
Learnability, Memorability and Preference. CHI
2009-Multimodal Mobile Interaction: 1513-1522.
5. Sandra G. Hart and Lowell E. Staveland. 1988.
Development of NASA-TLX (Task Load Index):
Results of Empirical and Theoretical Research.
Advances in Psychology 52, C: 139-183.
6. E. J. Hellier and J. Edworthy. 1989. Quantifying the
perceived urgency of auditory warnings. Canadian
Acoustics 17, 3-11.
7. Myounghoon Jeon, Thomas M. Gable, Benjamin K.
Davison, Michael A. Nees, Jeff Wilson, and Bruce
Walker. 2015. Menu Navigation With In-Vehicle
Technologies: Auditory Menu Cues Improve Dual
Task Performance, Preference, and Workload.
International Journal of Human-Computer
Interaction 31, 1: 1-16.
8. Pontus Larsson and Mathias Niemand. 2015. Using
Sound to Reduce Visual Distraction from In-Vehicle
HumanMachine Interfaces. Traffic Injury
Prevention 16, sup1: S25-S30.
9. Denis McKeown and Sarah Isherwood. 2007.
Mapping candidate within-vehicle auditory displays
to their referents. Human Factors 49, 3: 417-428.
10. Michael A. Nees, B. Helbein, and A. Porter. 2016.
Speech auditory alerts promote memory for alerted
events in a video-simulated self-driving car ride.
Human Factors 58, 3: 416-426.
11. Dianne K. Palladino and Bruce N. Walker. 2007.
Learning rates for auditory menus enhanced with
spearcons versus earcons. 13th International
Conference on Auditory Display, 274-279.
12. Edin Sabic and Jing Chen. 2016. Threshold of
Spearcon Recognition for Auditory Menus.
Proceedings of the Human Factors and Ergonomics
Society Annual Meeting, 60, 1539-1543.
13. Jeremiah Singer, Neil Lerner, Carryl Baldwin, and
Eric Traube. 2015. Auditory Alerts in Vehicles:
Effects of Alert Characteristics and Ambient Noise
Conditions on Perceived Meaning and Detectability.
24th International Technical Conference on the
Enhanced Safety of Vehicles (ESV)(No. 15-0455).
14. Julien Tardieu, Nicolas Misdariis, Sabine Langlois,
Pascal Gaillard, and Céline Lemercier. 2015.
Sonification of in-vehicle interface reduces gaze
movements under dual-task condition. Applied
Ergonomics, 50, 41-49.
15. Bruce N. Walker, Amanda Nance, and Jeff Lindsay.
2006. Spearcons: Speech-based earcons improve
navigation performance in auditory menus. 12th
International Conference on Auditory Display, 63-
68.
16. Bruce N. Walker, Jeff Lindsay, Amanda Nance, Yoko
Nakano, Dianne K. Palladino, Tilman Dingler, and
Myounghoon Jeon. 2013. Spearcons (speech-based
earcons) improve navigation performance in
advanced auditory menus. Human Factors, 55,
157-182.
Late-Breaking Work
CHI 2017, May 6–11, 2017, Denver, CO, USA
2016
... For the warning signals, we chose auditory warnings due to their easily manipulated directionality and wide utilization in modern vehicles. Although visual warning systems can also be used, auditory warnings appear to be most suitable because driving is already a visually demanding task (Hergeth et al., 2015;Sabic et al., 2017). Tactile warnings have been shown to yield faster response time than auditory and visual warnings (Mohebbi et al., 2009;Scott & Gray, 2008). ...
Article
Objective The present study investigated the design of spatially oriented auditory collision-warning signals to facilitate drivers’ responses to potential collisions. Background Prior studies on collision warnings have mostly focused on manual driving. It is necessary to examine the design of collision warnings for safe takeover actions in semi-autonomous driving. Method In a video-based semi-autonomous driving scenario, participants responded to pedestrians walking across the road, with a warning tone presented in either the avoidance direction or the collision direction. The time interval between the warning tone and the potential collision was also manipulated. In Experiment 1, pedestrians always started walking from one side of the road to the other side. In Experiment 2, pedestrians appeared in the middle of the road and walked toward either side of the road. Results In Experiment 1, drivers reacted to the pedestrian faster with collision-direction warnings than with avoidance-direction warnings. In Experiment 2, the difference between the two warning directions became nonsignificant. In both experiments, shorter time intervals to potential collisions resulted in faster reactions but did not influence the effect of warning direction. Conclusion The collision-direction warnings were advantageous over the avoidance-direction warnings only when they occurred at the same lateral location as the pedestrian, indicating that this advantage was due to the capture of attention by the auditory warning signals. Application The present results indicate that drivers would benefit most when warnings occur at the side of potential collision objects rather than the direction of a desirable action during semi-autonomous driving.
... In addition, auditory warnings have been shown to effectively attract a driver's attention while performing a driving task, which is often visually demanding (Baldwin, 2011;Petermeijer, Doubek, & de Winter, 2017). An auditory warning can be either speech-based containing sematic information (e.g., "car in blind spot") or non-sematic (e.g., a tone, or an earcon), which can influence driver behaviors differently (Sabic, Mishler, Chen, & Hu, 2017). The purpose of the current study was to examine the effect of level of automation and warning type on driver responses to novel critical events, using vehicle hacking attempts as a concrete example, in a driving simulator. ...
Article
Modern surface transportation vehicles often include different levels of automation. Higher automation levels have the potential to impact surface transportation in unforeseen ways. For example, connected vehicles with higher levels of automation are at a higher risk for hacking attempts, because automated driving assistance systems often rely on onboard sensors and internet connectivity (Amoozadeh et al., 2015). As the automation level of vehicle control rises, it is necessary to examine the effect different levels of automation have on the driver-vehicle interactions. While research into the effect of automation level on driver-vehicle interactions is growing, research into how automation level affects driver’s responses to vehicle hacking attempts is very limited. In addition, auditory warnings have been shown to effectively attract a driver’s attention while performing a driving task, which is often visually demanding (Baldwin, 2011; Petermeijer, Doubek, & de Winter, 2017). An auditory warning can be either speech-based containing sematic information (e.g., “car in blind spot”) or non-sematic (e.g., a tone, or an earcon), which can influence driver behaviors differently (Sabic, Mishler, Chen, & Hu, 2017). The purpose of the current study was to examine the effect of level of automation and warning type on driver responses to novel critical events, using vehicle hacking attempts as a concrete example, in a driving simulator. The current study compared how level of automation (manual vs. automated) and warning type (non-semantic vs. semantic) affected drivers’ responses to a vehicle hacking attempt using time to collision (TTC) values, maximum steering wheel angle, number of successful responses, and other measures of response. A full factorial between-subjects design with the two factors made four conditions (Manual Semantic, Manual Non-Semantic, Automated Semantic, and Automated Non-Semantic). Seventy-two participants recruited using SONA ( odupsychology.sona-systems.com ) completed two simulated drives to school in a driving simulator. The first drive ended with the participant safely arriving at school. A two-second warning was presented to the participants three quarters of the way through the second drive and was immediately followed by a simulated vehicle hacking attempt. The warning either stated “Danger, hacking attempt incoming” in the semantic conditions or was a 500 Hz sine tone in the non-semantic conditions. The hacking attempt lasted five seconds before simulating a crash into a vehicle and ending the simulation if no intervention by the driver occurred. Our results revealed no significant effect of level of automation or warning type on TTC or successful response rate. However, there was a significant effect of level of automation on maximum steering wheel angle. This is a measure of response quality (Shen & Neyens, 2017), such that manual drivers had safer responses to the hacking attempt with smaller maximum steering wheel angles. In addition, an effect of warning type that approached significance was also found for maximum steering wheel angle such that participants who received a semantic warning had more severe and dangerous responses to the hacking attempt. The TTC and successful response results from the current experiment do not match those in the previous literature. The null results were potentially due to the warning implementation time and the complexity of the vehicle hacking attempt. In contrast, the maximum steering wheel angle results indicated that level of automation and warning type affected the safety and severity of the participants’ responses to the vehicle hacking attempt. This suggests that both factors may influence responses to hacking attempts in some capacity. Further research will be required to determine if level of automation and warning type affect participants ability to safely respond to vehicle hacking attempts. Acknowledgments. We are grateful to Scott Mishler for his assistance with STISIM programming and Faye Wakefield, Hannah Smith, and Pettie Perkins for their assistance in data collection.
... Together with visual displays, auditory displays have been widely investigated based on multiple resource theory (MRT) [7,8] that predicts better performance when different modalities are employed. These studies have evaluated the difference in performance between different types of auditory displays such as speech, auditory icons, earcons, and spearcons [9][10][11][12]. However, most studies were conducted through empirical human-subject studies requiring extensive efforts and resources. ...
Conference Paper
In semi-autonomous vehicles (SAE level 3) that require driver's engagement in critical situations, it is important to secure reliable control transitions. There have been many studies on investigating appropriate auditory displays for takeover request (TOR) but most of them were empirical experiments. In the present study, we established two computational models using a Queuing Network Model Human Processor (QN-MHP) framework to predict a driver's reaction time to auditory displays for TOR. The reaction time for different sound types were modeled based on the results of subjective questionnaire in empirical studies. Separately, the reaction times for various non-speech sounds were modeled by using acoustical characteristics of sounds and previous empirical studies. It is one of a few attempts modeling the effects of auditory displays for TOR on the reaction time in autonomous driving. The current study will contribute to driving research by allowing us to simulate and predict drivers' behavior.
... For the warning signals, we chose auditory warnings due to their easily manipulated directionality and wide utilization in modern vehicles. Although visual warning systems can also be used, auditory warnings appear to be most suitable because driving is already a visually demanding task (Hergeth et al., 2015;Sabic et al., 2017). Tactile warnings have been shown to yield faster response time than auditory and visual warnings (Mohebbi et al., 2009;Scott & Gray, 2008). ...
Article
Assistance driving systems aim to facilitate human behavior and increase safety on the road. These systems comprise common systems such as forward collision warning systems, lane deviation warning systems, and even park assistance systems. Warning systems can communicate with the driver through various modalities, but auditory warnings have the advantage of not further tasking visual resources that are primarily used for driving. Auditory warnings can also be presented from a certain location within the cab environment to be used by the driver as a cue. Beattie, Baillie, Halvey, and McCall (2014) assessed presenting warnings in stereo configuration, coming from one source, and bilateral configuration, panned fully from left or right, and found that drivers felt more in control with lateral warnings than stereo warnings when the car was in self-driving mode. Straughn, Gray, and Tan (2009) examined laterally presented auditory warnings to signal potential collisions. They found that the ideal presentation of warnings in either the avoidance direction, in which the driver should direct the car to avoid a collision, or the collision direction, in which the potential collision is located, was dependent on time to collision. Wang, Proctor, and Pick (2003) applied the stimulus-response compatibility principle to auditory warning design by using a steering wheel in a non-driving scenario and found that a tone presented monaurally in the avoidance-direction led to the fastest steering response. However, the reverse finding occurred when similar experiments utilized a driving simulator in a driving scenario (Straughn et al., 2009; Wang, Pick, Proctor, & Ye, 2007).
Article
Existing driver models mainly account for drivers’ responses to visual cues in manually controlled vehicles. The present study is one of the few attempts to model drivers’ responses to auditory cues in automated vehicles. It developed a mathematical model to quantify the effects of characteristics of auditory cues on drivers’ response to takeover requests in automated vehicles. The current study enhanced queuing network-model human processor (QN-MHP) by modeling the effects of different auditory warnings, including speech, spearcon, and earcon. Different levels of intuitiveness and urgency of each sound were used to estimate the psychological parameters, such as perceived trust and urgency. The model predictions of takeover time were validated via an experimental study using driving simulation with resultant R squares of 0.925 and root-mean-square-error of 73 ms. The developed mathematical model can contribute to modeling the effects of auditory cues and providing design guidelines for standard takeover request warnings for automated vehicles.
Chapter
Driver Assistance System (DAS) is significant to reduce human errors and in turn road accidents, which is a major threat to the Indian road and transport infrastructure. DASs make use of computer vision and machine learning to learn the environment and update the driver with information for further action. The computer vision enables the system to see the environment through the various sensors equipped on the vehicles such as radars, lidars, etc. and machine learning gives the ability to understand it. Deep Learning is a subset of machine learning which works like humans. It processes the data, understands and creates patterns to take decisions. Object detection using deep learning identifies the objects present in an image or a video and classifies those objects as car, person, building, etc. In the present work, pretrained object detection model, SSD MobileNet is implemented on DATS_2020, a dataset for Indian traffic scenario.KeywordsTransfer learningObject detectionMachine learningAnnotations
Chapter
This paper presents the first and unique technique to predict unexpected circumstances using telematics data. Detailed statistics and analysis of drivers’ behavior with respect to the provided telematics data, is also performed. Part one, a new angle of using telematics, includes prediction of situations like high traffic congestion, increasing pedestrian and cyclist traffic as well as poor road conditions in turn trying to avoid probable road mishaps as the user is alerted in time. This application of telematics isn’t fully established in research and analysis and can be immensely useful to individual drivers as the end users. The next part can be helpful to large organizations providing taxi services to assess their drivers’ behavior and control rash driving for enhanced service. The prime objective of the project is to ensure the user’s safety and provide emergency assistance while trying to avoid varied types of road accidents based on the research and insights drawn from the provided fleet data. Post profound research, it is observed that the telematics data is utilized in applications such as air pollution tracking, training automated vehicles, reporting rash driving and route prediction. On the contrary, this idea of traffic prediction is off the mainstream, something never thought of earlier. Unlike the conventional methods like camera-based surveillance, this is a new advancement in traffic prediction techniques as well.KeywordsTelematicsData analysisRoad mishapsTraffic predictionDriver behaviorVehicles
Article
Full-text available
Objective: Auditory displays could be essential to helping drivers maintain situation awareness in autonomous vehicles, but to date, few or no studies have examined the effectiveness of different types of auditory displays for this application scenario. Background: Recent advances in the development of autonomous vehicles (i.e., self-driving cars) have suggested that widespread automation of driving may be tenable in the near future. Drivers may be required to monitor the status of automation programs and vehicle conditions as they engage in secondary leisure or work tasks (entertainment, communication, etc.) in autonomous vehicles. Method: An experiment compared memory for alerted events-a component of Level 1 situation awareness-using speech alerts, auditory icons, and a visual control condition during a video-simulated self-driving car ride with a visual secondary task. The alerts gave information about the vehicle's operating status and the driving scenario. Results: Speech alerts resulted in better memory for alerted events. Both auditory display types resulted in less perceived effort devoted toward the study tasks but also greater perceived annoyance with the alerts. Conclusion: Speech auditory displays promoted Level 1 situation awareness during a simulation of a ride in a self-driving vehicle under routine conditions, but annoyance remains a concern with auditory displays. Application: Speech auditory displays showed promise as a means of increasing Level 1 situation awareness of routine scenarios during an autonomous vehicle ride with an unrelated secondary task.
Article
Full-text available
Objective: Driver distraction and inattention are the main causes of accidents. The fact that devices such as navigation displays and media players are part of the distraction problem has led to the formulation of guidelines advocating various means for minimizing the visual distraction from such interfaces. However, although design guidelines and recommendations are followed, certain interface interactions, such as menu browsing, still require off-road visual attention that increases crash risk. In this article, we investigate whether adding sound to an in-vehicle user interface can provide the support necessary to create a significant reduction in glances toward a visual display when browsing menus. Methods: Two sound concepts were developed and studied; spearcons (time-compressed speech sounds) and earcons (musical sounds). A simulator study was conducted in which 14 participants between the ages of 36 and 59 took part. Participants performed 6 different interface tasks while driving along a highway route. A 3 × 6 within-group factorial design was employed with sound (no sound /earcons/spearcons) and task (6 different task types) as factors. Eye glances and corresponding measures were recorded using a head-mounted eye tracker. Participants' self-assessed driving performance was also collected after each task with a 10-point scale ranging from 1 = very bad to 10 = very good. Separate analyses of variance (ANOVAs) were conducted for different eye glance measures and self-rated driving performance. Results: It was found that the added spearcon sounds significantly reduced total glance time as well as number of glances while retaining task time as compared to the baseline (= no sound) condition (total glance time M = 4.15 for spearcons vs. M = 7.56 for baseline, p =.03). The earcon sounds did not result in such distraction-reducing effects. Furthermore, participants ratings of their driving performance were statistically significantly higher in the spearcon conditions compared to the baseline and earcon conditions (M = 7.08 vs. M = 6.05 and M = 5.99 respectively, p =.035 and p =.002). Conclusions: The spearcon sounds seem to efficiently reduce visual distraction, whereas the earcon sounds did not reduce distraction measures or increase subjective driving performance. An aspect that must be further investigated is how well spearcons and other types of auditory displays are accepted by drivers in general and how they work in real traffic.
Article
Full-text available
In-car infotainment systems (ICIS) often degrade driving performances since they divert the driver's gaze from the driving scene. Sonification of hierarchical menus (such as those found in most ICIS) is examined in this paper as one possible solution to reduce gaze movements towards the visual display. In a dual-task experiment in the laboratory, 46 participants were requested to prioritize a primary task (a continuous target detection task) and to simultaneously navigate in a realistic mock-up of an ICIS, either sonified or not. Results indicated that sonification significantly increased the time spent looking at the primary task, and significantly decreased the number and the duration of gaze saccades towards the ICIS. In other words, the sonified ICIS could be used nearly exclusively by ear. On the other hand, the reaction times in the primary task were increased in both silent and sonified conditions. This study suggests that sonification of secondary tasks while driving could improve the driver's visual attention of the driving scene. Copyright © 2015 Elsevier Ltd and The Ergonomics Society. All rights reserved.
Article
Full-text available
Auditory display research for driving has mainly examined a limited range of tasks (e.g., collision warnings, cell phone tasks). In contrast, the goal of this project was to evaluate the effectiveness of enhanced auditory menu cues in a simulated driving context. The advanced auditory cues of ‘spearcons’ (compressed speech cues) and ‘spindex’ (a speech-based index cue) were predicted to improve both menu navigation and driving. Two experiments used a dual task paradigm in which users selected songs on the vehicle’s infotainment system. In Experiment 1, 24 undergraduates played a simple, perceptual-motor ball-catching game (the primary task; a surrogate for driving), and navigated through an alphabetized list of 150 song titles—rendered as an auditory menu—as a secondary task. The menu was presented either in the typical visual-only manner, or enhanced with text-to-speech (TTS), or TTS plus one of three types of additional auditory cues. In Experiment 2, 34 undergraduates conducted the same secondary task while driving in a simulator. In both experiments, performance on both the primary task (success rate of the game or driving performance) and the secondary task (menu search time) was better with the auditory menus than with no sound. Perceived workload scores, as well as user preferences favored the enhanced auditory cue types. These results show that adding audio, and enhanced auditory cues in particular, can allow a driver to operate the menus of in-vehicle technologies more efficiently while driving more safely. Results are discussed with multiple resources theory.
Article
Full-text available
Awareness of features in our environment is essential for many daily activities. While often awareness of such features comes from vision, this modality is sometimes unavailable or undesirable. In these instances, auditory cues can be an excellent method of representing environmental features. The study reported here investigated the learnability of well known (auditory icons, earcons, and speech) and more novel (spearcons, earcon-icon hybrids, and sized hybrids) sonification techniques for representing common environmental features. Spearcons, which are speech stimuli that have been greatly sped up, were found to be as learnable as speech, while earcons unsurprisingly were much more difficult to learn. Practical implications are discussed.
Article
The present study examined aspects of a type of auditory feedback that aims at making interactions with menus more efficient. Specifically, manipulations were made to the auditory counterparts of common computer commands and unrelated words to examine identification performance. By investigating the point at which sped-up versions of text-to-speech, spearcons, can be identified accurately as words, the boundaries at which the efficiency of spearcon use begins to decline can be better defined. Additionally, by examining the effect of category membership on speed and accuracy of spearcon identification, whether an overarching category can facilitate spearcon use can be examined. Results in two experiments demonstrated that spearcon identification began to decline drastically after linear compression leading to 40% of the original audio length. Reaction time data also demonstrated that spearcon efficiency began to decline after the same level of linear compression. Efficiency scores combining reaction time and accuracy also supported the bottom limit of spearcons at 40% of original audio length.
Article
With shrinking displays and increasing technology use by visually impaired users, it is important to improve usability with non-GUI interfaces such as menus. Using non-speech sounds called earcons or auditory icons has been proposed to enhance menu navigation. We compared search time and accuracy of menu navigation using four types of auditory representations: speech only; hierarchical earcons; auditory icons; and a new type called spearcons. Spearcons are created by speeding up a spoken phrase until it is not recognized as speech. Using a within-subjects design, participants searched a 5 x 5 menu for target items using each type of audio cue. Spearcons and speech-only both led to faster and more accurate menu navigation than auditory icons and hierarchical earcons. There was a significant practice effect for search time, within each type of auditory cue. These results suggest that spearcons are more effective than previous auditory cues in menu-based interfaces, and may lead to better performance and accuracy, as well as more flexible menu structures.