ArticlePDF Available

The shallow of your smile: The ethics of expressive vocal deep-fakes

The Royal Society
Philosophical Transactions B
Authors:

Abstract

Rapid technological advances in artificial intelligence are creating opportunities for real-time algorithmic modulations of a person’s facial and vocal expressions, or ‘deep-fakes’. These developments raise unprecedented societal and ethical questions which, despite much recent public awareness, are still poorly understood from the point of view of moral psychology. We report here on an experimental ethics study conducted on a sample of N = 303 participants (predominantly young, western and educated), who evaluated the acceptability of vignettes describing potential applications of expressive voice transformation technology. We found that vocal deep-fakes were generally well accepted in the population, notably in a therapeutic context and for emotions judged otherwise difficult to control, and surprisingly, even if the user lies to their interlocutors about using them. Unlike other emerging technologies like autonomous vehicles, there was no evidence of social dilemma in which one would, for example, accept for others what they resent for themselves. The only real obstacle to the massive deployment of vocal deep-fakes appears to be situations where they are applied to a speaker without their knowing, but even the acceptability of such situations was modulated by individual differences in moral values and attitude towards science fiction. This article is part of the theme issue ‘Voice modulation: from origin and mechanism to social impact (Part II)’.
royalsocietypublishing.org/journal/rstb
Research
Cite this article: Guerouaou N, Vaiva G,
Aucouturier J-J. 2021 The shallow of your
smile: the ethics of expressive vocal
deep-fakes. Phil. Trans. R. Soc. B 377:
20210083.
https://doi.org/10.1098/rstb.2021.0083
Received: 31 March 2021
Accepted: 28 July 2021
One contribution of 12 to a theme issue Voice
modulation: from origin and mechanism to
social impact (Part II).
Subject Areas:
cognition, behaviour
Keywords:
voice transformation, ethics, deep-fake,
moral psychology, emotions
Author for correspondence:
Nadia Guerouaou
e-mail: nadia.guerouaou@chru-lille.fr
Electronic supplementary material is available
online at https://doi.org/10.6084/m9.figshare.
c.5662276.
The shallow of your smile: the ethics of
expressive vocal deep-fakes
Nadia Guerouaou
1,2
, Guillaume Vaiva
2
and Jean-Julien Aucouturier
3,4
1
Science and Technology of Music and Sound, IRCAM/CNRS/Sorbonne Université, Paris, France
2
Lille Neuroscience and Cognition Center (LiNC), Team PSY, INSERM U-1172/CHRU Lille, France
3
FEMTO-ST, UBFC/CNRS, Besançon, France
4
Alta Voce SAS, Houilles, France
4
Centre National de Ressource et Résilience (CN2R Lille Paris), Lille, France
NG, 0000-0003-2319-623X
Rapid technological advances in artificial intelligence are creating opportu-
nities for real-time algorithmic modulations of a persons facial and vocal
expressions, or deep-fakes. These developments raise unprecedented
societal and ethical questions which, despite much recent public awareness,
are still poorly understood from the point of view of moral psychology.
We report here on an experimental ethics study conducted on a sample of
N= 303 participants (predominantly young, western and educated), who
evaluated the acceptability of vignettes describing potential applications of
expressive voice transformation technology. We found that vocal deep-
fakes were generally well accepted in the population, notably in a thera-
peutic context and for emotions judged otherwise difficult to control, and
surprisingly, even if the user lies to their interlocutors about using them.
Unlike other emerging technologies like autonomous vehicles, there was
no evidence of social dilemma in which one would, for example, accept
for others what they resent for themselves. The only real obstacle to the mas-
sive deployment of vocal deep-fakes appears to be situations where they are
applied to a speaker without their knowing, but even the acceptability of
such situations was modulated by individual differences in moral values
and attitude towards science fiction.
This article is part of the theme issue Voice modulation: from origin and
mechanism to social impact (Part II).
1. Introduction
The human facial and vocal expressions have evolved as signals to inform and
manipulate others [1,2]. By continuously modulating our facial muscles and the
phonatory and articulatory structures of our vocal apparatus, we provide a rich,
flexible non-verbal back-channel to our daily conversations, communicating our
emotional states such as joy or surprise [3,4], our social intents such as warmth
or dominance [5,6], or our epistemic attitudes, such as certainty or doubt [7,8].
While our facial and vocal expressions were shaped by a long and delicate
interplay of biological and cultural evolution [9,10], spectacular technological
advances occurring in the past few years may soon dramatically alter how we
use and experience these behaviours in daily life. Recent progress in signal pro-
cessing has indeed made possible the real-time manipulation of e.g. facial
expressions such as smiles [11] and vocal expressive cues such as pitch [12] or
timbre [11]. Perhaps even more radically, recent advances in deep neural network
architectures have provided a flexible way to learn and generate mappings (or
deep-fakes[13]) between pairs of stimuli, and opened possibilities to para-
metrically manipulate individual facial actions [14] (figure 1b) or convert one
voice into several emotional variants [15]. In just a few years, combined with
the unprecedented adoption of remote communication software such as video
conferencing and virtual meetings, we have come to a situation where it is
© 2021 The Authors. Published by the Royal Society under the terms of the Creative Commons Attribution
License http://creativecommons.org/licenses/by/4.0/, which permits unrestricted use, provided the original
author and source are credited.
difficult to trust whether the smiles, laughs and frowns of
our conversation partners are genuine or algorithmically
modulated (figure 1c).
The goal of this paper is to initiate the data-driven study
of expressive deep-fakes ethics (specifically here, vocal deep-
fakes) and, inspired by the methodology of experimental
ethics[16], to quantify societal expectations about the
principles that should guide their deployment.
The realistic, artificial manipulation of expressive behav-
iour raises unprecedented societal and ethical questions.
First, it raises concerns about truthfulness. Because expressive
behaviours are often thought to provide genuine cues about
the senders emotional states [17], the ability to arbitrarily
manipulate these displays opens avenues for deception: one
may use, e.g. a facial filter to fake a smile despite having no
intent to affiliate, or a voice transformation to appear more
certain than one really is. Second, they raise concerns about
fairness. Expressive behaviours in verbal interactions strongly
influence subsequent behaviours. It is already known that
vendors displaying positive, authentic expressions while
interacting with customers sell more mobile phones [18], or
that negotiators faking anger in commercial discussions
obtain better prices [19]. The algorithmic manipulation of
expressions designed for such situations may coerce people
into making unwarranted or unfair decisions. Third, they
raise concerns about autonomy. Non-verbal influences on
behaviour are often non-conscious: in a study with voice trans-
formation, mock patients calling 911 medical triage with a
more dominant voice obtained more urgent medical responses
from doctors but doctors did not attribute the cause of their be-
haviour to the patients voice; rather, they wrongly attributed
it to more urgent medical situations [20] (figure 2, bottom).
Technologies able to trigger such unconscious reactions are
therefore intrinsically manipulative, as people may not be
able to identify the transformation as the cause for their sub-
sequent behaviour. Finally, they also raise concerns about
transparency, as their deployment in virtual conversations
lends itself to situations where a speaker does not know how
their interlocutor is hearing or seeing them, i.e. whether a
transformation of their own voice or face is applied without
their knowing.
However, none of these deontological concerns warrants a
straightforward moral objection to the deployment of expres-
sive transformation technologies, because each of them also
create opportunities for highly desirable situations. First, the
fact that, e.g. a smiling voice transformation can be used to
appear happier than one really is becomes highly desirable
in the case of patients who cannot easily express emotions
(e.g. amyotrophic lateral sclerosis patients who rely on assis-
tive voice technology for communication, [22]). Second, the
fact that voice or face transformations can coerce observers
into subsequent actions can be desirable in interventions
where people are nudged into positive behaviour [23], for
instance reducing aggressive behaviour in call-centre conver-
sations by transforming the operators fatigued voice [24], or
applying a gender voice transformation on an online hiring
platform to alleviate gender biases [21] (figure 2, top). Third,
the fact that expressive transformations can be processed
unconsciously may be desirable in situations where this
increases their effectiveness, as seen in emotional vocal
feedback [25].
Societal expectations in such situations are non-trivial and
important to understand in order to inform and regulate the
deployment of deep-fakes in commercial products or clinical
protocols. A recently emerging methodology for doing so is
that of experimental ethics, in which moral judgements about
various situational vignettes are collected from relatively large
samples of online participants. In recent years, this method-
ology has been applied to quantify societal attitudes towards
new technologies such as autonomous vehicles [16] or brain
stimulation [26], potential pub lic policies such as l egalizing pay-
ments to kidney donors [27], but also downright futuristic
scenarios such as mind upload [28], sex robots [29] or cognitive
enhancement with brain implants [30] (figure 1d). The exper-
imental ethics approach allows comparing different situation
variants that may make or break dilemmas (e.g. whether
imagining oneself as the conductor of an autonomous car
changes ones attitude to how to react to accidents[16]) and
whether these effects are modulated by individual differences
(e.g. whether a persons familiarity with science fiction
themes modulates their attitude towards robots[29]).
Here, we employed the methodology of experimental
ethics to gauge societal attitudes towards emotional voice
transformation technology. We asked N= 303 online partici-
pants to read 24 short text vignettes describing potential
applications of vocal deep-fakes, and rate how morally
acceptable they thought each scenario is. Participants were
presented with a cover story describing an imaginary
so apparently this filter
reveals what you will look
like with veneers
1872 2018 2020 science fiction
Figure 1. From Darwin to deep-learning: rapid technological advances in artificial intelligence create opportunities for real-time algorithmic modulations of facial
and vocal expressions, which raises unprecedented societal and ethical questions. From left to right: (a) original studies of human facial expressions employed electric
stimulation to induce muscle contraction (Guillaume Duchenne de Boulogne, reproduced in [1]); (b) manipulation of individual action units in still photographs using
Generative Adversarial Networks (GANimation [14]); (c) real-time smile filters in commercial video sharing plateforms (Tiktok, ByteDance Ltd., Beijing, China); (d) still
from the Arkangelepisode of dystopian science fiction television series Black Mirror (Endemol Shine UK Ltd., 2017) in which parents equip their children with anti-
violence visual filters via a brain implant. Here, the device visually filters out a dog aggressively barking at the child, directly in the childs mind. (Online version in
colour.)
royalsocietypublishing.org/journal/rstb Phil. Trans. R. Soc. B 377: 20210083
2
hardware device (MyVoicePlus) able to transform the
emotional quality of a voice in real-time, which was said to
be currently considered by a startup company for commercial
deployment in various situations. The vignettes describing
potential applications of the device varied among four factors:
(i) whether the user of the device was the participant or
an unknown other;
(ii) whether the transformations were used therapeutically,
or to enhance user capacities;
(iii) whether the transformations operated on positive
(enhancing smiling) or negative emotional expressions
(reducing anxiety, reducing anger);
(iv) and whether the transformation affected how the users
voice is heard by others (i.e. the users production),
how the user hears other personsvoices (i.e. the
users perception), or whether it is used in a situation
where the user hears their own manipulated voice
(i.e. feedback).
For each vignette, participants first rated the acceptability
of the situation, and were then presented with two potential
dilemmas involving lying about the true purpose of the trans-
formation in order to improve its effectiveness. Finally, for all
of these judgements, we examined associations with individ-
ual differences in participantsattitudes towards morality
(Moral Foundations Questionnaire, MFQ [31], measuring
factors of harmcare, fairnesscheating, loyaltybetrayal,
authoritysubversion and puritydegradation) and toward
technology and science fiction (Science Fiction Hobbyism
Scale, SFH; [28]), two factors found relevant in previous
research about the moral reception of new technologies
[26,28,29,32] (see §4 for details of the procedure).
Although our study is exploratory and we did not prereg-
ister any formal hypotheses, a number of loose predictions can
be made from the literature about how our variables of interest
impact participantsmoral judgements. First, similar exper-
iments with emerging technologies such as autonomous
vehicles [16] or brain stimulation [26] have documented situ-
ations of social dilemma, in which participants accept things
for themselves (i.e. a car that favours its driver, rather than
pedestrians) that they would otherwise reject for others.
Second, across diverse forms of enhancement (e.g. memory,
general intelligence, mood, etc.), participants are widely
reported to be more comfortable with technologies that
enhance capacities towards the norm (i.e. that are used thera-
peutically) than above the norm [30,33]. Finally, to the best of
our knowledge, there is no straightforward equivalent in the
literature of whether, e.g. manipulating positive or negative
emotions, or manipulating a users perception or production,
has any impact on a participants judgement of acceptability.
Whether participants feel more comfortable with, e.g. smiling
or anxiety filters, and filters that affect their produced voice or
their perception of how others sound, is an open non-trivial
question [34], which our study wishes to address.
2. Results
(a) Acceptability of overtly using the technology
We first evaluated how morally acceptable our participants
(N= 303) thought the use of a voice transformation device
was, when the true purpose of the technology was overtly
known to all involved parties.
(i) Voice transformations are in general well accepted in the
population
Across situations, the moral acceptability of overt vocal trans-
formation was strongly significantly higher than neutral
(M= 6.49 > 5; one-sample t-test against mid-point, averaging
all acceptance scores across vignettes: t(302) = 146, p< 0.001).
Because of heteroscedasticity (Breush-Pagan: F(6, 296) =
3.23, p= 0.004), we tested the effect of individual character-
istics on this judgement with multiple iterated re-weighted
least squares (IRLS) regression (Huber weights, HC3
A’s voice is transformed to mask their
belonging to a minority group
desirable outcome:
reduced discrimination,
better hiring decisions,
increased workplace diversity
A is from a minority
group discriminated
against in hiring
decisions
B is a recruiter
interviewing A for a
position via an online
platform
B is a triage
operator for 911
responding to A’s
phone call
problematic outcome:
B is coerced into giving A undue
medical resources, which are then
denied to more urgent cases
A is a patient
demanding urgent
service from 911
A’s voice is transformed to sound
more assertive and dominatin
g
FX
FX
AB
Figure 2. Very similar uses of voice transformation technology can lead to both desirable or problematic situations. Top: a voice transformation is used to mask the
sex, accent or ethnicity of a user to eliminate discrimination in online hiring services. Situation inspired by genuine practice by the interviewing.io company [21].
Bottom: a voice transformation is used to increase the perceived dominance of a patient calling emergency medical services, who consequently gets undue medical
resources from triage operators at the expense of other more urgent cases. Situation inspired by the authorsexperimental work [20].
royalsocietypublishing.org/journal/rstb Phil. Trans. R. Soc. B 377: 20210083
3
correction). Acceptability was significantly positively associ-
ated with the participantsfamiliarity with science fiction
(β= 0.014, z= 2.75, p= 0.006; figure 3c) and marginally posi-
tively associated to participants reliance on MFQ purity
(PU) (β= 0.04, z= 1.85, p= 0.064). No other MFQ factors
regressed significantly (all ps.0:1). The marginal positive
association with MFQ PU differed from other studies of simi-
lar technologies where purity was found negatively correlated
with acceptability (e.g. mind upload [28]; sex robots [30]).
(ii) A therapeutic context makes them even more acceptable
We tested the effect of the goal to repair or enhance on
situation acceptability by averaging within-participant
scores for overt acceptability over therapeutic (n= 6 vignettes)
and enhancing situations (n= 6 vignettes), and testing for
population differences with a one-way repeated-measure
ANOVA. Repairenhance had a significant main effect on
situation acceptability (F(1, 302) = 47, p< 0.001), with thera-
peutic situations (M= 6.7) being (even) more acceptable
than enhancing situations (M= 6.2; figure 3a).
To test whether the effect of repair or enhance was associ-
ated with individual characteristics, we computed the
within-participant difference between acceptability scores
averaged over both types of vignettes, and computed mul-
tiple ordinary least-square (OLS) regression (BreuschPagan
heteroscedasticity test: F(6, 296) = 0.39, p= 0.88). The better
acceptability of repair situations was not significantly associ-
ated with individual differences in MFQ or science fiction
familiarity (R
2
= 0.008, F(6, 296) = 0.38, p= 0.89).
(iii) Manipulating perception is less acceptable than
manipulating production
Similarly, we tested the effect of whether situations described
voice transformation as affecting how the users voice is
heard by others (condition production:n= 4 vignettes), how
the user hears other personsvoices (condition perception:
n= 4 vignettes), or whether the user hears their own manipu-
lated voice (condition feedback:n= 4 vignettes) by averaging
acceptability scores within-participant over the three types
of vignettes and testing for population differences with a
one-way repeated-measure ANOVA. There was a significant
effect of the productionperceptionfeedback variable (F(2,
604) = 7.5, p= 0.001), with transformations affecting the
users production being more acceptable than perception
and feedback. Both latter conditions share the fact that the
device manipulates what the participant hears, regardless
of whether it is the participants own voice or that of
another person.
We tested for associations with individual characteristics
by computing the within-participant pairwise differen-
ces between acceptability scores averaged over all three
types of vignettes, and computing multiple OLS regres-
sion (BreuschPagan heteroscedasticity test: perception
production F(6, 296) = 0.27, p= 0.96; feedbackproduction
F(6, 296) = 0.66, p= 0.6815). The difference of acceptability
between these situations was not associated with participant
MFQ or SFH (perceptionproduction: R
2
= 0.006, F(6, 296) =
0.31, p= 0.93; feedbackproduction: R
2
= 0.011, F(6, 296) =
0.571, p= 0.75).
7.0
6.8
6.6
6.4
6.2
6.0 2
3
4
5
6
7
8
9
6.5
6.0
5.5
5.0
4.5
4.0
3.5
3.0
2
1
3
4
5
6
7
8
9
2
1
3
4
5
6
7
8
9
repair smile anxiety anger 20 30 40 50 60 70 80 90 100
familiarity with SF
510152025
MFQ purit
y
10 12 14 16 18 20 22 24 26
MFQ fairness
participant other
who is being lied to
enhance
context transformed expression
who wears the device
participant
other
acceptability of overt use
acceptability of lying
covert use overt use
(b)(a) (c)
(e)(d)(f)
Figure 3. Moral judgements of overt and covert use of voice transformations. Top row: overt use. (a) The moral acceptability of overt vocal transformation was
higher than the neutral midpoint, and therapeutic transformations even more so than transformations used to enhance user capacities. (b) Situations in which
transformations aimed at weakening the two negative emotions of anxiety or anger were better accepted than situations in which transformations aimed to enhance
smiling. (c) Across situations, acceptability was positively associated with the participantsfamiliarity with science fiction. Bottom row: covert use involving lying
about the true purpose of the transformation in order to improve its effectiveness. (d) Participants considered it morally acceptable that the user of the trans-
formation hides its true purpose to others but hiding the transformation to the person using the device was totally unacceptable. (e,f) The acceptability of lying to
the person using the device was negatively associated with the participantsconcern with fairness, and positively with purity. (a,b,d) Across conditions, there was no
effect of whether the user of the device was the participant or an unknown other. Error bars: 95% confidence intervals. (Online version in colour.)
royalsocietypublishing.org/journal/rstb Phil. Trans. R. Soc. B 377: 20210083
4
(b) Acceptability of covert uses
For each situation, we then tested the acceptability of lying
about the true purpose of the device, in order to increase
the transformations effectiveness, in two situations which
either involved the users lying to their interlocutors, or the
devices prescribers lying to the user themselves. Because
using the transformation overtly was generally well accepted
(see above), and because we presented situations in a context
where lying about the transformations would also improve
their effectiveness (see §4), these situations can be regarded
as genuine moral dilemmas in which the deontologically
blamable act of lying is balanced by the utilitarian value of
the resulting improvement of performance.
(i) Using the transformation covertly is not a problem
Although more acceptable situations were more acceptable
to lie about (OLS regression over vignettes averaged
between-participants: R
2
= 0.61, F(1, 22) = 34.61, p< 0.001;
BreuschPagan heteroscedasticity test: F(1, 22) = 0.13, p=
0.72), lying was generally regarded as non acceptable by
our participants (M= 4.69 < 5, t(302) = 3.19, p= 0.001).
However, there was a very large interaction with which
person is being lied to (one-way rm-ANOVA: F(1, 302) =
631, p< 0.001; figure 3d): somewhat surprisingly, participants
considered it morally acceptable that the user of this device
hides its true purpose to others (M= 6.08 [5.87, 6.3];
one-sample t-test against mid-point: t(302) = 9.99, p< 0.001).
Because of marginal heteroscedasticity (BreuschPagan:
F(6, 296) = 1.72, p= 0.11), we tested the association of the
acceptability of userslying to others with individual charac-
teristics with multiple IRLS regression. The acceptability of
lying to others was not found associated with any of the
MFQ subscales (best, PU: β= 0.033, z= 1.33, p= 0.18), but
was positively influenced by science fiction familiarity
(β= 0.01, z= 1.95, p= 0.05).
(ii) unless it is hidden from the user of the device
However, hiding the transformation to the person using the
device appeared totally unacceptable (M= 3.3 < <5, one-
sample t-test against mid-point: t(302) = 14.9, p< 0.001),
even though the transformation was presented as more effec-
tive for the user when doing so (figure 3d).
As above, we tested the association of the acceptability of
lying to the devices user with individual characteristics,
using multiple IRLS regression (BreuschPagan heteroscedas-
ticity test: F(6, 296) = 1.28, p= 0.26). The low acceptability of
lying to the user was driven (i.e. negatively associated) by
participants high on the MFQ subscale of fairness (β=
0.1395; z=3.461, p= 0.001; figure 3e) but attenuated (i.e.
positively associated) for participants high on MFQ purity
(β= 0.0930, z= 3.65, p< 0.001; figure 3f) and loyalty (β=
0.09, z= 2.71, p= 0.007). The acceptability of lying to the
user was also associated with science fiction familiarity (β=
0.014, z= 2.38, p= 0.017), with greater familiarity making
lying to the user more acceptable.
(c) Acceptability of voice transformations is not
influenced by seeking self profits
To test for the effect of either depicting situations where the
user was the participant or an unknown person, we conducted
a mixed ANOVA with selfother as a between-participant
factor, and vignette conditions (repairenhance, positive
negative transformation, productionperceptionfeedback)
as within-participant factors. There was no statistical differ-
ence of acceptability between overt situations which depicted
the participant as the user benefiting of the device (M= 6.43),
and situations where the user was an unknown person (M=
6.55; no main effect, F(1, 301) = 0.37, p= 0.54). Neither did the
effect of selfother interact with any of the other variables:
regardless of whether the user was themselves or others, par-
ticipants thought similarly of differences between situations
meant to repair and enhance (no interaction selfother ×
repairenhance: F(1, 301) = 1.68, p= 0.20; figure 3a), of differ-
ences between situations involving smiling, anger or anxiety
(no interaction selfother × transformation: F(1, 301) = 1.43,
p= 0.23; figure 3b), and of differences between devices
affecting the users production, perception or feedback
(no interaction selfother × productionperceptionfeedback:
F(2, 602) = 0.047, p= 0.95).
Similarly, participants did not judge less acceptable
the covert situations where the true purpose of the device
was hidden from them (regardless of whether they were its
user, or not), compared to situations where it was hidden
from unknown others (rm-ANOVA, with concealed partici-
pantother as within-participant factor, F(1, 302) = 0.0026,
p= 0.87; figure 3d). In other words, the relatively high accept-
ability of userslying to others did not depend on whether the
participant was the user of the device or the person whom the
transformation is hidden from; and the low acceptability of
lying to the devices users did not depend either on whether
the user was the participant themselves or an unknown other.
In sum, contrary to situations like pedestrian dilemmas in
autonomous vehicles [16], there was radically no evidence of
social dilemma regarding the use of voice transformations, in
which one would e.g. accept for themselves what they would
blame others for, even in situations involving the blamable
act of lying.
(d) The nature of the emotion impacts the moral
acceptability of the transformation
Finally, we tested the impact of what emotion is transformed
on the acceptability of the situation, as well as the interaction
with the repairenhance factor. We averaged within-partici-
pant scores of overt acceptability over repairenhance
situations concerning anxiety (n= 4 vignettes; repair: 2),
anger (n= 2 vignettes; repair: 1) and smile vignettes (n=6;
repair: 3), and tested for population differences with a two-
way rm-ANOVA.
There was a main effect of emotion: situations in which
transformations aimed at weakening the two negative
emotions of anxiety (M= 6.8) or anger (M= 6.5) were better
accepted than situations involving transformations enhan-
cing smile (F(2, 604) = 24.47, p< 0.001), although the latter
remained well accepted at M= 6.3 ( figure 3b).
The effect of emotion also interacted significantly with
the repairenhance factor (F(2, 604) = 21.3, p< 0.001), with
transformations aiming to weaken negative emotions benefit-
ing more of the therapeutic condition (Δ= +0.56) than the
transformation targeting positive emotions (Δ= +0.35). The
effect was maximal for the repair of anxiety (repair: M=
7.17; enhance: M= 6.34).
Similarly, in covert situations, it was more more accepta-
ble to hide the purpose of a transformation aiming to
royalsocietypublishing.org/journal/rstb Phil. Trans. R. Soc. B 377: 20210083
5
weaken negative emotions than a transformation aiming
to enhance smile (one-way rm-ANOVA; main effect of
transformation: F(2, 604) = 8.3, p< 0.001).
Finally, we tested whether these differences between posi-
tive and negative transformations were associated with
individual differences, by computing the within-participant
pairwise differences between acceptability scores averaged
over all three types of transformations, and computing mul-
tiple OLS regression (BreuschPagan heteroskedasticity test:
anxietysmile F(6, 296) = 0.59, p= 0.74; angersmile F(6,
296) = 0.73, p= 0.62). The difference of acceptability between
these situations was not associated with participant MFQ or
SFH (anxietysmile: R
2
= 0.031, F(6, 296) = 1.58, p= 0.15;
angersmile: R
2
= 0.023, F(6, 296) = 1.137, p= 0.34).
3. Discussion
We reported here on an experimental ethics study in which
N= 303 online participants evaluated the acceptability of vign-
ettes describing potential applications of expressive voice
transformation technology. We found that vocal deep-fakes
were generally well accepted, notably in a therapeutic (versus
enhancement) context; when they corrected negative emotions
rather than enhanced positive emotions; and when they
manipulated a speakers production rather than perception.
Surprisingly, transformations remained well-accepted even
when the user lied to their interlocutors about using them
and, unlike other emerging technologies such as autonomous
vehicles, there was no evidence of social dilemma in which
one would accept for others what they resent for themselves.
The only real moral objection to vocal transformations
appeared related to situations in which they were applied to
a speaker without their knowing, with the acceptability of
such situations being modulated by individual differences in
moral values and attitude towards science fiction.
The fact that voice transformations are generally well-
accepted, with average scores across situations well above
the scale mid-point, first and foremost shows that the
western, young, educated population studied here is sym-
pathetic to the idea of customizing ones own emotional
expression with technology, when these technologies
become available. This attitude, at least for the range of scen-
arios tested here, seems consistent with transhumanistic
views for which technology should be used to enhance
human capacities and improve happiness [35] as well as
control for emotional or neurological limitations (e.g. taking
anti-love drugs to curb affect in divorce situations [36]).
Contrary to other moral psychology studies where individ-
ual attitudes to MFQ purity negatively correlated with
acceptability of cognitive enhancement or mind upload [28],
acceptability here was facilitated by the participantsreliance
on the purity dimension. This may suggest that voice trans-
formations are not seen as a breech of human integrity, but
rather as a way to improve control and self-determinacy
(i.e. an anthropotechnical tool for self-customization [37]). In a
contemporary society promoting continuous self improvement,
the good reception of this kind of technology is thus perhaps
not surprising [38]. However, it should be noted that the MFQ
purity construct has come under recent debate (e.g. it may be
interpreted differently by religious and non-religious individ-
uals [39]), and further research is needed to ascertain what
this construct measures in our specific sample of participants.
The good general acceptance of voice transformations
was further improved in therapeutic situations, which were
judged more acceptable that situations merely aiming to
enhance user capacities [40]. This attitude is consistent with
what is reported in other empirical studies of cognitive
enhancement [26,30,33], and with imperatives put forward
by the bioethics literature [41,42]. It confirms that the
therapy-enhancement distinction is morally salient to the
public concerning potential display of expressive voice
transformation technology.
Acceptance was also higher for situations which manipu-
lated the production of an expression than situations which
manipulated its perception. That participants should be
biased against the latter somehow contradicts the expectation
that covert changes that are internal to the individual would
have less broad impact on others than changes affecting their
outward expression [43]. This preference may reflect a worry
about having ones real experience distorted, as one could
worry e.g. about mood-enhancer drugs such as SSRIs altering
ones sense of living truly (is it me or the Prozac enjoying this?
[40,44]), even though in the case of Prozac these bioethical
concerns do not seem shared by the general population
[45]. Since the production and perception situations could
be compared respectively with the use of Instagram filters
(which are now common; figure 1c) over augmented-reality
(AR) glasses (which are not yet), it would be interesting to
follow up on these results in the next few months, as several
announced AR devices such as Apple Glasses may gain
popularity and modify these attitudes ([46]; see also below
about science fiction familiarity).
In a second set of questions, we collected judgements
about concealed-use situations, and presented them in a con-
text where lying about the transformations would also
improve their effectiveness (see §4 Judge how acceptable it is
to lie to your entourage [], knowing that this would improve
the effectiveness of the device). The fact that voice transform-
ations are generally thought desirable in overtsituations
makes these covertsituations appear as genuine moral
dilemmas, in which the deontological imperative against
lying is balanced against the utilitarian benefits of self-
improvement. For these situations, both sides of the debate
were clearly reflected in participant judgements: on the one
hand, acceptance of lying was negatively associated with
MFQ fairness; on the other hand, as was the case in overt
situations, acceptance scores for these situations were also
positively associated with MFQ purity, which attenuated
the generally low acceptability of covert use.
Strikingly though, in all of these dilemma as well as in the
less problematic overtsituations, we found radically no evi-
dence of a social dilemma where a participant would refuse
for themselves what they think acceptable for others. This
held whether participants envisioned modifying their own
voice, or that of others; and whether participants were
being lied to regarding their perception, or whether they
lied to others. This absence of effect of who benefits from
the device when judging its acceptability is in stark contrast
with typical sacrificial scenarios (like the trolley problem or,
more recently, pedestrian versus driver dilemma in auton-
omous vehicles), in which participants tend to value self-
preservation [16,47]. This suggests that participants judge
voice-transformation technology primarily with a utilitarian
perspective, treating the welfare of everyone as of equal
importance, from the point of view of the universe[48]
royalsocietypublishing.org/journal/rstb Phil. Trans. R. Soc. B 377: 20210083
6
regardless of whether they are near or far, our children and
friends or absolute strangers, human or animal [43]. While
this does not mean that self-preservation biases could not be
created, for instance for situations involving finite supply
[26] or larger individual cost [16], the fact that voice transform-
ation should be judged so impartially suggests that there
currently is no social obstacle to the massive deployment of
such technologies in (here, western) societies.
Even though there was no effect of selfother, covert
dilemma was very strongly biased against lying to the
person wearing the prosthesis (i.e. regardless of who that
person was: self or other). This attitude may be an effect of
describing the device in our cover story as a physical prosthe-
sis, for which installingit covertly would be seen as an
unacceptable breach of consentautonomy [49]. To control
for physicality, future work could e.g. extend this study to
assess the acceptability of software effects (filters) deployed
in virtual meeting software.
Unexpectedly, transformations aiming to enhance positive
expressions (smiles) were judged less acceptable than those
aiming to reduce negative expressions (anxiety, anger). This
asymmetric pattern of result contrasts with a purely hedonic
view, in which making people feel as good as possible, and
feel least bad[50] would be equally valued. Rather, it may
indicate that curbing negative expressions is valued less for
the gain of valence than for an Aristotelician inclination for
control over oneself, because negative emotions such as por-
trayed here (stress, anxiety, fear) are viewed as less deliberate
and more automatic than smiling [51]. This view is also con-
sistent with our interpretation of MFQ purity as valuing self-
determinacy. If true, this may prefigure a situation where,
when broadly adopted, expressive technology would shift
the moral responsibility associated to certain emotions or
behaviours: expressions which were once normal to not con-
trol (e.g. one cannot be blamed for stress [51]) may become
controllable, and thus blamable and subjected to social
demand (e.g. why didnt you put stress-control on?, [52]). To
further test this idea, it would be interesting to examine scen-
arios involving non-deliberate positive expressions (eg. using
a transformation to avoid giggling uncontrollably at a fun-
eral) or to examine how the present results are modulated
by cultural differences in emotional display norms [53].
Finally, across-the-board positive associations with the
participantsfamiliarity with science fiction indicate a robust
effect of cultural conditioning on the acceptance of voice trans-
formation technology. As already remarked for brain implants
[28] or cognitive enhancement [30], exposure to futuristic
themes and ideas appears associated with less resistance to
technologies which challenge our conception of human
nature. The influence of science fiction themes is already
well studied as a source of inspiration for real-world techno-
logical innovation, e.g. in space [54] or nanotechnologies
[55], but it appears that it also plays a role in the reception of
new technology by the general public [56]. One consequence
is that the attitude towards voice transformations may
co-vary with cultural differences in attitudes towards new
technology (e.g. robots in Japan [57]).
One obvious limitation of this work is our focus on a
sample of predominantly young and educated western par-
ticipants (i.e. college students), which is representative
neither of the generation population in western countries
(as would survey pools constructed to match the composition
of a given adult population by gender, age, education and
ethnicity[27]), nor of the more global non-WEIRD popu-
lation [58]. Although research suggests that instruments
such as the MFQ are relatively stable across cultures [59],
there is an emerging corpus of work attempting to diversify
moral psychology research samples [60,61], and to conduct
cross-cultural comparisons with massive online method-
ologies [62]. Such initiatives will be particularly needed
when evaluating the acceptability of information technol-
ogies such as deep-fakes, which are spreading equally fast
in western and non-western countries [63].
The use of vignettes in experimental ethics approaches
also comes with several limitations. First, the intensity of
reactions elicited by the stories may be limited by the
immersion of the participant, or the vividness of their
imagination [30], and reading a vignette, especially one
describing an intense emotional situation, may not elicit reac-
tions as strong as in the corresponding real-life situations [64].
Here, we moderate these limitations by including an elabor-
ate cover story presenting the device as being considered
for commercialization by an actual voice technology com-
pany, and stating that participant responses will have
weight in future commercial decisions. Second, all of these
scenarios consider idealized transformations which are
assumed to be non-identifiable as fake, and properly recog-
nized as their intended emotion. As these technologies soon
become available, future work could consider measuring
reactions to more tangible situations (e.g. upon hearing
ones own voice modified by the device), studying situations
in which voice transformations are not recognized as genuine
behaviour (e.g. how comfortable am I to use a filter that may
sound robotic at times?), or combining the approach with
qualitative ethnographic methods documenting the appro-
priation of the device by potential users (e.g. how real call-
centre operators end up using a smile transformation) [65].
Finally, it should also be noted that, even though we designed
the present 12 vignettes to span a wide range of situations, it
remains an open question whether our conclusions generalize
to other types of vocal deep-fakes, and/or other types of
situations than those tested here.
Feelings and emotions are at the forefront of the political
behaviour of citizens and policy makers [66]. It will be essen-
tial for our societal future to clarify the determinants of moral
judgements about technologies able to customize and control
these behaviours, in order to guide norm-setting regarding
their applications.
4. Material and methods
(a) Participants
N= 303 participants (M= 25.7; female: 156) took part in an online
study, administered via a Qualtrics questionnaire (Qualtrics Inter-
national Inc., Seattle, WA). All were French residents, recruited by
the INSEAD-Sorbonne Université Behavioural Laboratory among
a population consisting mainly of university students. Of partici-
pants, 213 (70.3%) had completed at least a Bachelors degree, and
116 (38%) had at least a Masters degree. Participants were ran-
domised into one of two selfother conditions. For each
condition, participants were presented 12 vignettes of scenarios
assessing three within-participant factors tested for their possible
impact on moral acceptability (see §4c). For each vignette, partici-
pants answered three questions about their perceived moral
acceptability of the situation (see §4d), which creates a total of
36 answers for each participant.
royalsocietypublishing.org/journal/rstb Phil. Trans. R. Soc. B 377: 20210083
7
(b) Procedure
Participants were initially presented a cover story describing an
imaginary hardware device able to transform the emotional
quality of a voice in real-time, both in the users voice (for
others to hear) and in the users ear (i.e. transforming the
emotions of othersvoices). The device, named MyVoicePlus,
was presented as being considered for possible commercial
and/or clinical deployment by a French startup company. The
cover story included mock photographs of the device (consisting
of both an in-ear prosthesis and a larynx piece, disguised as
jewelry), as well as references to technical voice-transformation
literature (e.g. [67]) allegedly describing the algorithms imple-
mented in the device (see electronic supplementary material).
Participants were told that the startup was commissioning the
study to evaluate the societal acceptability of their technology
in various usage scenarios, and that their collective judgements
would condition the deployment of the technology.
After reading the cover story, participants were presented a
series of n= 12 short situational vignettes, each describing a poten-
tial application of the voice-transformation device (see following
section). There were two between-participant conditions, in
which participants either read vignettes that described the partici-
pant as the user of the device (condition self;N= 150), or vignettes
describing otherwise-identical situations in which the device was
applied to others and in which participants were in the position
of the users conversation partners (condition other;N= 153). In
each selfother condition, vignettes included a number of
within-participant conditions, which we describe below. For
each vignette, participants were asked to answer three questions
about how morally acceptable they think the situation was
(see §4d).
Finally, after completing the questions for all vignettes, par-
ticipants were asked to complete two standard questionnaires
measuring attitudes towards morality (Moral Foundations
Questionnaire MFQ; [31]) and toward technology and science
fiction (Science Fiction Hobbyism Scale; [28]). The study lasted
on average 30 min.
(c) Vignettes
We created n= 12 short text vignettes describing potential appli-
cations of the voice-transformation device in concrete daily life
situations. Vignettes varied among three situational factors,
which were encoded as within-participant variables to test for
their impact on the acceptability of the device:
(i) whether voice transformations are used to repair (e.g.
therapeutically) or enhance user capacities (condition
repair:n=6; enhance:n= 6). Examples of repair vignettes
included using the device to help a depressive patient
communicate with their close ones with a more enthusias-
tic tone of voice; examples of enhance situations included
using the same transformation to help a politician gather
more following. In repair vignettes, the device was
described as being prescribed to the user by a doctor; in
enhance vignettes, the device was recommended by a life
coach.
(ii) the kind of voice transformation operated by the device,
either reducing anger (n= 2; e.g. making angry customers
voices less taxing to attend to, for call-centre operators),
reducing anxiety (n= 4; e.g. helping a budding actor over-
come stage-fright) or enhancing smile (n= 6; e.g. helping a
waiter gather more tips from customers).
(iii) whether the voice transformation affects how the users
voice is heard by others (condition production:n= 4),
how the user hears other personsvoices (condition per-
ception:n= 4), or whether it is used in a situation where
the user hears their own manipulated voice (condition
feedback:n= 4). Examples of the feedback condition
include having a post-traumatic stress disorder (PTSD)
patient listen to their own voice made less anxious as
they retell their traumatic event [68].
All 12 vignettes were written in two matched versions, in
which the user of the device was either the participant (e.g. imagine
you are a depressive patient, and your doctor is advising you to use a
voice-transformation device; condition self :n= 12) or an unknown
other (e.g. A depressive patient; condition other: n= 12. Condition
selfother was randomly assigned between-participant; all other
conditions were varied within-participant, in random order.
All vignettes are available with English translation in the electronic
supplemental material.
(d) Measures
After reading each vignette, participants answered three ques-
tions about:
(i) how morally acceptable they think the situation is (Judge
how morally acceptable it is to use the MyVoicePlus device in
such a situation; FR: A quel point jugez-vous cette utilisation
du produit MyVoicePlusmoralement acceptable?)
(ii) how morally acceptable they think it would be for the
user to use the device covertly, i.e. to lie to their conversa-
tion partners that they are either talking to them, or
hearing them, with a modified voice, knowing that this
may improve the effectiveness of the device by up to
70%. (Judge how acceptable it is to lie to your entourage
about using the voice transformation, knowing that this
would improve the effectiveness of the device; FR: A quel
point jugez-vous acceptable le fait de cacher -votre entourage
lexistence de la transformation de voix, en sachant que cela
augmente considérablement lefficacité du dispositif?)
(iii) how morally acceptable they think it would be to hide the
true purpose of the device from its own user, i.e. that the
users themselves do not know that they are either talking,
or hearing others, with a modified voice. (Judge how accep-
table it is that the [doctor/coach] should lie to the user about the
voice transformation, knowing that this would improve the
effectiveness of the device; FR: A quel point jugez-vous
acceptable le fait que le médecin vous cache lexistence de
la transformation de voix, en sachant que cela augmente
considérablement son efficacité?)
Answers to all three questions were rated using a 9-point
Likert scale, anchored by 1 totally unacceptable and 9 totally
acceptable.
(e) Attitude questionnaires
In addition to providing moral judgements about the vignettes,
participants completed two questionnaires measuring their atti-
tudes toward morality (Moral Foundations Questionnaire MFQ;
[31]) and toward technology and science fiction (Science Fiction
Hobbyism Scale SFH; [28]).
The MFQ consists of 32 short questions (30 items + 2 foil items)
about how relevant various considerations are (e.g. whether or not
someone suffered emotionally) when deciding whether something is
right or wrong, rated from 1 (not at all relevant)to7(extremely rel-
evant), and how much the participant agrees with various moral
positions (e.g. compassion for those who are suffering is the most cru-
cial virtue; rated from 1 (strongly disagree)to7(strongly agree). In
accordance with typical MFQ analysis [31], we grouped and aver-
aged each participant responses along the five subscales of care
harm (6 items; e.g. whether or not someone suffered emotionally), fair-
nesscheating (5 items; e.g. whether or not some people were treated
differently from others), loyaltybetrayal (6 items; e.g. whether or
royalsocietypublishing.org/journal/rstb Phil. Trans. R. Soc. B 377: 20210083
8
not someone did something to betray their group), authoritysubver-
sion (5 items; e.g. whether or not an action caused chaos or disorder),
and puritydegradation (6 items; e.g. whether or not someone vio-
lated standards of purity and decency). None of the items were
reverse-coded. In this work, we used the back-translated French-
language version of the MFQ designed by Métayer & Pahlavan
[69]. A previous study [69] validated MFQ-French on a sample
of similar demographics as the present study and found it had
acceptable internal validity (N= 538 participants; care: Cronbach
α= 0.64; fairness: α= 0.67; loyalty: α= 0.65; authority: α= 0.73;
purity: α= 0.79) and one-month test-retest validity (N= 40; care:
r¼0:53, 95% CI [0.26,0.72]; fairness: r¼0:66, ½0:43, 0:80; loyalty:
r¼0:66, ½0:44, 0:81; authority: r¼0:75, [0.57, 0.86], purity:
r¼0:88, ½0:78, 0:94; all ps,0:01). The fit to a 5-factor structure,
while significantly better than an alternative 3-factor model, was
comparably poorer (N= 538; Comparative Fit Index, CFI: 0.82;
root mean squared error of approximation, RMSEA: 0.065), a
known issue common to the American version and discussed else-
where [70,71].
In accordance with recommendations of Métayer & Pahlavan
[69] and compared to the American version, two MFQ items (fair-
ness: I think its morally wrong that rich children inherit a lot of money
while poor children inherit nothing; authority: Men and women each
have different roles to play in society) were removed from the French
translation to improve internal consistency. In our data (N= 303),
the internal validity of the 5 MFQ constructs was comparable to
the sample of Métayer & Pahlavan [69] for puritydegradation
(α= 0.74, [0.70, 0.78]), fairnesscheating (α= 0.65, [0.58, 0.70]), loy-
altybetrayal (α= 0.63, [0.56, 0.69]) and authoritysubversion (α=
0.68, [0.62, 0.73]), but was poor for the careharm construct (α=
0.55, 95% CI [0.45, 0.61]). Confirmatory factor analysis for the 5-
factor model was significant (GLS fit, χ
2
(340) = 575, p< 0.001), fit-
ting data adequately on some measures (RMSEA = 0.048) but
relatively poorly on others (CFI = 0.48). We opt to conform to
the recommendations of a more extensive validation (with a
sample size nearly twice as big as our current sample [69]) and
use the 5-factor model for our current analysis. However, the pre-
sent data adds evidence to the fact that, as already discussed
elsewhere [71], significant elements of the MFQ covariance
structure are not captured by this model.
The SFH scale [28] consists of 12 items and measures individ-
ualscultural exposure to futuristic technology and science
fiction themes (examples of items: I often think about what machines
are like in the future,I often spot science or technology related errors in
science fiction films, TV series, or books). All items are rated from 1
(strongly disagree)to7(strongly agree), with higher scores indi-
cating higher science fiction familiarity. None of the items were
reverse-coded. A previous study [29] validated the scale on
N= 172 participants and found it had good psychometric proper-
ties (all factor loadings > 0.57; Cronbachsα= 0.92). In this work,
we used our own, non-validated French-language translation of
the SFH. In our data (N= 303), the internal validity of the SF con-
struct was also good (α= 0.89, [0.888, 0.898]).
(f) Statistical analyses
There were two dependent variables (DVs) in the study, measur-
ing the acceptability of overt (DV1) and covert (DV2) use of voice
transformations. DV2 was constructed by recoding the two
questions about concealed use (lying to the user, and lying to
others) as a single DV measured in two conditions (who is
being lied to).
The studys vignettes spanned a number of situation character-
istics, each described as a combination of independent variables
(IVs). There was one between-participant IV (selfother), three
within-participant IVs for DV1 (repairenhance, smileanxiety
anger, productionperceptionfeedback) and an additional two
within-participant IVs for covert DV2 (lying to userother, and
lying to participantother). We analysed the effect of IVs on
both DVs using one-way, repeated-measures or mixed ANOVAs,
by averaging acceptability scores within-participant over the
vignettes corresponding to each condition tested.
In addition, there were six measures of individual character-
istics (MFQ: 5 constructs; SFH: 1 construct). We tested the
association of these individual characteristics with the studys
DVs by computing within-participant averages of acceptability
scores (one data point per participant) and multiple regression.
We tested for residualprediction heteroscedasticity with the
BreuschPagan test. In case of homoscedasticity, we used mul-
tiple ordinary least square (OLS) regression; in case of
heteroscedasticity, we used iterated re-weighted least-square
(IRLS) regression with Huber weighting and HC3 correction.
All analyses were conducted in Python (3.6.8), using the
pingouin (0.3.12) and statsmodels (0.12.2) packages.
Ethics. All participants were tested at the Sorbonne-INSEAD Center for
Behavioral Science. The experiment was approved by the Institut
Européen dAdministration des Affaires (INSEAD) IRB (Study
202058; Study of the moral attitudes and willingness towards
the use of a voice transformation device; decision of 18 June 2020).
All participants gave their informed consent for the study,
were debriefed after the study, and were compensated for their
participation at a standard rate.
Data accessibility. Experimental data and analysis code (open-source,
Python) are available as electronic supplementary material at
https://github.com/creamlab/deep-ethics.
Authorscontributions. N.G. and J.J.A. designed the study and analysed
data. N.G. and J.J.A. wrote the manuscript, with contributions
from G.V.
Competing interests. J.J.A. is scientific advisor for voice transformation
start-up Alta Voce SAS.
Funding. Study funded by European Research Council Starting Grant
CREAM 335634, Proof of concept grant ACTIVATE (875212),
Agence Nationale de la Recherche PRC grants REFLETS and
SEPIA, and Fondation Pour lAudition (FPA RD-2018-2).
Acknowledgements. The authors thank Gilles Degottex and Marco Liuni
(Alta Voce SAS) for comments on the design of the study and Pablo
Arias (Lund University/IRCAM) for help with data analysis.
References
1. Darwin C. 1872 The expression of the emotions in
man and animals, 1998 edn. Oxford, UK: Oxford
University Press.
2. Knapp ML, Hall JA, Horgan TG 2013 Nonverbal
communication in human interaction. Boston, MA:
Cengage Learning.
3. Bachorowski J-A, Owren MJ. 1995 Vocal expression
of emotion: acoustic properties of speech are
associated with emotional intensity and context.
Psychol. Sci. 6, 219224. (doi:10.1111/j.1467-9280.
1995.tb00596.x)
4. Jack RE, Sun W, Delis I, Garrod OGB, Schyns PG.
2016 Four not six: revealing culturally common
facial expressions of emotion. J. Exp. Psychol.
General 145, 708. (doi:10.1037/xge0000162)
5. Oosterhof NN, Todorov A. 2008 The functional basis
of face evaluation. Proc. Natl Acad. Sci. USA 105,
11 08711 092. (doi:10.1073/pnas.0805664105)
6. Ponsot E, Arias P, Aucouturier JJ. 2018 Uncovering
mental representations of smiled speech using
reverse correlation. J. Acoust. Soc. Am. 143,
EL19EL24. (doi:10.1121/1.5020989)
7. Bitti PER, Bonfiglioli L, Melani P, Caterina R, Garotti
P. 2014 Expression and communication of doubt/
uncertainty through facial expression. Ricerche
Pedagogia DidatticaJ. Theories Res. Edu. 9,
159177.
royalsocietypublishing.org/journal/rstb Phil. Trans. R. Soc. B 377: 20210083
9
8. Goupil L, Ponsot E, Richardson D, Reyes G,
Aucouturier J-J. 2021 Listenersperceptions of the
certainty and honesty of a speaker are associated
with a common prosodic signature. Nat. Commun.
12,117. (doi:10.1038/s41467-020-20649-4)
9. Jack RE, Garrod OGB, Yu H, Caldara R, Schyns PG.
2012 Facial expressions of emotion are not culturally
universal. Proc. Natl Acad. Sci. USA 109, 72417244.
(doi:10.1073/pnas.1200155109)
10. Safra L, Chevallier C, Grèzes J, Baumard N. 2020
Tracking historical changes in trustworthiness using
machine learning analyses of facial cues in
paintings. Nat. Commun. 11,17. (doi:10.1038/
s41467-020-18566-7)
11. Arias P, Rachman L, Liuni M, Aucouturier J-J. 2020
Beyond correlation: acoustic transformation methods
for the experimental study of emotional voice and
speech. Emotion Rev. 13,1224. (doi:10.1177/
1754073920934544)
12. Rachman L et al. 2017 David: an open-source
platform for real-time transformation of infra-
segmental emotional cues in running speech.
Behav. Res. Methods 50, 323343. (doi:10.3758/
s13428-017-0873-y)
13. Westerlund M. 2019 The emergence of deepfake
technology: a review. Technol. Innovation Manage.
Rev. 9,4053. (doi:10.22215/timreview/1282)
14. Pumarola A, Agudo A, Martinez AM, Sanfeliu A,
Moreno-Noguer F. 2018 Ganimation: anatomically-
aware facial animation from a single image.
In Proc. Eur. Conf. on Computer Vision (ECCV),
814 September, Munich, Germany.
pp. 818833.
15. Luo Z, Chen J, Takiguchi T, Ariki Y. 2017 Emotional
voice conversion using neural networks with
arbitrary scales f0 based on wavelet transform.
EURASIP J. Audio Speech Music Process. 2017,113.
(doi:10.1186/s13636-016-0099-4)
16. Bonnefon J-F, Shariff A, Rahwan I. 2016 The social
dilemma of autonomous vehicles. Science 352,
15731576. (doi:10.1126/science.aaf2654)
17. Porter S, Brinke LT. 2008 Reading between the lies:
identifying concealed and falsified emotions in
universal facial expressions. Psychol. Sci. 19,
508514. (doi:10.1111/j.1467-9280.2008.02116.x)
18. Tang C, Seal CR, Naumann SE. 2013 Emotional labor
strategies, customer cooperation and buying
decisions. J. Manage. Market. Res. 13,1.
19. Côté S, Hideg I, Van Kleef GA. 2013 The
consequences of faking anger in negotiations.
J. Exp. Soc. Psychol. 49, 453463. (doi:10.1016/j.
jesp.2012.12.015)
20. Boidron L, Boudenia K, Avena C, Boucheix J-M,
Aucouturier J-J. 2016 Emergency medical triage
decisions are swayed by computer-manipulated
cues of physical dominance in callers voice. Sci.
Rep. 6,17. (doi:10.1038/srep30219)
21. Lerner A. We built voice modulation to mask gender
in technical interviews. Heres what happened.
https://blog.interviewing.io/we-built-voice-
modulation-to-mask-gender-in-technical-
interviews-heres-what-happened/ (accessed 16
March 2021).
22. Nijboer F, Morin FO, Carmien SP, Koene RA, Leon E,
Hoffmann U. 2009 Affective brain-computer
interfaces: psychophysiological markers of emotion
in healthy persons and in persons with amyotrophic
lateral sclerosis. In 2009 3rd Int. Conf. on Affective
Computing and Intelligent Interaction and
Workshops, pp. 111. Piscataway, NJ: IEEE.
23. Bovens L. 2009 The ethics of nudge. In Preference
change (eds T Grüne-Yanoff, SO Hansson), pp. 207
219. New York, NY: Springer.
24. Rohrmann S, Bechtoldt MN, Hopp H, Hodapp V,
Zapf D. 2011 Psychophysiological effects of
emotional display rules and the moderating role of
trait anger in a simulated call center. Anxiety Stress
Coping 24, 421438. (doi:10.1080/10615806.2010.
530262)
25. Goupil L, Johansson P, Hall L, Aucouturier J-J. 2021
Vocal signals only impact speakersown emotions
when they are self-attributed. Conscious Cogn. 88,
103072. (doi:10.1016/j.concog.2020.103072)
26. Medaglia JD, Yaden DB, Helion C, Haslam M.
2019 Moral attitudes and willingness to enhance
and repair cognition with brain stimulation.
Brain Stimulat. 12,4453. (doi:10.1016/j.brs.2018.
09.014)
27. Elias JJ, Lacetera N, Macis M. 2019 Paying for
kidneys? A randomized survey and choice
experiment. Am. Econ. Rev. 109, 28552888.
(doi:10.1257/aer.20180568)
28. Laakasuo M, Drosinou M, Koverola M, Kunnari A,
Halonen J, Lehtonen N, Palomäki J. 2018 What
makes people approve or condemn mind upload
technology? Untangling the effects of sexual
disgust, purity and science fiction familiarity.
Palgrave Commun. 4,114. (doi:10.1057/s41599-
018-0124-6)
29. Koverola M, Drosinou M, Palomäki J, Halonen J,
Kunnari A, Repo M, Lehtonen N, Laakasuo M. 2020
Moral psychology of sex robots: an experimental
studyhow pathogen disgust is associated with
interhuman sex but not interandroid sex. Paladyn
J. Behav. Rob. 11, 233249. (doi:10.1515/pjbr-
2020-0012)
30. Koverola M, Kunnari A, Drosinou M, Palomäki J,
Hannikainen IR, Sundvall J, Laakasuo M. 2020 Non-
human superhumans: understanding moral
disapproval of neurotechnological enhancement.
PsyArXiv. (doi:10.31234/osf.io/qgz9c).
31. Graham J, Nosek BA, Haidt J, Iyer R, Koleva S, Ditto
PH. 2011 Mapping the moral domain. J. Pers. Soc.
Psychol. 101, 366. (doi:10.1037/a0021847)
32. Körner A, Deutsch R, Gawronski B. 2020 Using the
cni model to investigate individual differences in
moral dilemma judgments. Pers. Soc. Psychol. Bull.
46, 13921407. (doi:10.1177/0146167220907203)
33. Cabrera LY, Fitz NS, Reiner PB. 2015 Empirical
support for the moral salience of the therapy-
enhancement distinction in the debate over
cognitive, affective and social enhancement.
Neuroethics 8, 243256. (doi:10.1007/s12152-014-
9223-2)
34. Rini R. 2020 Deepfakes and the epistemic backstop.
Phil. Imprint 20,116.
35. Pearce D. 1995 Hedonistic imperative. http://
happymutations.com/ebooks/david-pearce-the-
hedonistic-imperative.pdf (accessed 26 March 2021).
36. Earp BD, Wudarczyk OA, Sandberg A, Savulescu J.
2013 If I could just stop loving you: anti-love
biotechnology and the ethics of a chemical breakup.
Am. J. Bioethics 13,317. (doi:10.1080/15265161.
2013.839752)
37. Goffette J 2006 Naissance de lanthropotechnie: de la
médecine au modelage de lhumain. Paris, France:
Librarie Philosophique J. Vrin.
38. Clarke AE, Shim JK, Mamo L, Fosket JR, Fishman JR.
2003 Biomedicalization: technoscientific
transformations of health, illness, and us biomedicine.
Am. Sociol. Rev. 68,161194. (doi:10.2307/1519765)
39. Davis D, Dooley M, Hook J, Choe E, McElroy-Heltzel
S. 2016 The purity/sanctity subscale of the moral
foundations questionnaire does not work similarly
for religious versus non-religious individuals.
Psychol. Religion Spirituality 9, 124. (doi:10.1037/
rel0000057)
40. Kass L 2003 Beyond therapy: biotechnology and the
pursuit of happiness. New York, NY: HarperCollins.
41. Persson I, Savulescu J. 2008 The perils of cognitive
enhancement and the urgent imperative to enhance
the moral character of humanity. J. Appl. Phil. 25,
162177. (doi:10.1111/j.1468-5930.2008.00410.x)
42. Sahakian BJ, Morein-Zamir S. 2011
Neuroethical issues in cognitive enhancement.
J. Psychopharmacol. 25, 197204. (doi:10.1177/
0269881109106926)
43. Kahane G. 2015 Sidetracked by trolleys: why
sacrificial moral dilemmas tell us little (or nothing)
about utilitarian judgment. Social Neurosci. 10,
551560. (doi:10.1080/17470919.2015.1023400)
44. Schermer M 2015 Ethics of pharmacological mood
enhancement. In Handbook of Neuroethics (eds J
Clausen, N Levy), pp. 11771190. New York, NY:
Springer.
45. Kramer PD, Brody EB. 1994 Listening to prozac: a
psychiatrist explores antidepressant drugs and the
remaking of the self. J. Nervous Mental Dis. 182,
362.
46. Michaud T 2020 Science fiction and innovation
design. Hoboken, NJ: John Wiley & Sons.
47. Greene JD, Sommerville RB, Nystrom LE, Darley JM,
Cohen JD. 2001 An fMRI investigation of
emotional engagement in moral judgment.
Science 293, 21052108. (doi:10.1126/science.
1062872)
48. Sidgwick H 1874 The methods of ethics. New York,
NY: The Macmillan Company.
49. Trippett D. 2017 Music and the transhuman ear:
ultrasonics, material bodies, and the limits of
sensation. Musical Q. 100, 199261. (doi:10.1093/
musqtl/gdy001)
50. Kahane G. 2011 Reasons to feel, reasons to take
pills. In Enhancing Human Capacities (eds J
Savulescu, R ter Meulen, G Kahane), pp. 166178.
Hoboken, NJ: Wiley-Blackwell.
51. Ben-Zeev A. 1997 Emotions and morality. J. Value
Inquiry 31, 195212. (doi:10.1023/
A:1004236823330)
royalsocietypublishing.org/journal/rstb Phil. Trans. R. Soc. B 377: 20210083
10
52. Theriault JE, Young L, Barrett LF. 2021
The sense of should: a biologically-based
framework for modeling social pressure.
Phys. Life Rev. 36, 100136. (doi:10.1016/j.plrev.
2020.01.004)
53. Matsumoto D, Yoo SH, Fontaine J. 2008 Mapping
expressive differences around the world: the
relationship between emotional display rules and
individualism versus collectivism. J. Cross-Cultural
Psychol. 39,5574. (doi:10.1177/00220221
07311854)
54. Raitt D, Gyger P, Woods A. 2001 Innovative
technologies from science fiction. ESA-BR-176.
Noordwijk, The Netherlands: European Space
Agency.
55. Le Blanc T. 2014 Nanotech ideas in science fiction
literature. [German original: Hessen-Nanotech series
vol. 24, Hessen, Germany: Ministry of Economics,
Energy, Transport and Regional Development]
56. Michael K, Abbas R, Roussos G, Scornavacca E,
Fosso-Wamba S. 2020 Dealing with technological
trajectories: where we have come from and where
we are going. IEEE Trans. Technol. Soc. 1,27.
(doi:10.1109/TTS.2020.2976425)
57. Haring KS, Mougenot C, Ono F, Watanabe K. 2014
Cultural differences in perception and attitude
towards robots. Int. J. Affect. Eng. 13, 149157.
(doi:10.5057/ijae.13.149)
58. Henrich J, Heine SJ, Norenzayan A. 2010 Most
people are not weird. Nature 466,2929. (doi:10.
1038/466029a)
59. Doğruyol B, Alper S, Yilmaz O. 2019 The five-factor
model of the moral foundations theory is stable
across weird and non-weird cultures. Pers. Individual
Differences 151, 109547. (doi:10.1016/j.paid.2019.
109547)
60. Atari M, Graham J, Dehghani M. 2020 Foundations
of morality in Iran. Evol. Hum. Behav. 41, 367384.
(doi:10.1016/j.evolhumbehav.2020.07.014)
61. Berniūnas R. 2020 Mongolian yos surtakhuun and
WEIRD morality.J. Cultural Cogn. Sci. 4,5971.
(doi:10.1007/s41809-019-00045-1)
62. Awad E, Dsouza S, Shariff A, Rahwan I, Bonnefon
J-F. 2020 Universals and variations in moral
decisions made in 42 countries by 70,000
participants. Proc. Natl Acad. Sci. USA 117,
23322337. (doi:10.1073/pnas.1911517117)
63. Wu F, Ma Y, Zhang Z. 2021 I found a more
attractive deepfaked self: the self-enhancement
effect in deepfake video exposure. Cyberpsychol.
Behav. Soc. Netw. 24, 173181. (doi:10.1089/cyber.
2020.0173)
64. Parkinson B, Manstead ASR. 1993 Making
sense of emotion in stories and social life. Cogn.
Emotion 7, 295323. (doi:10.1080/0269993
9308409191)
65. Rey B, Simoncini N, Triclot M. 2021 Les sciences
humaines et sociales en recherche technologique:
vers une démarche de conception fondée sur
lethnographie. In La démocratie technique en
travail (eds P Lamard, Y Lequin). Belfort, France:
Presses de lUTBM.
66. Tsakiris M. 2021 How should the political animals of
the 21st century feel?: Comment on The sense of
should: a biologically-based framework for
modelling social pressureby JE Theriault et al.
Phys. Life Rev. 36,7779. (doi:10.1016/j.plrev.2020.
06.008)
67. Khosravani S, Mahnan A, Yeh I-L, Aman JE,
Watson PJ, Zhang Y, Goding G, Konczak J. 2019
Laryngeal vibration as a non-invasive
neuromodulation therapy for spasmodic
dysphonia. Sci. Rep. 9,111. (doi:10.1038/s41598-
019-54396-4)
68. Aucouturier J-J, Johansson P, Hall L, Segnini R,
Mercadié L, Watanabe K. 2016 Covert digital
manipulation of vocal emotion alter speakers
emotional states in a congruent direction. Proc. Natl
Acad. Sci. USA 113, 948953. (doi:10.1073/pnas.
1506552113)
69. Métayer S, Pahlavan F. 2014 Validation of the moral
foundations questionnaire in French. Revue Int. de
Psychologie Sociale 27,79107. (doi:10.1037/
t60220-000)
70. Curry OS, Chesters MJ, Van Lissa CJ. 2019 Mapping
morality with a compass: testing the theory of
morality-as-cooperationwith a new questionnaire.
J. Res. Personal. 78, 106124. (doi:10.1016/j.jrp.
2018.10.008)
71. Zakharin M, Bates T. 2021 Remapping the
foundations of morality: well-fitting
structural model of the moral foundations
questionnaire. PsyArXiv. (doi:10.31234/osf.io/pksnt)
royalsocietypublishing.org/journal/rstb Phil. Trans. R. Soc. B 377: 20210083
11
... It should be noted that the ethical criterion is set aside here, the idea is to say what is technically possible. Moreover, it seems obvious that some of these applications are potentially undesirable, do not answer any need and are object of potential totalitarian drifts and anthropological mutations (Guerouaou et al., 2021). ...
... This perspective, as interesting as it is terrifying, deserves to be discussed a lot more broadly, gathering insights from a variety of disciplines such as machine learning, cognitive science and psychology. This is the purpose of the collaboration I started with clinical psychologist and PhD candidate Nadia GEROUAOU of the Perception and Sound Design (PDS) team at Ircam (Guerouaou et al., 2021). ...
... Deep neural networks Source: Gil et al., (2023) A pair of algorithms have been employed in combat with one another in order to generate fake data. Each is frequently referred to as a marker, while the other as a machine that generates electricity (Etienne, 2021;Guerouaou et al., 2022). The tool for discrimination has been charged in figuring out if the multimedia material generated through the program that creates it synthetic or legitimate. ...
Chapter
Full-text available
Innovative approaches and regulatory laws for fake technology are clarified in this study. This chapter contributors explores the regulatory and legal implications of one specific real-world deepfake case study. The writers note that individuals are initially frightened of these deep fakes, similar to how they have been with numerous other innovations in the past. Although deepfakes are anticipated to be applied more often in the years to come, the writers focus on a number of societal and constitutional concerns that Humans and government will have to contend effectively. The probable role of legislatures and digital broadcasting networks in countering deepfakes is further addressed in the study. Furthermore, the researchers of this study offer suggestions and ideas on how to mitigate and reduce the growth of deepfakes from a standpoint of regulation and one from a point of view of technology.
... This consideration is in line with a survey among the general public which finds that deepfake (voice) transformations are more accepted when they are used inside a therapeutic context. 26 Specific measures should be taken in both cases to protect the depicted person's privacy. First, the deepfake and any personal data used for its creation should be stored locally and securely. ...
Article
Full-text available
Deepfakes are hyper-realistic but fabricated videos created with the use of artificial intelligence. In the context of psychotherapy, the first studies on using deepfake technology are emerging, with potential applications including grief counselling and treatment for sexual violence-related trauma. This paper explores these applications from the perspective of medical ethics and health law. First, we question whether deepfake therapy can truly constitute good care. Important risks are dangerous situations or ‘triggers’ to the patient during data collection for the creation of a deepfake, and when deepfake therapy is started, there are risks of overattachment and blurring of reality, which can complicate the grieving process or alter perceptions of perpetrators. Therapists must mitigate these risks, but more research is needed to evaluate deepfake therapy’s efficacy before it can be used at all. Second, we address the implications for the person depicted in the deepfake. We describe how privacy and portrait law apply and argue that the legitimate interests of those receiving therapy should outweigh the interests of the depicted, as long as the therapy is an effective and ‘last resort’ treatment option, overseen by a therapist and the deepfakes are handled carefully. We suggest specific preventative measures that can be taken to protect the depicted person’s privacy. Finally, we call for qualitative research with patients and therapists to explore dependencies and other unintended consequences. In conclusion, while deepfake therapy holds promise, the competing interests and ethicolegal complexities demand careful consideration and further investigation alongside the development and implementation of this technology.
... Hence, someone should be responsible for AI processes, policies, and protocols. Somebody must be responsible for determining if the output and performance are as per the given framework [58] a [40-48, 50, 53- quality of life [68,70,74,83,92]; improving efficiency and productivity at work [50,53,62,65,83] by reducing errors [77], relieving the burden of professionals and/or increasing accuracy in decisions [47,55,90]; and facilitating and expanding access to safe and fair healthcare [42,53,54] through earlier diagnosis, increased screening and monitoring, and personalized prescriptions [47,90]. To foster human rights, allowing people to make informed decisions, the last say was up to the person themselves [42,43,46,55,64,67,73,76]. People should determine where and when to use automated functions and which functions to use [44,54], developing "job sharing" arrangements with machines and humans complementing and enriching each other [56,65,90]. ...
Article
Full-text available
This scoping review examines the research landscape about publics’ views on the ethical challenges of AI. To elucidate how the concerns voiced by the publics are translated within the research domain, this study scrutinizes 64 publications sourced from PubMed® and Web of Science™. The central inquiry revolves around discerning the motivations, stakeholders, and ethical quandaries that emerge in research on this topic. The analysis reveals that innovation and legitimation stand out as the primary impetuses for engaging the public in deliberations concerning the ethical dilemmas associated with AI technologies. Supplementary motives are rooted in educational endeavors, democratization initiatives, and inspirational pursuits, whereas politicization emerges as a comparatively infrequent incentive. The study participants predominantly comprise the general public and professional groups, followed by AI system developers, industry and business managers, students, scholars, consumers, and policymakers. The ethical dimensions most commonly explored in the literature encompass human agency and oversight, followed by issues centered on privacy and data governance. Conversely, topics related to diversity, nondiscrimination, fairness, societal and environmental well-being, technical robustness, safety, transparency, and accountability receive comparatively less attention. This paper delineates the concrete operationalization of calls for public involvement in AI governance within the research sphere. It underscores the intricate interplay between ethical concerns, public involvement, and societal structures, including political and economic agendas, which serve to bolster technical proficiency and affirm the legitimacy of AI development in accordance with the institutional norms that underlie responsible research practices.
... To address more critical domains, such as psychotherapy or the service industry, AI applications should describe, understand or detect more emotions in real life. For instance, one can imagine training to express self-confidence before a job interview or dealing with anxiety in a therapeutic context [14]. Thus the FEM-based process needs to adapt to more various and fine-grained requirements. ...
Article
Full-text available
Recent deep-learning techniques have made it possible to manipulate facial expressions in digital photographs or videos, however, these techniques still lack fine and personalized ways to control their creation. Moreover, current technologies are highly dependent on large labeled databases, which limits the range and complexity of expressions that can be modeled. Thus, these technologies cannot deal with non-basic emotions. In this paper, we propose a novel interdisciplinary approach combining the Generative Adversarial Network (GAN) with a technique inspired by cognitive sciences, psychophysical reverse correlation. Reverse correlation is a data-driven method able to extract an observer’s ‘mental representation’ of what a given facial expression should look like. Our approach can generate 1) personalized facial expression prototypes, 2) of basic emotions, and non-basic emotions that are not available in existing databases, and 3) without the need for expertise. Personalized prototypes obtained with reverse correlation can then be applied to manipulate facial expressions. In addition, our system challenges the universality of facial expression prototypes by proposing the concepts of dominant and complementary action units to describe facial expression prototypes. The evaluations we conducted on a limited number of emotions validate the effectiveness of our proposed method. The code is available at https://github.com/yansen0508/Mental-Deep-Reverse-Engineering.
Article
: Deepfake technology is still on the rise, bringing serious risks to cybersecurity; detection and prevention are the next critical levels. This paper examines how AI and ML are critical components of the deepfake approach to combating cyber threats within the cybersecurity system. The first goals are to assess the efficiency of AI/ML in deepfake detection, to survey today's methods, and to discuss possible improvements regarding their future employment. In a quantitative method supplemented by perceptions and impressions from cybersecurity professionals, this cross-sectional study utilizes a mixed-method questionnaire. Principal findings show that pre-trained AI and ML enhance the accuracy and efficiency of deepfake identification with some limitations in identifying highly elaborate and real-time deepfake content. The study concludes that AI/ML-based solutions are needed to enhance anti-deepfake cybersecurity tools as they provide tailored and reliable frameworks for addressing the issue. They prove the importance of further implementing AI and ML technologies in improving the resistance of today's cybersecurity systems.
Article
We consider the pairing of audio chatbot technologies with voice-based deep fakes, that is, voice clones, examining the potential of this combination to induce consumer trust. We report on a set of controlled experiments based on the investment game, evaluating how voice cloning and chatbot disclosure jointly affect participants’ trust, reflected by their willingness to play with an autonomous, AI-enabled partner. We observe evidence that voice-based agents garner significantly greater trust from subjects when imbued with a clone of the subject’s voice. Recognizing that these technologies present not only opportunities but also the potential for misuse, we further consider the moderating impact of AI disclosure, a recent regulatory proposal advocated by some policymakers. We find no evidence that AI disclosure attenuates the trust-inducing effect of voice clones. Finally, we explore underlying mechanisms and contextual moderators for the trust-inducing effects, with an eye toward informing future efforts to manage and regulate voice-cloning applications. We find that a voice clone’s effects operate, at least in part, by inducing a perception of homophily and that the effects are increasing in the clarity and quality of generated audio. Implications of these results for consumers, policymakers, and society are discussed. This paper has been This paper was accepted by D. J. Wu for the Special Issue on the Human-Algorithm Connection. Funding: This work was supported by funding from the University of Wisconsin-Milwaukee Research Assistance Fund. Supplemental Material: The online appendix and data files are available at https://doi.org/10.1287/mnsc.2022.03316 .
Article
Social interaction research is lacking an experimental paradigm enabling researchers to make causal inferences in free social interactions. For instance, the expressive signals that causally modulate the emergence of romantic attraction during interactions remain unknown. To disentangle causality in the wealth of covarying factors that govern social interactions, we developed an open-source video-conference platform enabling researchers to covertly manipulate the social signals produced by participants during interactions. Using this platform, we performed a speed-dating experiment where we aligned or misaligned the facial smiles of participants in real time with face transformation algorithms. Even though participants remained totally unaware that their faces were being manipulated, aligning their smiles causally enhanced the romantic attraction they felt toward each other, compared to unaligned scenarios. Manipulations also influenced how participants synchronized and vocally reacted to each other. This paradigm causally manipulates the emergence of romantic attraction in free social interactions. Moreover, our methodology opens the possibility to perform causal inferences during free social interactions.
Article
Moral psychology was shaped around three categories of agents and patients: humans, other animals, and supernatural beings. Rapid progress in artificial intelligence has introduced a fourth category for our moral psychology to deal with: intelligent machines. Machines can perform as moral agents, making decisions that affect the outcomes of human patients or solving moral dilemmas without human supervision. Machines can be perceived as moral patients, whose outcomes can be affected by human decisions, with important consequences for human–machine cooperation. Machines can be moral proxies that human agents and patients send as their delegates to moral interactions or use as a disguise in these interactions. Here we review the experimental literature on machines as moral agents, moral patients, and moral proxies, with a focus on recent findings and the open questions that they suggest. Expected final online publication date for the Annual Review of Psychology, Volume 75 is January 2024. Please see http://www.annualreviews.org/page/journal/pubdates for revised estimates.
Article
Full-text available
With the introduction of deepfake technology, which enables digital face-swapping between two individuals, young women are no longer passive viewers of attractive celebrities, but are able to become part of the perfect images. This study used the ZAO app as the apparatus to investigate the impact of viewing the self-celebrity deepfaked videos (SCDV) on young female users' appearance self-evaluation (i.e., body image and state appearance self-esteem). A sample of 128 young women 18-31 years of age was randomly assigned to view either 10 SCDV or 10 purely celebrity videos (PCV). All videos were sourced from the ZAO app. Results showed that participants in the SCDV condition perceived themselves as more physically attractive, experienced greater satisfaction with their own facial features, and reported marginally higher state appearance self-esteem than those in the PCV condition, whereas body shape satisfaction did not differ between the conditions. In addition, SCDV exposure increased perceived overall physical attractiveness and facial features satisfaction, but did not impact body shape satisfaction, whereas PCV exposure decreased facial features satisfaction, but did not impact perceived overall physical attractiveness or body shape satisfaction. "Attractive possible self" (APS) perception positively mediated the effect of SCDV exposure on perceived overall physical attractiveness, facial features satisfaction, body shape satisfaction, and state appearance self-esteem, while state appearance comparison only negatively mediated the effect of SCDV exposure on facial features satisfaction. This study reveals the potential of deepfake technology as an intervention technique for body image disturbances.
Article
Full-text available
The success of human cooperation crucially depends on mechanisms enabling individuals to detect unreliability in their conspecifics. Yet, how such epistemic vigilance is achieved from naturalistic sensory inputs remains unclear. Here we show that listeners’ perceptions of the certainty and honesty of other speakers from their speech are based on a common prosodic signature. Using a data-driven method, we separately decode the prosodic features driving listeners’ perceptions of a speaker’s certainty and honesty across pitch, duration and loudness. We find that these two kinds of judgments rely on a common prosodic signature that is perceived independently from individuals’ conceptual knowledge and native language. Finally, we show that listeners extract this prosodic signature automatically, and that this impacts the way they memorize spoken words. These findings shed light on a unique auditory adaptation that enables human listeners to quickly detect and react to unreliability during linguistic interactions.
Article
Full-text available
Social trust is linked to a host of positive societal outcomes, including improved economic performance, lower crime rates and more inclusive institutions. Yet, the origins of trust remain elusive, partly because social trust is difficult to document in time. Building on recent advances in social cognition, we design an algorithm to automatically generate trustworthiness evaluations for the facial action units (smile, eye brows, etc.) of European portraits in large historical databases. Our results show that trustworthiness in portraits increased over the period 1500-2000 paralleling the decline of interpersonal violence and the rise of democratic values observed in Western Europe. Further analyses suggest that this rise of trustworthiness displays is associated with increased living standards.
Article
Full-text available
While acoustic analysis methods have become a commodity in voice emotion research, experiments that attempt not only to describe but to computationally manipulate expressive cues in emotional voice and speech have remained relatively rare. We give here a nontechnical overview of voice-transformation techniques from the audio signal-processing community that we believe are ripe for adoption in this context. We provide sound examples of what they can achieve, examples of experimental questions for which they can be used, and links to open-source implementations. We point at a number of methodological properties of these algorithms, such as being specific, parametric, exhaustive, and real-time, and describe the new possibilities that these open for the experimental study of the emotional voice.
Preprint
Moral foundations theory posits five moral foundations, however 5-factor models provide poor fit to the data. Here, in five studies, each with large samples (total N = 11,496), we construct and replicate a well-fitting model of the Moral Foundations Questionnaire (MFQ). In study 1 (N = 2,271) we tested previously theorised models, confirming none provide adequate fit. We then developed a well-fitting model of the MFQ. This required 7 domains resulted from splitting the loyalty/purity and loyalty/patriotism aspects of existing domains, adding a general moral factor as well as a moral “tilt” factor, corresponding to left-right political distinctions. The new model replicated in 4 independent samples including western and eastern countries (N = 9,225). These findings demonstrate the first well-fitting replicable model of the MFQ. They also highlight the importance of modelling measurement structure, improving measurement precision and revealing important additional factors and group-factor structure within the moral foundations questionnaire.
Article
Emotions are often accompanied by vocalizations whose acoustic features provide information about the physiological state of the speaker. Here, we ask if perceiving these affective signals in one’s own voice has an impact on one’s own emotional state, and if it is necessary to identify these signals as self-originated for the emotional effect to occur. Participants had to deliberate out loud about how they would feel in various familiar emotional scenarios, while we covertly manipulated their voices in order to make them sound happy or sad. Perceiving the artificial affective signals in their own voice altered participants’ judgements about how they would feel in these situations. Crucially, this effect disappeared when participants detected the vocal manipulation, either explicitly or implicitly. The original valence of the scenarios also modulated the vocal feedback effect. These results highlight the role of the exteroception of self-attributed affective signals in the emergence of emotional feelings.
Article
Most moral psychology research has been conducted in Western, Educated, Industrialized, Rich, and Democratic (WEIRD) societies. As such, moral judgment, as a psychological phenomenon, might be known to researchers only by its WEIRD manifestations. Here, we start with evaluating Moral Foundations Theory (MFT) using the Moral Foundations Questionnaire, and follow up by building a bottom-up model of moral values, in Iran, a non-WEIRD, Muslim-majority, understudied cultural setting. In six studies (N = 1945) we examine the structural validity of the Persian translation of the Moral Foundations Questionnaire, compare moral foundations between Iran and the US, conduct qualitative interviews regarding moral values, expand the nomological network of “Qeirat” as a culture-specific set of moral values, and investigate the pragmatic validity of “Qeirat” in the Iranian culture. Our findings suggest an additional moral foundation in Iran, above and beyond the five foundations identified by MFT. Specifically, qualitative studies highlighted the role of “Qeirat” values in the Iranian culture, which are comprised of guarding and protectiveness of female kin, romantic partners, broader family, and country. Significant cultural differences in moral values are argued in this work to follow from the psychological systems that, when brought to interact with particular socio-ecological environments, produce different moral structures. This evolutionarily-informed, cross-cultural, mixed-methods research sheds light on moral concerns and their cultural, demographic, and individual-difference correlates in Iran.
Article
Technological progress is widely recognized as having considerable impact on human societies, which have become more and more dependent on human-made tools that entail intricate scientific processes, considered technology [item 1) in the Related Works]. The adoption of such tools is borne out of the necessity of survival at first [item 2) in the Related Works], and subsequently as a means to improve access to food, shelter, and clothing; arguably things that we cannot live without. Nevertheless, the urgency of understanding the role of technology in human life has never been as vital as today. Indeed, one could say that technological progress is the key ingredient driving human civilization: the invention of ever more complex tools traces the history of human civilization and its development. Each generation of technology lays the foundation for the next, the invention of simple machines, such as the wheel, for example, enabled the invention of the wheelbarrow, which enabled in turn building larger structures. In parallel, the wheel enabled the construction of gears to transmit power from crankshaft to driveshaft and also the rotation of magnetic tapes permitting the capture of binary data leading to the spinning hard disk drives and their superior digital data storage capacity.