Conference PaperPDF Available

Differentiation in Personality Emotion Mappings From Self Reported Emotion and Automatically Classified Emotion

Authors:

Abstract and Figures

How does the relationship between personality traits and the basic emotions vary across the modalities of self-report and facial expression analysis? This article presents the results of an exploratory study that quantifies consistencies and differences in personality-emotion mappings across these two modalities. Twenty-four participants answered a personality questionnaire before watching twelve emotionally provocative videos. Participant's self-reported their emotional reactions per video, while their facial expressions were being recorded for automated emotional analysis. The results indicated that overall there was greater consistency than differences in personality-emotion mappings across the two modalities. The robustness of this relationship enables direct applications of emotional-state-to-personality-trait in academic and industrial domains.
Content may be subject to copyright.
Differentiation in Personality Emotion Mappings
From Self Reported Emotion and Automatically
Classified Emotion
Ryan Donovan1, Aoife Johnson1, and Ruairi O’Reilly1
Cork Institute of Technology, Ireland
brendan.donovan@mycit.ie
aoife.johnson@cit.ie
ruari.oreilly@cit.ie
Abstract. How does the relationship between personality traits and the
basic emotions vary across the modalities of self-report and facial expres-
sion analysis? This article presents the results of an exploratory study
that quantifies consistencies and differences in personality-emotion map-
pings across these two modalities. Twenty-four participants answered a
personality questionnaire before watching twelve emotionally provocative
videos. Participant’s self-reported their emotional reactions per video,
while their facial expressions were being recorded for automated emo-
tional analysis. The results indicated that overall there was greater con-
sistency than differences in personality-emotion mappings across the two
modalities. The robustness of this relationship enables direct applica-
tions of emotional-state-to-personality-trait in academic and industrial
domains.
Keywords: Personality ·Five-Factor Model ·Emotion ·Facial Expres-
sion Analysis ·Multimodal.
1 Introduction
Cognitive science has shown that the functioning of both personality and emotion
is necessary for positive well-being [1] [2]. Personality represents the idiosyncratic
way we perceive, feel, and interact with the world. It is a psychological system
that structures one’s desires, goals, and our methods for fulfilling our wants and
attaining our goals in the medium to long-term [3]. Emotions had historically
been considered as an impediment to clear thinking and action. However, emo-
tions guide our thoughts, behaviour, feelings, and motivation towards stimuli
that can satisfy our needs and desires in the present; emotions are signposts
towards our destination, not obstacles [4]. People with a malfunctioning person-
ality are aimless; people with malfunctioning emotions are chaotic.
There exists a wealth of research that has investigated the phenomena of
personality and emotion [5] [6]. However, only in a small proportion on such
research has focused on how these two phenomena interact with one another.
The results of such research showcase that there exists a quantifiable link between
2 R. Donovan, A. Johnson, R. O’Reilly
personality and emotion [7]. The existence of this relationship enables potential
applications in the domains of academia (e.g. understanding the affective nature
of our personality), clinical care (e.g. personality-based screenings for the onset
of affective disorders), occupational and marketing (e.g. personalised content and
services).
However, whilst the potential for applying personality-emotion mappings in
practical domains is exciting, there is a need to assess the robustness and gener-
alisability of this relationship. If emotions are a reliable indicator of personality
(or vice versa), then it needs to be demonstrated that this relationship is robust
across important factors [8]. Otherwise, the effectiveness of such applications will
be erratic and imprecise.
In terms of generalisability, an important factor is the modality of emotional
expression, e.g. subjective self-report and facial expressions. Research methodol-
ogy for assessing personality, emotion, and their relationship is largely reliant on
self-report-based questionnaires. From a researcher’s point of view, self-report-
based questionnaires are cost-effective and quick to administer. However, this
reliance on questionnaires can weaken the validity of results in cases where re-
peated self-report is required per participant (e.g. “retest artifact” effects [9])
and where it is an obstacle to recruiting participant due to the time taken to com-
plete such questionnaires. If the results from self-report-based questionnaires can
generalise across modalities, then this enables alternative and automatic meth-
ods of data capturing that requires minimal input from participants, even in
repeated sessions.
How consistent are emotional-state-to-personality trait mappings converge
across multiple modalities? This paper describes an exploratory research ex-
periment that investigated this question. The experiment investigated the level
of consistency of personality-to-emotion mappings across the modalities of self-
report and facial expressions. If it can be shown that there exists a large degree of
consistency with self-reported and automatically extracted emotions, then this
fortifies the concept of “state-to-trait” mappings as a usable tool [7].
Personality was conceptualised as personality traits, which are the typical
expressions of cognition, behaviour, affect, and motivation across time [10]. The
bedrock model of personality traits is the Five-Factor Model (FFM), which cat-
egorises personality across five broad traits: Openness to Experience, Consci-
entiousness, Extraversion, Agreeableness, and Neuroticism [11]. Emotion was
conceptualised as the basic emotions, which are a group of distinct emotions
that are reliably indicated by psychological, behavioural, and physiological sig-
nals. The basic emotions considered in this article are Anger, Disgust, Fear, Joy,
Sadness, and Surprise [12].
The research study utilised a machine learning-based detection platform,
Emotion Viewer [13], to automate the classification of facial expressions. Emotion
Viewer analyses real-time video to automatically classify emotions via facial
expressions from video recordings of participants taking part in the study. Facial
landmarks are defined as the detection and localization of certain key points
on a human face. The Emotion Viewer was trained on two data sets (Cohn-
Differentiation in Personality-Emotion Mappings 3
Kanade and Multimedia Understanding Group) to detect the basic emotions.
The Emotion Viewer was then tested with real-time video clips resulting in
an average accuracy of 88.76% per basic emotion [13]. The Emotion Viewer
provided a strong foundation for analysing differences between personality-to-
emotion mappings.
The paper is structured in the following manner. Section 2 describes the
experiment’s methodology in terms of the design, participant pool, materials
used, and procedure. Section 3 presents the results for descriptive and inferential
statistical analyses. Section 4 discusses the results in the context of the research
question. Section 5 concludes the paper and provides recommendations for future
research in this area.
2 Method
The experimental method adopted generated a range of interesting results wor-
thy of publication. Experimental results quantifying the links between person-
ality traits and basic emotions via self-reported emotions were published in [7].
In this paper, the work is extended to include analysis of automated emotion
categorisation. The work focuses on the differentiation in emotion-personality
mappings between self-reported emotions and automated emotion detection from
facial expressions.
2.1 Participants
The sample consisted of 24 participants (n = 24, females = 16, males = 8, M-
age = 31.96, SD = 13.73) from a subset of 38 participants. The age range of
the sample (range = 19-63) is larger than in most social science research, which
tends to primarily consist of 18-23 undergraduate students.
2.2 Materials
Emotions Scale - A Likert-scale was created for the purposes of this study.
Participants were asked to answer the question “While watching the previous
video, to what extent did you experience these following emotions? Please use
the following scale in your self-assessment: 1 = Not at All; 2 = A little bit; 3 =
Moderately; 4 = A lot; 5 = A great deal/an extreme amount”. This scale was
designed to capture the experience of each emotion, but not the level of valence
or arousal. Participants completed the scale 12 times each, consisting of 84
questions overall (with alpha = .95). Each individual emotion was assessed 12
times throughout the duration of the study: Anger (alpha = .80), Disgust
(alpha =.83), Fear (alpha = .85), Sadness (alpha = .76), Joy (alpha =.66),
Surprised (alpha =.88).
Personality Questionnaire - The Big Five Aspects Scale is a reliable
measure for Big Five traits and their associated sub-traits [13]. The scale is
composed of 100 questions. The results showed that scale had satisfactory
4 R. Donovan, A. Johnson, R. O’Reilly
test-retest reliability across each FFM trait: Openness to Experience (0.81),
Conscientiousness (.85), Extraversion (.85), Agreeableness (.74), Neuroticism
(0.85).
Technology - Participant’s reactions were recorded with the camera of a 2015
MacBook Pro, which included a 1080p web-camera. The software,
Screen-Cast-O-Matic, was used in order to both record the participant and the
MacBook screen simultaneously. Participants were also given a pair of Bose
QuietComfort Noise-Cancelling Headphones to wear whilst watching the
videos.
Emotional Stimuli - Previous research by independent groups have
demonstrated that video clips from movies and TV shows are a reliable
method for evoking emotional reactions [14] [15] [16].
Twelve video clips were used in this experimental design. Nine of those video
clips were chosen based on prior research demonstrating their ability to evoke
emotional reactions. Three new video clips were also selected - the rationale
being that the nine tested video clips provided a solid foundation to
empirically evaluate new stimuli. The overall list for the videos is presented in
Table 2, along with the clip’s length and the expected emotional reaction it
would elicit.
Emotion Viewer - The Emotion Viewer analyses real-time video to
automatically classify emotions using a machine learning supported support
vector machine. Figure 1 depicts a demonstration of the tool in operation. The
tool comes with three options: track face, track expressions, and to set the
voting count. The first two are required to enable facial expression analysis.
The voting count refers to the amount of consecutive classified emotions on a
frame-by-frame basis required to register a particular emotion. The voting
count for this research study was set to 10, which is the highest supported by
the Emotion Viewer. This means that for the Emotion Viewer to output an
emotion classification (for this example, Anger) it would require 10 consecutive
frames where it detected the emotion Anger. This represented a conservative
approach and was chosen to reduce the risk for Type 1 errors. More
information on the design of the Emotion Viewer is available from [13].
2.3 Procedure for Participants
Participants initially filled out a demographic information form online, which
included questions regarding their age, gender, nationality, and previous experi-
ence with psychometric tests. Participants were then invited to the laboratory
stage of the experiment, provided they fit inclusion criteria for the study (over
18 years of age and have not been diagnosed with an affective disorder).
In the first part of the laboratory stage, participants completed the Big Five
Aspects scale questionnaire, which on average took about 15 minutes to com-
plete. In the second part of the laboratory stage, the researcher set up the video
recording on the MacBook Pro. Given that participants varied in height, there
Differentiation in Personality-Emotion Mappings 5
Fig. 1. The Emotion Viewer Platform. Examples shown are of automated emotional
analysis of Joy (Happy), Anger, and Surprise
had to be a manual check to ensure that each participant’s face occupied the
camera frame. Once this was settled, the researcher would leave the room, and
the participants watched 12 video clips always in the same order (see Table 2).
Participants were alone when viewing videos to elicit a more natural reaction
and to prevent the participant from feeling self-conscious about their response.
Participants were given a pair of noise-cancelling headphones whilst watching the
videos, to help immerse themselves in the video. After each video, participants
completed a short emotion questionnaire, asking how they felt whilst watching
the video clip. Overall, the study took a participant 1 hour to complete.
3 Results
This section presents the key descriptive and inferential statistics from the study
in relation to the level of differentiation between mappings between personality
and emotions across modalities.
3.1 Descriptive Statistics
Emotions - Participants did not seem to experience a dominant emotion through-
out the study. The mean and standard deviations for both self-report and auto-
matically classified emotions are presented in Table 1.
Self-Reported Emotions per Video Clip - Table 2 presents the mean and
standard deviation for self-reported emotional reactions per video clip.
Personality Self-Reported Scores - The descriptive statistics for self-reported
personality scores are presented in Table 3.
6 R. Donovan, A. Johnson, R. O’Reilly
Emotions M SD Emotions M SD
A-Anger 1.96 1.15 SR-Anger 1.92 0.55
A-Fear 1.93 1.27 SR-Fear 2.00 0.70
A-Joy 1.56 1.07 SR-Joy 1.54 0.34
A-Sadness 2.06 1.24 SR-Sadness 2.22 0.41
A-Surprise 2.33 1.29 SR-Surprise 2.78 0.79
A-Disgust 1.42 0.87 SR-Disgust 2.50 0.51
Table 1. Descriptive Statistics for Automatically Classified Emotions
View Sequence: Source Anger Disgust Fear Joy Sadness Surprise
1: When Harry Met Sally 1.30 1.53 1.03 3.23 1.00 2.50
2: Peep Show 1.87 1.80 1.37 2.10 2.10 2.87
3: The Champ 1.77 1.53 1.63 1.17 4.13 1.80
4: The Lion King 2.77 2.33 2.27 1.20 4.20 1.77
5: Schindler List 3.93 4.10 1.93 1.03 3.30 2.40
6: American History X 3.67 4.33 3.00 1.03 3.57 3.13
7: Trainspotting 1.30 4.47 1.20 2.17 1.47 3.37
8: Pink Flamingo 1.67 4.67 1.50 1.17 1.33 3.87
9: Test Your Awareness 1.03 1.03 1.07 2.50 1.30 4.43
10: Sea of Love 1.07 1.00 2.30 1.00 1.00 2.43
11: Annabelle 1.37 1.63 4.10 1.10 1.57 3.20
12: The Blaire Witch Project 1.33 1.40 3.53 1.07 1.67 2.23
Table 2. Mean Emotion per Video Clip Across Sample.
Personality Traits M SD Personality Traits M SD
Openness to Experience 3.53 0.42 Agreeableness 1.03 0.29
>Openness 1.87 0.51 >Compassion 1.37 0.40
>Intellect 1.77 0.51 >Politeness 1.63 0.36
Conscientiousness 2.77 0.53 Neuroticism 2.27 0.45
>Industriousness 3.93 0.56 >Withdrawal 1.93 0.45
>Orderliness 3.67 0.60 >Volatiltiy 3.00 0.67
Extraversion 1.30 0.43
>Enthusiasm 1.67 0.47
>Assertiveness 1.03 0.60
Table 3. Descriptive Statistics for Personality Traits.
Differentiation in Personality-Emotion Mappings 7
3.2 Inferential Statistics
Relationship Between Personality Traits and Self-Reported Emotions
and Automatically Extracted Emotions Table 4 presents the correlated
mapping only for sub-sample (n = 24). These results are presented as a means for
comparison with the mapping between personality traits, self-reported emotions,
and automatically classified emotions.
Traits Anger Fear Joy Sadness Surprise Disgust
SR—A SR—A SR—A SR—A SR—A SR—A
Openness to Experience 0.01 -0.21 -0.04 0.05 0.20 0.29 0.11 -0.18 0.09 0.22 0.10 0.19
Openness 0.05 -0.06 -0.06 -0.03 0.17 0.28 0.10 -0.14 0.26 0.18 0.18 0.07
Intellect -0.03 -0.28 0.00 0.11 0.16 0.20 0.08 -0.16 -0.12 0.18 -0.02 0.25
Conscientiousness 0.16 0.58 -0.06 0.01 0.01 0.09 0.30 -0.30 -0.11 -0.19 0.10 0.10
Industriousness 0.11 0.51 -0.15 0.04 -0.15 -0.16 0.19 -0.12 -0.19 -0.14 -0.01 -0.01
Orderliness 0.16 0.50 0.03 -0.01 0.14 0.28 0.33 -0.39 -0.02 -0.19 0.18 0.17
Extraversion -0.13 0.15 -0.39 0.07 0.12 0.16 -0.09 -0.14 0.03 -0.12 -0.08 -0.18
Enthusiasm -0.13 0.17 -0.35 0.08 0.12 0.18 -0.19 -0.18 0.04 -0.07 -0.04 -0.20
Assertiveness -0.08 0.09 -0.29 0.04 0.08 0.09 0.03 -0.06 0.00 -0.13 -0.09 -0.10
Agreeableness -0.13 0.20 -0.16 -0.15 -0.35 -0.08 0.07 -0.34 0.15 0.26 0.13 0.05
Compassion -0.15 0.13 -0.29 -0.18 -0.32 0.04 0.01 -0.32 0.23 0.18 0.06 0.06
Politeness -0.04 0.18 0.07 -0.04 -0.20 -0.16 0.10 -0.19 -0.02 0.21 0.13 0.01
Neuroticism 0.22 0.26 0.14 -0.42 0.40 0.14 0.41 0.08 0.21 0.05 0.25 0.28
Withdrawal -0.01 0.04 0.24 -0.21 0.50 0.02 0.14 0.10 0.32 0.21 0.15 -0.03
Volatility 0.29 0.32 0.03 -0.42 0.20 0.18 0.45 0.04 0.07 -0.07 0.24 0.39
Table 4. Bi-Directional Mappings of Personality Traits and the Basic Emotions for
both Self-Reported (SR) and Automatically Classified Emotions (A).
Relationship between Personality Traits and Automatically Classified
Emotions Conscientiousness positively correlated with recognition of Anger
with a large effect size (df = 23, p0.01, r= 0.58). The sub-traits of Con-
scientiousness, Industriousness and Orderliness also positively correlated with
Anger with large effect sizes (df = 23, p0.05, r= 0.51; df = 23, p0.05,
r= 0.50). Orderliness negatively correlated with recognition of Sadness with a
medium effect size (df = 23, p= .059, r= -0.39).
Neuroticism negatively correlated with recognition of Fear with a medium-to-
large effect size (df = 23, p= .04, r= -0.41). The sub-trait Volatility negatively
correlated with recognition of Fear with a medium-to-large effect size (df = 23,
p= .04, r= -0.42). Volatility positively correlated with recognition of Disgust
with a medium effect size (df = 23, p= .05, r= 0.39)
Relationship between Self-Reported and Automatically Extracted Emo-
tions A Pearson correlation was conducted on the matrix of self-reported emo-
8 R. Donovan, A. Johnson, R. O’Reilly
Fig. 2. Overlap between automatic emotions and self-reported emotions. Green indi-
cates consistency, white no overlap, and red differences. The numbers within each cell
indicate the Pearson R effect size.
tions and A-emotions. A matrix depicting the level of consistency and differences
between self-reported emotions and automatically classified emotional expres-
sions,is presented in Figure 2 along with the effect size for each comparison. In
terms of consistency/differences for same-emotions across modalities, then 5 of
the 7 emotions are positively correlated: Fear (df = 23, p= 0.10), Sadness (df
= 23, p= 0.65), and Surprise (df = 23, p= 0.08). The emotion Disgust (df =
23, p= 0.72) and Anger (df = 23, p= 0.88) diverge with negative correlations
with their modality counterpart.F
4 Discussion
This study aimed to investigate whether the relationship between personality
traits and emotions is robust enough to generalise across the modalities of self-
report and facial expression analysis. The results showed that there existed more
consistency than differences in personality-emotion mappings. This section dis-
cusses both (i) which personality-emotion mappings showed greater consistency
than difference and (ii) which personality-emotion mappings showed greater dif-
ferences than consistency. Potential reasons for (i) and (ii) are also discussed.
Differentiation in Personality-Emotion Mappings 9
Additionally, an evaluation of the video-clips used is conducted and recommen-
dations for future researchers interested in employing video clips in emotion-
elicitation research is provided.
4.1 Consistency in Personality to Self-Reported and Automatically
Extracted Emotion Mappings
A consistent mapping is defined here as similarity in the direction of correlation
(e.g. both positive, both negative) and the existence of non-trivial effect sizes
(both must be greater than r=±.10) across both self-reported and automati-
cally classified emotions. Overall, 11 of the 15 personality traits showed greater
consistency than differences across both modalities.
Openness to Experience mapped consistently for Joy and Disgust. Openness
mapped consistently for Joy and Surprise; Intellect mapped consistently for Joy.
The largest consistency found across both emotion modalities for these traits
was the emotion Joy, which positively correlated with small-to-medium effect
sizes. This is consistent with prior research linking Openness to Experience with
positive emotion systems in the brain [2].
Conscientiousness mapped consistently for emotions Anger, Surprise, Dis-
gust. Industriousness mapped consistently for Anger and Surprise; Orderliness
mapped consistently for Anger and Disgust. The positive relationship found be-
tween Conscientiousness and Anger was the strongest across the entire data-set
for both self-reported and automatically classified emotions.
Extraversion mapped consistently for Joy. Enthusiasm mapped consistently
for Joy and Sadness; Assertiveness did not map consistently across any emotion.
The consistent relationship found between Extraversion and Joy is consistent
with past research findings showing that people high in Extraversion experience
more joy on a daily basis [17].
Agreeableness mapped consistently for Fear and Surprise. Compassion also
mapped consistently for Fear and Surprise; Politeness mapped consistently for
Joy. As detailed out in the next sub-section, there was a considerable amount of
differences between the modalities.
Neuroticism mapped consistently for Anger, Joy, and Disgust. Withdrawal
mapped consistently for Sadness and Surprise; Volatility mapped consistently
for Anger, Joy, and Disgust. The positive correlation found between Joy and the
personality traits associated with Neuroticism is inconsistent with past research.
Neuroticism has been consistently considered a personality trait that is nega-
tively valenced and it has been repeatedly linked to the experience of mental
illnesses [18].
4.2 Differences in Personality to Self-Reported and Automatically
Extracted Emotions Mappings
An inconsistent mapping (difference) is defined here as when the direction of cor-
relation for both modalities is the opposite of one another (e.g. one positive, one
10 R. Donovan, A. Johnson, R. O’Reilly
negative) and the existence of non-trivial effect sizes (both must be greater than
r=±.10). Overall, 2 of the 15 personality traits showed greater inconsistency
(difference) than consistency across both modalities.
Openness to Experience mapped differently for Sadness. Openness also mapped
differently for Sadness; Intellect mapped differently for Surprise and Disgust. For
both Openness to Experience and Openness, there were more consistencies than
differences across modalities. This was the opposite for the sub-trait Intellect.
Conscientiousness mapped differently for Sadness. Industriousness and Or-
derliness also mapped differently for Sadness. Overall, there were more consis-
tencies than differences for Conscientiousness and its two sub-traits.
Extraversion mapped differently for Anger. Enthusiasm also mapped differ-
ently for Anger; Assertiveness did not map differently across both modalities. It
should be stated although the differences between self-reported and automati-
cally classified Fear were still striking, it did not hit our threshold.
Agreeableness mapped differently for Anger. Compassion also mapped differ-
ently for Anger. Politeness mapped differently for Anger and Sadness. Anger was
the major source of divergence for mappings related to Agreeableness. Agree-
ableness, and in particular Politeness, has been linked to reduced experience of
Anger [7]. However, it is an open question as to whether people high in these
traits experience less anger or are less likely to report experiencing anger [19].
The self-reported emotion mappings support the first hypothesis, that people
high in these traits are less likely to experience Anger. However, the automated
emotional analysis suggests that these participants do experience Anger, but
are less likely to recognise or admit it. This latter finding is consistent with
the description of people high in politeness withholding their feelings to avoid
conflict [20].
Neuroticism mapped differently for Fear. Withdrawal also mapped differently
for Fear; Volatility did not map differently. There existed several mappings with
Neuroticism that were stronger in a given modality (e.g. Sadness), however,
these mappings did not have a large enough effect size in both modalities to
meaningfully compare and contrast.
4.3 Recommendations for Emotional Stimuli
Twelve video clips were used as emotional stimuli in this research study. Three
of those videos had not been previously used in prior experimental research. Two
of those new videos, Annabelle and Who Dunnit? Test Your Awareness, were
successfully elicited high levels of self-reported emotion. Annabelle evoked the
strongest mean experience of Fear across the participant group and Who Dunnit?
Test Your Awareness evoked the strongest mean experience of Surprise across the
participant group. Both videos are recommended as reliable emotional stimuli for
future research. However, the third inclusion, Peep Show, is not recommended for
future usage. The results of this study also cast suspicion on the reliability of the
video clip, Sea of Love, which has been used in prior research experiments. but
only a minimal amount of its targeted emotion, Surprise. Researchers requiring
Differentiation in Personality-Emotion Mappings 11
a video to reliably produce Surprise are recommended to use the video clip Who
Dunnit? Test Your Awareness instead.
5 Conclusion
Personality traits and basic emotions are significant predictors of human be-
haviour and the functioning of both phenomena is necessary for positive well-
being. Previous research has also found quantifiable links between these two
phenomena that can enable state-trait inferences, i.e. personality-emotion map-
pings [7]. However, empirical observations of personality and emotion and their
relationship are largely reliant on self-report-based methodology (i.e. question-
naires). This reliance on self-report limits the validity of empirical research in
direct (e.g. “retest artifact”) and indirect ways (e.g. makes participant recruit-
ment more difficult).
This paper described an empirical research study that tested the generalis-
ability of personality-emotion mappings across a self-reported based approach
and an automatically classified emotion-based approach via video. If personality-
emotion mappings were robust across both modalities, then this would be an
indicator that technological-based approaches can directly analyse personality,
emotion, and their relationship. Technological-based approaches can enable in-
telligent and automatic observations of these phenomena with minimal input
from participants.
The results showed greater consistency than differences in personality-emotion
mappings across the two modalities. For the 15 personality traits captured in
this study, the results showed that: (i) 11 personality traits showed greater con-
sistency than differences personality-emotion mappings across modalities; (ii)
2 traits showed an equal amount of consistency and differences in personality-
emotion mappings across modalities; (iii) 2 traits showed greater differences than
consistency in personality-emotion mappings across modalities.
The criteria for assessing a personality-emotion mapping was two-fold, a com-
parison between the direction of the correlations and the existence of non-trivial
effect sizes. However, there is scope for future research to assess the degree of
consistency/difference between the two modalities. Quantifying the degree of
consistency/difference would enable assessments about the strengths and weak-
nesses of self-report and automatically classified approaches.
Overall, the results are a promising indication that the relationship between
personality traits and basic emotion is robust enough to enable research and
commercial applications. However, given the small sample size of the research
study, caution is required in generalising the results. Future research in this
area that can (a) scale the number of participants and (b) incorporate other
modalities of emotional expression (e.g. language, speech) and (c) quantify the
degree of consistency/difference across modalities would represent a significant
next step in this research area.
12 R. Donovan, A. Johnson, R. O’Reilly
References
1. Antonio R Damasio. Descartes’ error. Random House, 2006.
2. Colin G DeYoung. Cybernetic big five theory. Journal of Research in Personality,
56:33–58, 2015.
3. Dan P McAdams. The art and science of personality development. Guilford Pub-
lications, 2015.
4. Lisa Feldman Barrett. How emotions are made: The secret life of the brain.
Houghton Mifflin Harcourt, 2017.
5. Daniel K Mroczek and Todd D Little. Handbook of personality development. Psy-
chology Press, 2014.
6. Lisa Feldman Barrett, Michael Lewis, and Jeannette M Haviland-Jones. Handbook
of emotions. Guilford Publications, 2016.
7. Ryan Donovan, Aoife Johnson, Aine deRoiste, and Ruairi O’Reilly. Quantifying
the Links between Personality Sub-Traits and the Basic Emotions. In Affective
Computing and Emotion Recognition, Lecture Notes in Computer Science, Cagliari,
July 2020. Springer.
8. Alessandro Vinciarelli and Gelareh Mohammadi. More personality in personality
computing. IEEE Transactions on Affective Computing, 5(3):297–300, 2014.
9. Chris J Durham, Lisa Ducommun McGrath, Gary M Burlingame, G Bruce
Schaalje, Michael J Lambert, and D Rob Davies. The effects of repeated admin-
istrations on self-report and parent-report scales. Journal of Psychoeducational
Assessment, 20(3):240–257, 2002.
10. Rebecca L Shiner and Colin G DeYoung. 6 CHAPTER The Structure of Temper-
ament and Personality Traits: A Developmental. The Oxford handbook of develop-
mental psychology, Vol. 2: Self and other, 2:113, 2013.
11. Oliver P John, Laura P Naumann, and Christopher J Soto. Paradigm shift to the
integrative big five trait taxonomy. Handbook of personality: Theory and research,
3(2):114–158, 2008.
12. Paul Ekman. Basic emotions. Handbook of cognition and emotion, pages 45–60,
1999.
13. Michael Healy, Ryan Donovan, Paul Walsh, and Huiru Zheng. A machine learning
emotion detection platform to support affective well being. pages 2694–2700. IEEE,
2018.
14. Crystal A. Gabert-Quillen, Ellen E. Bartolini, Benjamin T. Abravanel, and
Charles A. Sanislow. Ratings for emotion film clips. Behavior Research Meth-
ods, 47(3):773–787, September 2015.
15. T Lee Gilman, Razan Shaheen, K Maria Nylocks, Danielle Halachoff, Jessica Chap-
man, Jessica J Flynn, Lindsey M Matt, and Karin G Coifman. A film set for
the elicitation of emotion in research: A comprehensive catalog derived from four
decades of investigation. Behavior research methods, 49(6):2061–2082, 2017.
16. James J. Gross and Robert W. Levenson. Emotion elicitation using films. Cognition
and Emotion, 9(1):87–108, 1995.
17. David Watson and Lee Anna Clark. Extraversion and its positive emotional core.
In Handbook of personality psychology, pages 767–793. Elsevier, 1997.
18. DH Saklofske, IW Kelly, and BL Janzen. Neuroticism, depression, and depression
proneness. Personality and individual differences, 18(1):27–31, 1995.
19. William G Graziano and Ren´ee M Tobin. Agreeableness. 2009.
20. Colin G DeYoung and Timothy A Allen. Personality neuroscience. Handbook of
Personality Development, page 79, 2018.
ResearchGate has not been able to resolve any citations for this publication.
Conference Paper
Full-text available
This paper describes a new emotional detection system based on a video feed in real-time. It demonstrates how a bespoke machine learning support vector machine (SVM) can be utilized to provide quick and reliable classification. Features used in the study are 68-point facial landmarks. In a lab setting, the application has been trained to detect six different emotions by monitoring changes in facial expressions. Its utility as a basis for evaluating the emotional condition of people in situations using video and machine learning is discussed.
Article
Full-text available
The use of repeated measures to track the outcome of psychosocial interventions is becoming common practice in most organizations. One threat with this methodology is the retest artifact. This artifact is defined by a systematic change in scores for individuals who repeatedly complete an outcome instrument yet do not receive any formal treatment. The retest artifact can call into question the validity, of change associated with school-based treatment programs when this change is based upon repeated testing using outcome instruments. This study directly assessed the retest effect for both a youth and an adult psychosocial outcome measure, The frequency of retest administrations was systematically varied (weekly, biweekly, monthly, pre/post testing) over a 9-week period using a youth parent-report and an adult self-report Outcome Questionnaire. Results indicate a statistically significant improvement in functioning at the time of the second administration regardless of the lapse in time for the adult measure and a steady improvement in functioning for the parent-report youth measure. Implications of this finding for evaluating psychosocial interventions are discussed.
Article
Full-text available
Cybernetics, the study of goal-directed, adaptive systems, is the best framework for an integrative theory of personality. Cybernetic Big Five Theory attempts to provide a comprehensive, synthetic, and mechanistic explanatory model. Constructs that describe psychological individual differences are divided into personality traits, reflecting variation in the parameters of evolved cybernetic mechanisms, and characteristic adaptations, representing goals, interpretations, and strategies defined in relation to an individual’s particular life circumstances. The theory identifies mechanisms in which variation is responsible for traits in the top three levels of a hierarchical trait taxonomy based on the Big Five and describes the causal dynamics between traits and characteristic adaptations. Lastly, the theory links function and dysfunction in traits and characteristic adaptations to psychopathology and well-being.
Article
Personality is a psychological construct aimed at explaining the wide variety of human behaviours in terms of a few, stable and measurable individual characteristics. In this respect, any technology involving understanding, prediction and synthesis of human behaviour is likely to benefit from Personality Computing approaches, i.e. from technologies capable of dealing with human personality. The presentation introduces a conceptual model underlying the three main problems addressed in the field, namely Automatic Personality Recognition, Automatic Personality Perception and Automatic personality Synthesis. Furthermore, it shows a few examples of how personality traits can be inferred from different types of behavioural evidence.
Article
Film clips are widely utilized to elicit emotion in a variety of research studies. Normative ratings for scenes selected for these purposes support the idea that selected clips correspond to the intended target emotion, but studies reporting normative ratings are limited. Using an ethnically diverse sample of college undergraduates, selected clips were rated for intensity, discreteness, valence, and arousal. Variables hypothesized to affect the perception of stimuli (i.e., gender, race-ethnicity, and familiarity) were also examined. Our analyses generally indicated that males reacted strongly to positively valenced film clips, whereas females reacted more strongly to negatively valenced film clips. Caucasian participants tended to react more strongly to the film clips, and we found some variation by race-ethnicity across target emotions. Finally, familiarity with the films tended to produce higher ratings for positively valenced film clips, and lower ratings for negatively valenced film clips. These findings provide normative ratings for a useful set of film clips for the study of emotion, and they underscore factors to be considered in research that utilizes scenes from film for emotion elicitation.
Article
Researchers interested in emotion have long struggled with the problem of how to elicit emotional responses in the laboratory. In this article, we summarise five years of work to develop a set of films that reliably elicit each of eight emotional states (amusement, anger, contentment, disgust, fear, neutral, sadness, and surprise). After evaluating over 250 films, we showed selected film clips to an ethnically diverse sample of 494 English-speaking subjects. We then chose the two best films for each of the eight target emotions based on the intensity and discreteness of subjects' responses to each film. We found that our set of 16 films successfully elicited amusement, anger, contentment. disgust, sadness, surprise, a relatively neutral state, and, to a lesser extent, fear. We compare this set of films with another set recently described by Philippot (1993), and indicate that detailed instructions for creating our set of film stimuli will be provided on request.
Article
The relationship between EPQ dimensions, proneness to depression and severity of depressive symptoms was examined in a sample of university students. The results suggest that Neuroticism is related to mood variation reflecting current depressive symptoms and depression vulnerability within nonclinical groups.
Article
Publisher Summary This chapter focuses on the theoretical views of extraversion. Extraversion has been included as a higher-order factor in every major taxonomic scheme of personality traits that have been developed during the past 50 years. Extraversion consistently emerges as a higher-order disposition in taxonomic schemes of personality traits. Higher-order traits represent the most general level in the hierarchy of dispositions. This is the level at which personologists attempt to explain individual differences with the fewest possible and most broadly applicable dimensions. Extraversion is divided into two distinct subfactors: (1) it involves successful adaptation through satisfying interpersonal relationships, and (2) another that entails adaptation through dominance, mastery, and achievement. Theoretical conceptualizations of extraversion have gradually but, systematically evolved over the past 75 years. This evolution necessarily involves elements of both continuity and change. The continuity is largely reflected in the interpersonal components of the trait. Recent conceptualizations have also stressed the positive affective component of the trait. These models emphasize that extraverts are happy, enthusiastic, confident, active, and energetic. More fundamentally, it now appears that extraversion essentially taps individual differences in affectively rewarding performance. Compared to introverts, extraverts view themselves as more effectively and pleasurably engaged in various aspects of their lives.