Research ProposalPDF Available

Investigating Objective and Subjective Social Exclusion in a Group Conversation Led by a Robot

Authors:

Abstract

This study investigates ostracism-based social exclusion in multi-person interactions with robots. To examine this phenomenon, we will conduct a laboratory study in which participants engage in a simulated job interview with the robot Ameca acting as the interviewer. The study compares objective exclusion (measured by the proportion of the robot's attention directed toward each participant) and subjective exclusion (participants' self-reported feelings of being ignored or excluded). We aim to identify the point at which objective exclusion leads to subjective feelings of exclusion and how this impact need fulfillment. After the interview, participants are allowed to stand somewhere else and are asked why they chose the same or a different standing position. Exploratory analyses will examine whether factors such as gender, height, or physical position (angle) relative to the robot influence the actual or assumed likelihood of being excluded.
1
Invesgang Objecve and Subjecve
Social Exclusion in a Group Conversaon
Led by a Robot
Authors:
Clarissa Sabrina Arlinghaus
Ashita Ashok
Karsten Berns
Günter W. Maier
Published: 03rd February 2025 DOI: hps://doi.org/10.17605/OSF.IO/HYB2S
Descripon:
(What is this study about?)
This study invesgates ostracism-based social exclusion in mul-person interacons with
robots. To examine this phenomenon, we will conduct a laboratory study in which
parcipants engage in a simulated job interview with the robot Ameca acng as the
interviewer.
The study compares objecve exclusion (measured by the proporon of the robot's aenon
directed toward each parcipant) and subjecve exclusion (parcipants' self-reported
feelings of being ignored or excluded). We aim to idenfy the point at which objecve
exclusion leads to subjecve feelings of exclusion and how this impact need fulllment. Aer
the interview, parcipants are allowed to stand somewhere else and are asked why they
chose the same or a dierent standing posion. Exploratory analyses will examine whether
factors such as gender, height, or physical posion (angle) relave to the robot inuence the
actual or assumed likelihood of being excluded.
Data collecon:
(Have any data been collected for this study already?)
No, no data have been collected for this study already.
2
Hypothesis:
(What’s the main queson being asked, or hypothesis being tested in this
study?)
Hypothesis:
1. Higher levels of objecve exclusion (measured by the robot's distribuon of aenon)
will lead to worse mood, and this relaonship will be mediated by subjecve
exclusion (parcipants’ self-reported feelings of being ignored or excluded). The more
people subjecvely feel excluded, the worse their mood is.
2. Higher levels of objecve exclusion (measured by the robot's distribuon of aenon)
will lead to lower need fulllment, and this relaonship will be mediated by subjecve
exclusion (parcipants’ self-reported feelings of being ignored or excluded). The more
people subjecvely feel excluded, the worse their need fulllment is.
Exploratory Quesons:
1. At what level of objecve exclusion do individuals begin to feel subjecvely excluded?
2. How do parcipants explain their experience of objecve exclusion by the robot?
3. Are there specic factors (e.g., age, gender, height, physical posion/angle) that
increase the likelihood of being excluded by the robot?
Dependent variables:
(Describe the key dependent variable(s) specifying how they will be measured.)
Objecve Exclusion:
Objecve exclusion will be measured by analyzing the robot's distribuon of aenon during
the simulated job interview. Precisely, we will calculate the proporon of interacons the
robot directed toward each parcipant. To improve interpretability, we will reverse the scale,
so that higher values represent higher levels of exclusion. The measure will be
operaonalized as follows:
𝑂𝑏𝑗𝑒𝑐𝑡𝑖𝑣𝑒 𝐸𝑥𝑐𝑙𝑢𝑠𝑖𝑜𝑛
= 1 𝑁𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑞𝑢𝑒𝑠𝑡𝑖𝑜𝑛𝑠 𝑎𝑛𝑑 𝑟𝑒𝑠𝑝𝑜𝑛𝑠𝑒𝑠 𝑑𝑖𝑟𝑒𝑐𝑡𝑒𝑑 𝑡𝑜 𝑎 𝑝𝑎𝑟𝑡𝑖𝑐𝑖𝑝𝑎𝑛𝑡
𝑇𝑜𝑡𝑎𝑙 𝑛𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑞𝑢𝑒𝑠𝑡𝑖𝑜𝑛𝑠 𝑎𝑛𝑑 𝑟𝑒𝑠𝑝𝑜𝑛𝑠𝑒𝑠 𝑡𝑜 𝑎𝑙𝑙 𝑝𝑎𝑟𝑡𝑖𝑐𝑖𝑝𝑎𝑛𝑡𝑠
Subjecve Exclusion:
Subjecve exclusion will be assessed using two self-report items from the manipulaon
check by Williams (2009):
For each queson, please select the number that best represents the feelings you were
experiencing during the interview.
"I was ignored."
"I was excluded."
Parcipants will rate these items on a 5-point Likert scale (1 = Strongly disagree, 5 = Strongly
agree) to capture their perceived exclusion during the interacon with the robot.
We adjusted the instrucon to reect parcipants' feelings during the job interview rather
than during a game.
3
Change in Mood:
Mood will be assessed using a scale from Williams (2009) that covers four posive aects
and four negave aects:
For each queson, please select the number that best represents the feelings you were
experiencing during the interview.
"Good"
"Bad"
"Friendly"
"Unfriendly"
"Angry"
"Pleasant"
"Happy"
"Sad"
Parcipants will rate these items on a 5-point Likert scale (1 = Not at all, 5 = Extremely)
before and aer their interacon with the robot. The dierence between the two measured
values is included in further analyses as a change in mood.
We adjusted the instrucon to reect parcipants' feelings during the job interview rather
than during a game.
Need Fulllment:
Need fulllment will be measured using a scale from Rudert and Greifeneder (2016). This
scale assesses parcipants' sasfacon with fundamental psychological needs following the
interacon. Four Items will be rated on a 9-point Likert scale to measure overall Need
Fulllment:
Please indicate how you felt during the interview. During the interview, I felt…
"invisible ……… recognized"
"devalued ……… valued"
"rejected ……… accepted"
"powerless ……… powerful"
We adjusted the instrucon to reect parcipants' feelings during the job interview rather
than during a game.
Standing posion:
Aer the interview, the parcipants are allowed to choose a new posion where they would
prefer to stand during a new round of interviews with the robot. We ask them "Where do
you want to stand for a second round of interviews?" and they can choose one out of ve
standing posions that are marked on the ground. We have also documented their standing
posion from the rst round so we can see if they want to change their posion of if they
decide to remain standing in the same posion.
Aer their decision, we ask parcipants the following two open-ended quesons to
understand their moves beer:
"Why did you choose this posion?"
"What do you think has a greater inuence on the conversaon with the robot: your
standing posion or the content of your answers?"
These responses will be analyzed qualitavely with a large language model using the LLM-
Assisted Inducve Categorizaon (LAIC) method (Arlinghaus et al., 2024).
4
Feedback:
At the end, parcipants have the chance to provide feedback though an open-ended
queson:
"Is there anything else you would like to tell me about your job interview experiences
or how they can be improved?"
These responses will be analyzed qualitavely with a large language model using the LLM-
Assisted Inducve Categorizaon (LAIC) method (Arlinghaus et al., 2024).
Condions:
(How many and which condions will parcipants be assigned to?)
There are no predened experimental groups in this study. All parcipants will engage in a
simulated group job interview with the humanoid robot Ameca, which operates in its
standard mode without any behavioral manipulaon.
During the interview, Ameca will naturally select which parcipants to ask quesons and
respond to. This interacon will be video recorded to accurately capture and later analyze
the robot's distribuon of aenon among parcipants.
Parcipants will not be assigned to specic condions. Instead, their experience of objecve
exclusion will be quaned based on how frequently Ameca directed quesons and
responses to them during the interacon.
All parcipants stand during the interview so that height dierences are more visible.
Analyses:
(Specify exactly which analyses you will conduct to examine the main
queson/hypothesis.)
1. Mediaon Analysis:
To test whether objecve exclusion indirectly aects change in mood or need
fulllment through subjecve exclusion, two simple mediaon analyses will be conducted.
Predictor: Objecve Exclusion
Mediator: Subjecve Exclusion
Outcome: Change in Mood or Need Fulllment
A (bootstrapped) mediaon analysis (e.g., with PROCESS Model 4) will be used to test the
indirect eect. 95% condence intervals will determine signicance.
2. Threshold Analysis (Objecve to Subjecve Exclusion):
To idenfy the point at which objecve exclusion leads to subjecve exclusion, a piecewise
(segmented) regression analysis will be conducted.
Predictor: Objecve Exclusion
Outcome: Subjecve Exclusion
This analysis will detect whether a threshold eect exists. If the relaonship is linear, a
standard regression will be used. For non-linear paerns, a piecewise regression will idenfy
crical turning points. Alternavely, a Johnson-Neyman analysis may explore the exact point
where the relaonship becomes signicant.
5
3. Qualitave Analysis of Open Responses:
Parcipants’ open-ended responses will be analyzed qualitavely. The LLM-Assisted Inducve
Categorizaon (LAIC) method will be used so that a large language model will inducvely
generate categories and assign parcipant statements accordingly (Arlinghaus et al., 2024).
4. Exploratory Analysis of Addional Factors:
To explore whether individual characteriscs inuence the likelihood of being excluded, the
following exploratory analyses will be conducted:
Predictors: Age, gender, internaonal student, language prociency, height, and
physical posion (angle relave to the robot)
Outcome: Objecve Exclusion
A mulple regression analysis will be performed to determine if these factors predict how
oen parcipants were engaged by the robot. If the data structure allows, interacon
terms may be included to examine potenal moderang eects.
5. Assumpon Checks:
Normality, linearity, homoscedascity, and mulcollinearity of variables will be
checked.
If assumpons are violated, non-parametric tests or data transformaons will be
applied.
Outliers and Exclusions:
(Describe exactly how outliers will be dened and handled, and your precise
rule(s) for excluding observaons.)
Outliers:
For connuous variables, outliers will be idened using the 3 × IQR rule (values
falling three mes the interquarle range below the rst quarle or above the third
quarle).
If extreme outliers (+- 3 x IQR) are detected, analyses will be conducted with and
without these cases to assess their impact on the results. Outliers will only be
excluded if they substanally distort the ndings.
Exclusion Criteria:
Parcipants will be excluded from the analysis if any of the following condions apply:
Technical Issues: Signicant technical problems during the interacon (e.g., recording
failures, robot malfuncons) that prevent accurate measurement of objecve
exclusion.
Non-compliance: Parcipants who do not follow the task instrucons (e.g., refusing
to engage with the robot or disrupng the interview).
Early Terminaon: Parcipants who do not complete the experiment and drop out
before the session is nished.
Incomplete Data: Parcipants who fail to complete the post-interacon
quesonnaire assessing subjecve exclusion and need fulllment.
All exclusions and the nal sample size will be transparently reported.
6
Sample Size:
(How many observaons will be collected or what will determine the sample
size?)
We plan to recruit approximately 35 parcipants for this study to account for potenal
exclusions due to technical issues or non-compliance.
This sample size was determined using a G*Power 3.1 (Faul et al., 2009) analysis for a linear
mulple regression (xed model, R² deviaon from zero) with:
A large eect size of = 0.35 (based on Hartgerink et al., 2015)
An α level of 0.05
A desired power of 0.80
Two predictors in the regression model.
The analysis indicated a required sample size of 31 parcipants. To ensure sucient power
and account for possible exclusions, we aim to recruit approximately 35 parcipants.
Recruitment will connue unl this number is reached or unl praccal constraints (e.g.,
me, resources) limit further data collecon. Any deviaons from the planned sample size
will be transparently reported.
Other:
(Anything else you would like to pre-register? E.g., data exclusions, variables
collected for exploratory purposes, unusual analyses planned)
The robot-mediated job interview scenario was inspired by prior research (Kumazaki et al.,
2017; Nørskov et al., 2020; Zafar et al., 2021). Even though this is a simulaon, we are
following the high-quality standards that apply to real interviews. We simulate a structured
job interview with pre-prepared queson since structured interviews are superior to
unstructured interviews (Levashina et al., 2014; Sacke et al., 2022). Apart from a short
round of introducons, which contains biographical elements, the job interview consists of
situaonal quesons and quesons about past behavior, as these have a parcularly high
validity (Taylor, 2002). The eight behavioral or situaonal quesons are presented randomly
and stem from the Career Development Services of the North Central State College (2016):
"Tell me about a challenge or conict you’ve faced at work (or university), and how
you dealt with it."
"How do you deal with pressure or stressful situaons?"
"Tell me about a me you failed. What did you learn from that experience?"
"How do you handle working with people who annoy you?"
"Tell me about your proudest achievement. What are you most proud of?"
"If I were your supervisor and asked you to do something that you disagreed with,
what would you do?"
"What if you were given an assignment that was too dicult for you? How would you
resolve the issue? Has this happened?"
"Describe how you would handle a situaon if you were required to nish mulple
tasks by the end of the day, and there was no conceivable way that you could nish
them."
7
We will use the robot Ameca. The Ameca hardware specicaons are the following (cf.
Roboc Research Lab, 2022):
Height 1870 mm
Weight 49 kgs
Binocular eye mounted cameras
Binaural ear mounted microphones
Speaker on chest
51 degrees of freedom
Eye tracking based on face detecon
External ZED2 camera mounted on chest
External microphone
References:
Arlinghaus, C. S., Wul, C., & Maier, G. W. (2024). Inducve Coding with ChatGPT – An
evaluaon of dierent GPT models clustering qualitave Data into categories. OSF Preprints.
hps://doi.org/10.31219/osf.io/gpnye
Career Development Services of North Central State College (2016). Common Interview
Quesons – Pracce List.
hps://ncstatecollege.edu/documents/StudentSrvcs/CareerSvcs/Interview%20Quesons_Lis
t.pdf
Faul, F., Erdfelder, E., Buchner, A., & Lang, A.-G. (2009). Stascal power analyses using
G*Power 3.1: Tests for correlaon and regression analyses. Behavior Research Methods,
41(4), 1149–1160. hps://doi.org/10.3758/BRM.41.4.1149
Hartgerink, C. H. J., van Beest, I., Wicherts, J. M., & Williams, K. D. (2015). The ordinal eects
of ostracism: A meta-analysis of 120 Cyberball studies. PloS ONE, e0127002.
hps://doi.org/10.1371/journal.pone.0127002
Kumazaki, H., Warren, Z., Corbe, B. A., Yoshikawa, Y., Matsumoto, Y., Higashida, H., Yuhi, T.,
Ikeda, T., Ishiguro, H. & Kikuchi, M. (2017). Android Robot-Mediated Mock Job Interview
Sessions for Young Adults with Ausm Spectrum Disorder: A Pilot Study. Froners in
Psychiatry, 8, 169. hps://doi.org/10.3389/fpsyt.2017.00169
Levashina, J., Hartwell, C. J., Morgeson, F. P., & Campion, M. A. (2014). The Structured
Employment Interview: Narrave and Quantave Review of Research Literature. Personnel
Psychology, 67(1), 241-293. hps://doi.org/10.1111/peps.12052
Nørskov, S., Damholt, M. F., Uhlhøi, J. P., Jensen, M. B., Ess, C., & Seibt, J. (2020). Applicant
Fairness Percepons of a Robot-Mediated Job Interview: A Video Vignee-Based
Experimental Survey. Froners in Robocs and AI, 7, 586263.
hps://doi.org/10.3389/frobt.2020.586263
8
Roboc Research Lab (2022). Emah. Empathic Mechanized Anthropomorphic Humanoid
System. hps://rrlab.cs.rptu.de/en/robots/ameca
Rudert, S. C. & Greifeneder, R. (2016). When it’s okay that I don’t play: Social norms and the
situated construal of social exclusion. Personality and Social Psychology Bullen, 42(7), 955-
969. hps://doi.org/10.1177/0146167216649606
Sacke, P. R., Zhang, C., Berry, C. M., & Lievens, F. (2022). Revising meta-analyc esmates
of validity in personnel selecon: Addressing systemac overcorrecon for restricon of
range. Journal of Applied Psychology, 107(11), 2040-2068.
hps://doi.org/10.1037/apl0000994
Taylor, P. J. (2002). Asking applicants what they would do versus what they did do: A meta-
analyc comparison of situaonal and past behaviour employment interview quesons.
Journal of Occupaonal and Organizaonal Psychology, 75(3), 277-294.
hps://doi.org/10.1348/096317902320369712
Williams, K. D. (2009). Ostracism: A temporal need-threat model. In M. P. Zanna (Ed.),
Advances in experimental social psychology (Vol. 41, pp. 275–314). Elsevier Academic Press.
hps://doi.org/10.1016/S0065-2601(08)00406-1
Zafar, Z., Ashok, A. and Berns, K. (2021). Personality Traits Assessment using P.A.D. Emoonal
Space in Human-robot Interacon. In Proceedings of the 16th Internaonal Joint Conference
on Computer Vision, Imaging and Computer Graphics Theory and Applicaons (VISIGRAPP
2021) (pp. 111-118). SciTePress. hps://doi.org/10.5220/0010161801110118
ResearchGate has not been able to resolve any citations for this publication.
Preprint
Full-text available
Qualitative data is invaluable, yet its analysis is very time-consuming. To prevent the loss of valuable information and to streamline the coding process for developing and assigning inductive categories, we introduce LLM-Assisted Inductive Categorization (LAIC), a novel method of categorizing text responses using a Large Language Model (LLM). In two pre-registered studies, we tested two Generative Pre-trained Transformer (GPT) models that are commonly used in ChatGPT (GPT-3.5 Turbo and GPT-4o) across three temperature settings (0, 0.5, 1) with 10 repetitions each (120 runs in total). Outputs were evaluated based on established qualitative research criteria (credibility, dependability, confirmability, transferability, transparency). Two human coders also generated inductive categories and assigned text responses accordingly for comparison. Our findings demonstrate that both GPT models are highly effective in developing and assigning inductive categories, even outperforming human coders in agreement rates. Overall, GPT-4o achieved the best results (e.g., better explanations and higher agreement) and is recommended for inductive category formation and assignment with a temperature setting of 0 and 10 repetitions. This approach saves significant time and resources while enhancing analysis quality. Instructions and Python scripts for applying our new coding technique are freely available under a CC-BY 4.0 International license: https://osf.io/h4dux/
Article
Full-text available
This paper systematically revisits prior meta-analytic conclusions about the criterion-related validity of personnel selection procedures, and particularly the effect of range restriction corrections on those validity estimates. Corrections for range restriction in meta-analyses of predictor-criterion relationships in personnel selection contexts typically involve the use of an artifact distribution. After outlining and critiquing five approaches that have commonly been used to create and apply range restriction artifact distributions, we conclude that each has significant issues that often result in substantial overcorrection and that therefore the validity of many selection procedures for predicting job performance has been substantially overestimated. Revisiting prior meta-analytic conclusions produces revised validity estimates. Key findings are that most of the same selection procedures that ranked high in prior summaries remain high in rank, but with mean validity estimates reduced by .10-.20 points. Structured interviews emerged as the top-ranked selection procedure. We also pair validity estimates with information about mean Black-White subgroup differences per selection procedure, providing information about validity-diversity tradeoffs. We conclude that our selection procedures remain useful, but selection predictor-criterion relationships are considerably lower than previously thought. (PsycInfo Database Record (c) 2021 APA, all rights reserved).
Article
Full-text available
It is well-established in the literature that biases (e. g., related to body size, ethnicity, race etc.) can occur during the employment interview and that applicants' fairness perceptions related to selection procedures can influence attitudes, intentions, and behaviors toward the recruiting organization. This study explores how social robotics may affect this situation. Using an online, video vignette-based experimental survey (n = 235), the study examines applicant fairness perceptions of two types of job interviews: a face-to-face and a robot-mediated interview. To reduce the risk of socially desirable responses, desensitize the topic, and detect any inconsistencies in the respondents' reactions to vignette scenarios, the study employs a first-person and a third-person perspective. In the robot-mediated interview, two teleoperated robots are used as fair proxies for the applicant and the interviewer, thus providing symmetrical visual anonymity unlike prior research that relied on asymmetrical anonymity, in which only one party was anonymized. This design is intended to eliminate visual cues that typically cause implicit biases and discrimination of applicants, but also to prevent biasing the interviewer's assessment through impression management tactics typically used by applicants. We hypothesize that fairness perception (i.e., procedural fairness and interactional fairness) and behavioral intentions (i.e., intentions of job acceptance, reapplication intentions, and recommendation intentions) will be higher in a robot-mediated job interview than in a face-to-face job interview, and that this effect will be stronger for introvert applicants. The study shows, contrary to our expectations, that the face-to-face interview is perceived as fairer, and that the applicant's personality (introvert vs. extravert) does not affect this perception. We discuss this finding and its implications, and address avenues for future research.
Article
Full-text available
The feasibility and preliminary efficacy of an android robot-mediated mock job interview training in terms of both bolstering self-confidence and reducing biological levels of stress in comparison to a psycho-educational approach human interview was assessed in a randomized study. Young adults (ages 18–25 years) with autism spectrum disorder (ASD) were randomized to participate either in a mock job interview training with our android robot system (n = 7) or a self-paced review of materials about job-interviewing skills (n = 8). Baseline and outcome measurements of self-reported performance/efficacy and salivary cortisol were obtained after a mock job interview with a human interviewer. After training sessions, individuals with ASD participating in the android robot-mediated sessions reported marginally improved self-confidence and demonstrated significantly lower levels of salivary cortisol as compared to the control condition. These results provide preliminary support for the feasibility and efficacy of android robot-mediated learning.
Article
Full-text available
Being excluded and ignored has been shown to threaten fundamental human needs and cause pain. Such reflexive reactions to social exclusion have been conceptualized as direct and unmoderated (temporal need threat model of ostracism). Here, we propose an extension and argue that reflexive reactions depend on how social exclusion situations are construed. If being excluded is understood as a violation of an inclusion norm, individuals will react with pain and threat. In contrast, if being excluded is consistent with the prevailing norm, the exclusion situation is interpreted as less threatening, and negative reflexive reactions to ostracism should be attenuated. Four studies empirically support this conceptual model. Studies 3 and 4 further show that to guide situated construal, the norm has to be endorsed by the individual. In both Studies 1 and 3, the effect of the norm is mediated by the objective situation’s subjective construal.
Article
Full-text available
We examined 120 Cyberball studies (N = 11,869) to determine the effect size of ostracism and conditions under which the effect may be reversed, eliminated, or small. Our analyses showed that (1) the average ostracism effect is large (d > |1.4|) and (2) generalizes across structural aspects (number of players, ostracism duration, number of tosses, type of needs scale), sampling aspects (gender, age, country), and types of dependent measure (interpersonal, intrapersonal, fundamental needs). Further, we test Williams's (2009) proposition that the immediate impact of ostracism is resistant to moderation, but that moderation is more likely to be observed in delayed measures. Our findings suggest that (3) both first and last measures are susceptible to moderation and (4) time passed since being ostracized does not predict effect sizes of the last measure. Thus, support for this proposition is tenuous and we suggest modifications to the temporal need-threat model of ostracism.
Article
Full-text available
Criterion‐related validities and inter‐rater reliabilities for structured employment interview studies using situational questions (e.g. “Assume that you were faced with the following situation … what would you do?”) were compared meta‐analytically with studies using past behaviour questions (e.g. “Can you think of a time when … what did you do?”). Validities and reliabilities were further analysed in terms of whether descriptively‐anchored rating scales were used to judge interviewees' answers, and validities for each question type were also assessed across three levels of job complexity. While both question formats yielded high validity estimates, studies using past behaviour questions, when used with discriptively anchored answer rating scales, yielded a substantially higher mean validity estimate than studies using the situational question format with descriptively‐anchored answer rating scales (.63 versus .47). Question type (situational versus past behaviour) was found to moderate interview validity, after controlling for whether studies used answer rating scales. No support was found for the hypothesis that situational questions are less valid for predicting job performance in high‐complexity jobs. Sample‐weighted mean inter‐rater reliabilities were similar for both situational and past behaviour questions, provided that descriptively‐anchored rating scales were used (.79 and .77, respectively), although they were slightly lower (.73) for past behaviour question studies lacking such rating scales.
Article
In the 20 years since frameworks of employment interview structure have been developed, a considerable body of empirical research has accumulated. We summarize and critically examine this literature by focusing on the 8 main topics that have been the focus of attention: (a) the definition of structure; (b) reducing bias through structure; (c) impression management in structured interviews; (d) measuring personality via structured interviews; (e) comparing situational versus past-behavior questions; (f) developing rating scales; (g) probing, follow-up, prompting, and elaboration on questions; and (h) reactions to structure. For each topic, we review and critique research and identify promising directions for future research. When possible, we augment the traditional narrative review with meta-analytic review and content analysis. We concluded that much is known about structured interviews, but there are still many unanswered questions. We provide 12 propositions and 19 research questions to stimulate further research on this important topic.