ChapterPDF Available

The Past, Present, and Future of Research on Interviewer Effects

Authors:

Figures

Content may be subject to copyright.
University of Nebraska - Lincoln University of Nebraska - Lincoln
DigitalCommons@University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln
Sociology Department, Faculty Publications Sociology, Department of
2020
The Past, Present, and Future of Research on Interviewer Effects The Past, Present, and Future of Research on Interviewer Effects
Kristen M. Olson
Jolene Smyth
Jennifer Dykema
Allyson L. Holbrook
Frauke Kreuter
See next page for additional authors
Follow this and additional works at: https://digitalcommons.unl.edu/sociologyfacpub
Part of the Family, Life Course, and Society Commons, and the Social Psychology and Interaction
Commons
This Article is brought to you for free and open access by the Sociology, Department of at
DigitalCommons@University of Nebraska - Lincoln. It has been accepted for inclusion in Sociology Department,
Faculty Publications by an authorized administrator of DigitalCommons@University of Nebraska - Lincoln.
Authors Authors
Kristen M. Olson, Jolene Smyth, Jennifer Dykema, Allyson L. Holbrook, Frauke Kreuter, and Brady T. West
1
The Past, Present, and Future of
Research on Interviewer Eects
Kristen Olson,1 Jolene D. Smyth,1 Jennifer Dykema,2
Allyson L. Holbrook,3 Frauke Kreuter,4 & Brady T. West 5
1 University of Nebraska-Lincoln
2 University of Wisconsin–Madison
3 University o Illinois at Chicago
4 University of Maryland, University of Mannheim, & Institute for
Employment Research
5 University of Michigan
Contents
1 Introduction 1
2 Training, Managing, and Monitoring Interviewers 3

4 Interviewers and Nonresponse 7
5 Interviewer Pace and Behaviors 8

7 Closing Thoughts 12
Acknowledgments 15

1.1 Introduction
Interviewer-administered surveys are a primary method of collect-
ing information from populations across the United States and the

digitalcommons.unl.edu
Published in  ed. Kristen Olson,
Jolene D. Smyth, Jennifer Dykema, Allyson L. Holbrook, Frauke Kreuter, and Brady T. West
(2020), Boca Raton: CRC Press, pp. 3-15.
Copyright © 2020 Taylor & Francis Group, LLC. Used by permission.
(2020) 2
including large-scale government surveys that monitor populations
(e.g., the Current Population Survey), surveys used by the academic
community to understand what people think and do (e.g., the Gen-
eral Social Survey), and surveys designed to gauge public opinion
at a particular time point (e.g., the Gallup Daily Tracking Poll). In-
-
tude of ways, including creating lists of housing units for sampling,
persuading sampled units to participate, and administering survey
questions (Morton-Williams 1993). In an increasing number of sur-
veys, interviewers are also tasked with collecting blood, saliva, and
other biomeasures, and asking survey respondents for consent to
link survey data to administrative records (Sakshaug 2013). Inter-
-

have failed (e.g., the American Community Survey and the Agricul-
tural Resource Management Survey; de Leeuw 2005; Dillman, Smyth
and Christian 2014; Olson et al. 2019). In completing these varied

-
nard 2010; West and Blom 2017).
Errors introduced by interviewers can take the form of bias or
variance. Early research found that interviewers vary in how they

sample clusters in both face-to-face (Hansen, Hurwitz and Bershad

Mathiowetz and Cannell 1980). In particular, similar to a design ef-
fect for cluster samples, interviewers increase the variance of an
estimated mean as a function of their average workload (b) and the
intra-interviewer correlation (IIC) (the degree of within-interviewer
correlation in measurements): 1 + (b
1) IIC. IIC values range from
very small (0.001) to large (0.20) and larger (Elliott and West 2015;

with median values typically below 0.02. Given an IIC of 0.02 and
a workload of 50 interviews per interviewer, the variance of an es-
timated mean is almost doubled and standard errors are increased
by 40%. Thus, a fundamental goal of research on interviewers is
understanding what contributes to (and how to minimize) the IIC.
Even if the IIC is small, interviewer characteristics and behaviors
(2020) 3
can still bias responses. Interviewer sociodemographic characteris-
-

voices may be associated with responses to survey questions or indi-
cators of survey error (e.g., an indicator of a sampled unit responding

characteristics of interviewers, they typically bias estimates. In addi-
tion, research based on coding information about the interviewer-re-
-
reading questions, probing directively, and acting non-neutrally may
-

The goal of the 
in February 2019 was to convene an in-
ternational group of leading academic, government, and industry re-
searchers and practitioners to discuss methods, practical insights, and

-
ers on multiple error sources, (2) evaluate study design and estima-

agenda for future studies of interviewers. After two days of presenta-
tions and posters, workshop participants spent the third day discuss-
ing and identifying areas where more work is needed. This chapter
introduces an edited volume consisting of chapters written by work-


the survey process. Second, we situate the chapters from this volume
-
search that arose from the third day of focused discussion.
2 Training, Managing, and Monitoring Interviewers

on the resultant survey data. Although standardized interviews are
the gold standard, a strict implementation of standardization may not
-
toring, and feedback systems vary widely across survey organizations
(2020) 4
(Vitema and Maynard 2002). Standardized interviewing commonly in-
-
vided, following up inadequate answers nondirectively, and acknowl-
edging adequate answers (Fowler and Mangione 1990). Yet some of
the central tenants of standardized interviewing- including reading
questions verbatim - are inadequately operationalized in actual train-
ing materials. Additionally, survey practitioners often make decisions
about interviewer training with little to no empirical evidence regard-

standardized training philosophy or materials since Fowler and Man-

this important issue, updating basic training of General Interview-
ing Techniques for question administration based on decades of re-
search on interviewer-respondent interaction and survey practice.
Among other things, their update tackles thorny issues of how to rec-
-
edging the critical role of characteristics of survey questions in deter-
mining what makes an answer codable), how to maintain respondent
engagement, and what common conversational practices can be al-
lowed in the interview.
-
tions for survey operations, it was clear during the workshop that this
research has not been fully adopted by survey organizations. We have
few recent descriptions of how survey organizations select, train, and
monitor survey interviewers, especially for smaller survey organiza-
tions. Yet many operational concerns are researchable, and such re-
search could yield both theoretical and practical insights. Miller and



and what kind of practical training methods may address these con-
-

(Chapter 4) meta-analysis of interviewer training includes many stud-


-
toring interviewers, including how to identify successful interviewers
(2020) 5
from a pool of applicants, how the pool of interviewers has changed

the supervisor on the interviewer. Future research is also needed to
-
ment, to shaping interviewer perceptions about their tasks, what com-
bination of in-person and online methods are best for delivering con-


maintain.
Many have questioned the utility of standardization, particularly


Schober and Conrad 1997; Suchman and Jordan 1990). Interviewers
often break from standardization on these types of items and ques-
tionnaires. Based on qualitative in-depth interviews with survey in-
terviewers and their reactions to vignettes, Kaplan and Yu (Chapter 5)
-

standardization occur. Innovations in technology, increased computing

paradata have dramatically changed how we monitor and train inter-
viewers, particularly for in-person interviewing (e.g., Edwards, Mai-
tland and Connor 2017; Olson and Wagner 2015). Edwards, Sun, and


as detected via monitoring using computer-assisted recorded inter-
viewing (CARI). These two chapters highlight the need to train inter-

the value of reinforcing this training through real-time monitoring.
Throughout the workshop, participants echoed the need for more in-
formation on how survey organizations monitor and provide feedback

and telephone interviews.
Although in-depth interviews and CARI recordings provide unique

process, many organizations lack resources to conduct such studies
and instead use paradata and interviewer observations to evaluate

 
et al. (Chapter 7) use paradata on interview duration and indicators

-


indicators help to identify problems and where they fail. West et al.
(Chapter 8) use data from two large surveys to assess whether post-
survey interviewer observations about the survey process are associ-
ated with indicators of measurement error. The sets of observations
that are most important to collect for monitoring and evaluation pur-

the types of survey questions asked and observations collected across
studies. More work is needed to align these observations across sur-


3 Interviewer Eects Across Contexts and Modes

to conduct interviews in private, many interviews are conducted in

when reporting about sensitive topics (Mneimneh et al. 2015). In com-
munity-based studies where in-community interviewers are recruited
-



-
iors and attitudes in the in-person Saudi National Mental Health Sur-

-

community who is known to them in a community-based participatory
research study in an American Indian community. Combined, these
studies reveal the need for more research into interview privacy and

Additionally, the mode and/or device for the interview - in-person,
(2020) 7
landline phone, cellular phone, or audio computer-assisted self-in-
terviewing; and interviewer input into a desktop or laptop, tablet, or

-
brook, Smyth and Olson 2018). Notably, the mode or device for the
interaction changes the nature of the interaction between interview-
-
ferences in the interviewer-respondent interaction across telephone
-
ferences in 
message-based interviews. Conrad et al. (Chapter 11) replace human

the interviewer is virtual, but their voice is that of an audio-recorded
human. These chapters raise important questions about interview-

or otherwise) is an important feature of the surveyor where tasks are
critical for the interviewer to consider? Can virtual interviewers pro-
-
-

4 Interviewers and Nonresponse

-

Purdon 1997; Groves and Couper 1998) - due to both heuristic cues

their behaviors during the recruitment interaction (e.g., Couper and

nonresponse error variance (e.g., West, Kreuter and Jaenichen 2013;

-
amples of how tailoring is operationalized (e.g., Groves and Couper
1998). Research about tailoring generally relies on interviewer reports
with measures that vary across studies. To address this, Ackermann-
Piek, Korbmacher, and Krieger (Chapter 14) predict survey contact
(2020) 8


little replication in associations between the covariates and the non-
response outcomes across studies, they emphasize the importance of
real-time monitoring of interviewers.

15, Wescott discusses a case management model for a telephone sur-

about when to call cases. While interviewers report being more sat-

the model yields lower productivity than using a call scheduler. More
work is needed to understand how interviewer autonomy and insights
-
phone survey organization to increase interviewer engagement and
ultimately retain interviewers.
Survey interviews increasingly ask respondents to provide blood,
saliva, urine, or other biomeasures or for permission to link their sur-
vey data to administrative data (e.g., Jaszczak, Lundeen and Smith
-
deh, Cernat, and Sakshaug use nurse characteristics and paradata to

biomeasures for a general population survey. These stages—partici-
pating in the nurse visit, consenting to a blood sample, and obtaining
a blood sample—reveal substantial variation in the nonresponse out-
comes related to the nurses, and that the predictors of nonresponse
vary across the stages. This work and additional workshop discussion
suggest that we need more research on the antecedents and conse-
quences of interviewer variation in the ability to successfully collect

5 Interviewer Pace and Behaviors

al. (Chapter 17) test what interviewer and question characteristics

response latencies and indicators of respondent comprehension and

(2020) 9
time taken to administer and answer questions is associated with in-
terviewer, respondent, and question characteristics. Kelley (Chapter
-
erated from timing paradata can be used to identify question misread-
ings, testing three methods of setting thresholds. Olson and Smyth


behaviors over time.

—pace—their conceptualizations and operationalizations vary (e.g.,
interview duration, question duration, interviewer speaking time
or speed, response latencies, words per second, and questions per
minute; see also Chapter 21, in which Dahlhamer et al. use aver-
age number of seconds per question across a questionnaire). Other
-
ine interviewer and respondent behaviors (e.g., Fowler 2011; On-

how behavior coding is implemented (from live interviews, record-
ings, or transcripts, and focused on respondents, interviewers, or
both) and the variety of operational and analytical decisions often
made in such research. Such decisions include assigning codes at the
question level or the conversational turn level (Olson and Parkhurst
2013); coding the entire questionnaire or a subset of items (e.g., see

Fowler 2011; Mathiowetz and Cannell 1980) or in combination (e.g.,

-
-
ally or sequentially; and dealing with overlapping speech, interrup-
tions, and other normal conversational events. Even simple issues
of how many interviews to code and (acceptable) reliability of the
codes vary across studies.
Despite this heterogeneity, there are common patterns observed
-
ence (Holbrook et al., Chapter 17; Garbarski et al., Chapter 18; Olson
and Smyth, Chapter 20; Olson and Bilgen 2011; Olson and Peytchev
2007). Second, question characteristics drive this phenomenon. Fig-
ure 1 shows the percent of variance in question duration attributable
(2020) 10
to interviewers, respondents, and questions across multiple studies
with question-level duration as the outcome. Some models are esti-
mated as three-level multilevel models, so the residual variance ac-
counts for question- level variation. Other models are estimated as
-
itly estimated as part of the model. Across these four studies, inter-
viewer- and respondent- level variation is small and the question-
level (or residual that incorporates questions) variation is large.
There is a consistent tendency for longer questions and questions
written at higher grade levels to have longer durations (e.g., Gar-
barski et al. Chapter 18; Couper and Kreuter 2013; Olson and Smyth
2015). Yet other question characteristics, including question place-
ment, question sensitivity, type of question (attitude, factual knowl-
-
structions, parentheticals, battery items, and measures from survey

and Gallhofer 2007), are inconsistently parameterized or have incon-
-

the mechanisms for these connections still requires more work. One
clear direction from the workshop for future research was to identify
a common set of dependent variables related to pace and behaviors, a
Figure 1 Variance in question duration due to interviewers, respondents, and
questions.
(2020) 11
common set of question characteristics, and a common set of respon-
dent and interviewer characteristics, parameterize these identically,
-

-

of research is also of interest.
6 Estimating Interviewer Eects

-
-


-
els to study how properties of survey questions themselves may af-


-
panelli 1998, 1999).
-

-

over 100 outcomes in the National Health Interview Survey. They
-




those who administer questions at a slower pace. Similarly, Loosveldt
-
-

countries, comparing two analytic approaches to estimating these ef-
fects. Similar to Dahlhamer et al., Loosveldt and Wuyts meta-analyze
(2020) 12


-
viewer variance across independent groups of interviewers. In Chapter
23, West reviews design decisions needed to compare interviewer vari-
ance components across two groups of interviewers, using a unique
study in Germany comparing standardized and conversational ap-
-
viewers to conditions to power analyses to analytic decisions, this


practical insights into these complicated designs.
7 Closing Thoughts

chapters in this volume. Even though interviewers have been cen-
tral to data gathered to understand society since the beginning of
survey research, we know surprisingly little about them. Namely,
-
ganizations, how interviewers perceive their job, and how we can
-
portant issues - interviewers can do a lot of harm to survey data if
they try to do so (Chapter 7) but also inadvertently introduce error
into data even when they are attempting to follow their training. Un-
derstanding the challenges and constraints interviewers face will fa-
cilitate understanding the mechanisms underlying interviewer-re-
lated survey errors.
Most of the chapters in this volume use observational data. Obser-
vational research is constrained by the data that a research team has
available. Through the workshop and studies featured in this volume,
-
tionalizations and create more consistency across operational imple-

-
-

(2020) 13


-

nomenclature (e.g., Conrad and Schober 2000), as well as the con-

interviewer-related variance across countries (Loosveldt and Wuyts,
Chapter 22) reveals the need for understanding interviewing practice,

contains many studies conducted outside of the United States (e.g.,
-

future research.
As researchers and practitioners, we call on survey organizations
to make information on interviewers available in public-use analytic
Table 1 contains a list of concepts that workshop partici-

studies. At the bare minimum, an anonymized interviewer ID variable

-

-
viewer level, these include measures of work productivity and qual-
ity, measures of how thinly spread the interviewer is across multiple

from the interviewers themselves about their job, as well as mea-
-
formation about the amount, type, and content of interviewer train-
ing as well as details about the supervision and monitoring practices
and feedback provided to interviewers from supervisors or monitors
-
mation about the content of general interviewing techniques training


interviewers. Some of this information could be included in method-
ology reports. Although many organizations may consider information
about training, supervision, and monitoring to be proprietary, more
complete disclosure is certainly needed to understand these rarely
(2020) 14
Table 1 A List of Recommended Information About Interviewers to Include in Public-
use Data Files and Methodology Reports
Interviewer ID
Interviewer characteristics
Demographics (gender, race, age, educational status)
Personality assessments
Experience (within survey, within organization, across organizations)
Certication test scores
Number of other jobs they are currently working
Performance metrics and problems on other studies
Ever been red for other performance issues on other studies
Number of hours work on other studies
Interviewer expectation and attitudinal measures
Ratings of the importance of a completed interview
Ratings of the importance of obtaining high-quality data
Description of how sensitive and dicult questions are approached
Ratings of other variables related to job satisfaction and engagement
Interviewing training variables
Content of training and training methods (e.g., round-robin, type and content
of at-home study, hours/days of training by topic)
Participation in any specialized training (e.g., for refusal avoidance)
Number of trainings attended
Interview process variables
Sanitized post-survey observations
Number of other projects interviewers involved in current project worked on
during current project
Whether and how interviewers were matched to respondents
Measures of eld performance, including ICC information for variables
Adaptive design features and implementation
Interviewers’ notes about survey questions
Organization and study-specic characteristics
Information on recruiting, hiring, training, and attrition
Monitoring, feedback, and falsication detection activities
Description of the supervisory structure of the interviewer corps, e.g., number
of interviewers per supervisor
(2020) 15
studied, yet critically important, survey practices.
Finally, dissemination and integration of research on interviewers
into survey practice is hard. Many new practices may face cultural op-
position at organizations simply because it is not the way that work
has been done in the past. Clients, survey project managers, and su-
pervisors are often risk averse to trying something new, even if inte-
grating recommendations based on research can improve survey prac-

read the latest research, and survey methodologists are often discon-
-

We suggest a few ways forward. First, many professional associa-
-
ical researchers. Carving out space for these two disparate groups to
discuss mutually interesting problems can facilitate research translat-
ing into practice and practice informing research questions. For sev-
eral years the Midwest Association for Public Opinion Research has
sponsored a workgroup in which researchers and practitioners talk
about important topics on interviewers. In 2019, the discussion fo-
cused on interviewer training, for instance. Second, methodologists
who work at organizations with survey shops or who contract for re-
search with survey shops are well positioned for translation of re-
-

related to practice could inspire some changes in survey practice that
would improve data quality. Furthermore, recognizing contributions
other than simply research articles - for instance, data being available
in the public domain, availability of code, availability of interviewer
-
vey practices that worked or did not work would ease translation of
work from one research team to another. Although these are hard, we
think they are worthwhile future pursuits.
Acknowledgments The workshop and this edited volume were sup-
ported by the National Science Foundation (SES-1758834), the Charles
Cannell Fund in Survey Methodology of the Survey Research Center at
the University of Michigan, and the Rensis Likert Fund for Research
 
in Survey Methodology at the University of Michigan. Any opinions,


National Science Foundation.
References
Campanelli, P., P. Sturgis, and S. Purdon. 1997. 
London:
Survey Methods Centre at SCPR.

interactions while using a mobile computer-assisted personal interview
device. 18:335-351.
Conrad, F. G. and M. F. Schober. 2000. Clarifying question meaning in a household
telephone survey. 
Couper, M. P. and R. M. Groves. 2002. Introductory interactions in telephone
surveys and nonresponse. In 
ed. D. W. Maynard, H. Houtkoop-

Wiley & Sons.

time and data quality. de

21:233-255.
Dijkstra, W. 1983. How interviewer variance can bias the results of research on

Dillman, D. A., J. D. Smyth, and L. M. Christian. 2014. 
Hoboken, NJ: John Wiley &
Sons.
Edwards, B., A. Maitland, and S. Connor. 2017. Measurement error in survey
operations management. In ed. P. P. Biemer, E.
De Leeuw, S. Eckman, B. Edwards, F. Kreuter, L. E. Lyberg, N. C. Tucker, and B.
T. West, 253-277. Hoboken, NJ: John Wiley & Sons.

variance that is unaccounted for in single-stage health surveys. 

Fowler, F. J. 2011. Coding the behavior of interviewers and respondents to
evaluate survey questions. In 
ed. J. Madans, K. Miller, A. Maitland, and G. Willis, 7-21.
Hoboken, NJ: John Wiley & Sons.
Fowler, F. J. and T. W. Mangione. 1990. 
Newbury Park, CA: Sage Publications.

(2020) 17
conversational practices, and rapport: Responsiveness and engagement in the
standardized survey interview. 

understanding aid (QUAID): A web facility that helps survey methodologists
improve the comprehensibility of questions. 
Groves, R. M. and M. Couper. 1998. 
New York: John Wiley & Sons, Inc.



Telephone interviewer voice characteristics and the survey participation
decision. In ed. J. M. Lepkowski, C.
Tucker, J. M. Brick, E. D. de Leeuw, L. Japec, P. J. Lavrakas, M. W. Link, and R.
L. Sangster, 385-400. Hoboken, NJ: John Wiley & Sons.

in censuses and surveys. 



9:no pp.


Jaszczak, A., K. Lundeen, and S. Smith. 2009. Using nonmedically trained
interviewers to collect biomeasures in a national in-home survey. 



Mathiowetz, N. A. and C. F. Cannell. 1980. Coding interviewer behavior as a
method of evaluating performance. 


2002. 
New York: John Wiley & Sons, Inc.


for social conformity on reporting sensitive information. 

Morton-Williams, J. 1993. Cambridge: University Press.




of interviewers in survey non-response. 


(2020) 18

Olson, K. and B. Parkhurst. 2013. Collecting paradata for measurement error
evaluations. In 
ed. F. Kreuter, 43--72. Hoboken, NJ: John Wiley & Sons.

pace and interviewer attitudes. 

interviewers on response time. 

Olson, K., J. D. Smyth, R. Horwitz, S. Keeter, V. Lesser, S. Marken, N. Mathiowetz,

Gurtekin, C. Turakhia, and J. Wagner. 2019. 
to Oakbrook Terrace, IL: American
Association for Public Opinion Research.

interviewer travel behavior. 

interviews. 
Sakshaug, J. W. 2013. Using paradata to study response to within-survey requests.
In ed.
F. Kreuter, 171-190. Hoboken, NJ: John Wiley and Sons.
Sakshaug, J. W., M. P. Couper, M. B. Ofstedal, and D. R. Weir. 2012. Linking survey
and administrative records: Mechanisms of consent. &

Saris, W. E. and I. N. Gallhofer. 2007. 
Hoboken, NJ: John Wiley and Sons.

in the standardized interview. In ed. P. Biemer,

John Wiley & Sons, Inc.

interviewing. In ed. P. V. Marsden and J. D.
Wright, 437-470. Bingley, UK: Emerald Group Publishing.

model of the call for survey participation: Actions and reactions in the survey
recruitment call. 

and response: Predicting participation from the call opening. 

Schober, M. F. and F. G. Conrad. 1997. Does conversational interviewing reduce
survey measurement error? 
Suchman, L. and B. Jordan. 1990. Interactional troubles in face-to-face survey
interviews. 
Timbrook, J., J. Smyth, and K. Olson. 2018. Why do mobile interviews take longer?
(2020) 19
A behavior coding perspective. 
Viterna, J. and D. W. Maynard. 2002. How uniform is standardization?
Variation within and across survey research centers regarding protocols for
interviewing. In 
ed. D. W. Maynard, H. Houtkoop-Steenstra, N. C.


synthesis. 

face surveys: A function of sampling, measurement error, or nonresponse?

West, B. T. and K. Olson. 2010. How much of interviewer variance is really
nonresponse error variance? 
... The various tasks for which interviewers may be responsible (i.e., generating sampling frames, making contact, gaining cooperation, asking survey questions, conducting measurements, maintaining motivation, and recording answers and measurements) result in various errors affecting coverage, nonresponse, measurement, and processing errors (Schaeffer et al., 2010;West and Blom 2017). Variability in response distributions may arise due to variability in interviewer behavior (West and Blom, 2017), and thus understanding origins of the degree of within-interviewer correlations in measurements (IICs) and finding ways to minimize IICs is the 'fundamental goal of research on interviewers' (Olson et al., 2020). For a wide variety of state-of-the-art research on interviewers and the effects that they can have on survey data from the TSE perspective were refer the reader to the recently published volume Interviewer Effects from a Total Survey Error Perspective (Olson et al., 2020). ...
... Variability in response distributions may arise due to variability in interviewer behavior (West and Blom, 2017), and thus understanding origins of the degree of within-interviewer correlations in measurements (IICs) and finding ways to minimize IICs is the 'fundamental goal of research on interviewers' (Olson et al., 2020). For a wide variety of state-of-the-art research on interviewers and the effects that they can have on survey data from the TSE perspective were refer the reader to the recently published volume Interviewer Effects from a Total Survey Error Perspective (Olson et al., 2020). ...
... Participants were fully informed about their rights, dignity, and privacy and were provided with a consent form and information sheet prior to their involvement in the study. Data was collected through face-to-face semi-structured interviews and focus groups with the participants (Olson et al., 2020;Haginoya, 2020). The interviews focused on the participants' professional experiences and aspirations to advance in their respective fields or workplaces. ...
Article
Full-text available
Women's progress in leadership roles within the field of education has made limited headway despite the efforts of government institutions and progressive policies aimed at addressing diversity and equity practices. This study seeks to investigate the experiences of women in leadership within the Department of Education in Lubombo Region of Eswatini. Adopting a qualitative approach, this study employs an interpretive paradigm and a phenomenological design to delve into the lived experiences of female leaders and their interpretation of these experiences. Twelve female principals from 12 high schools were purposefully selected to participate in this study. Semi-structured interviews, conducted with audio recordings, were used to guide the data generation process. The collected data were systematically transcribed, coded, and analysed thematically. Feminist theory serves as the foundation of this study, guiding the retrospective storytelling process aimed at understanding the experiences of women in leadership. Despite the implementation of regulations and other initiatives, gender equality in Eswatini has not progressed significantly. The findings of this study reveal that deeply ingrained traditional beliefs, rooted in the customs and values of local communities, continue to strongly influence promotion decisions within the educational system. The resistance observed reflects a deeply entrenched belief that female principals have limited opportunities for upward mobility in leadership due to the societal stigma or "injustice prejudice" that permeates government institutions. This study recommends that government entities and other stakeholders embark on new leadership training programs that prioritise progressive ethos, values, and culture within the education industry.
... McMillian and Schumacher (2006;, argue that purposeful sampling identifies 'information-rich' participants who may be knowledgeable about the phenomenon under investigation. The participants were engaged in face-to-face semi-structured interviews (Olson et al., 2020;Haginoya, 2020). The interviews were conducted in their schools as scheduled by the participants. ...
Article
Full-text available
This interpretivist qualitative study explored lived experiences of rural secondary schools’ female principals' career development. Participants were female principals who were purposively selected from rural schools in South Africa to provide in-depth information to the study. Thus, eleven female principals from 11 public secondary schools within Ilembe District, Limpopo were engaged in semi-structured interviews, which lasted between 20-30 minutes each. The interviews were audio-recorded with permission of the participants to capture exact information. The audio-recorded data was systematically transcribed, coded to generate common themes for the presentation and discussion of findings. Findings revealed that some communities’ cultures continue to dominantly influence women principals’ career pathways, despite policies or efforts to promote gender equality in South Africa. The social attitude or ‘injustice’ against female principals in rural environments has limited their quest to pursue their career further or to aspire top administrative leadership positions in education. It is recommended that necessary government structures and other stakeholders in education sector should provide necessary support to female principals in rural schools.
... Interviewer-administered surveys are often treated as a superior form of data collection, for example, concerning response rates, communication with respondents, and administration of long questionnaires (Groves et al. 2009;Olson et al. 2020). By encouraging respondents' participation, answering their queries, and ensuring questionnaire completion, interviewers play a vital role for survey quality. ...
Article
Full-text available
Deviant interviewer behavior is a potential hazard of interviewer-administered surveys, with interviewers fabricating entire interviews as the most severe form. Various statistical methods (e.g., cluster analysis) have been proposed to detect falsifiers. These methods often rely on falsification indicators aiming to measure differences between real and falsified data. However, due to a lack of real-world data, empirical evaluations and comparisons of different statistical methods and falsification indicators are scarce. Using a large-scale nationally representative refugee survey in Germany with known fraudulent interviews, this study tests, evaluates, and compares statistical methods for identifying falsified data. We investigate the use of new and existing falsification indicators as well as multivariate detection methods for combining them. Additionally, we introduce a new and easy-to-use multivariate detection method that overcomes practical limitations of previous methods. We find that the vast majority of used falsification indicators successfully measure differences between falsifiers and nonfalsifiers, with the newly proposed falsification indicators outperforming some existing indicators. Furthermore, different multivariate detection methods perform similarly well in detecting the falsifiers.
Article
When and how to survey potential respondents is often determined by budgetary and external constraints, but choice of survey modality may have enormous implications for data quality. Different survey modalities may be differentially susceptible to measurement error attributable to interviewer assignment, known as interviewer effects. In this paper, we leverage highly similar surveys, one conducted face-to-face (FTF) and the other via phone, to examine variation in interviewer effects across survey modality and question type. We find that while there are no cross-modality differences for simple questions, interviewer effects are markedly higher for sensitive questions asked over the phone. These findings are likely explained by the enhanced ability of in-person interviewers to foster rapport and engagement with respondents. We conclude with a thought experiment that illustrates the potential implications for power calculations, namely, that using FTF data to inform phone surveys may substantially underestimate the necessary sample size for sensitive questions.
Article
Interviewers’ postinterview evaluations of respondents’ performance (IEPs) are paradata, used to describe the quality of the data obtained from respondents. IEPs are driven by a combination of factors, including respondents’ and interviewers’ sociodemographic characteristics and what actually transpires during the interview. However, relatively few studies examine how IEPs are associated with features of the response process, including facets of the interviewer-respondent interaction and patterns of responding that index data quality. We examine whether features of the response process—various respondents’ behaviors and response quality indicators—are associated with IEPs in a survey with a diverse set of respondents focused on barriers and facilitators to participating in medical research. We also examine whether there are differences in IEPs across respondents’ and interviewers’ sociodemographic characteristics. Our results show that both respondents’ behaviors and response quality indicators predict IEPs, indicating that IEPs reflect what transpires in the interview. In addition, interviewers appear to approach the task of evaluating respondents with differing frameworks, as evidenced by the variation in IEPs attributable to interviewers and associations between IEPs and interviewers’ gender. Further, IEPs were associated with respondents’ education and ethnoracial identity, net of respondents’ behaviors, response quality indicators, and sociodemographic characteristics of respondents and interviewers. Future research should continue to build on studies that examine the correlates of IEPs to better inform whether, when, and how to use IEPs as paradata about the quality of the data obtained.
Article
Full-text available
Public opinion surveys are a fundamental tool to measure support for women’s political rights. This article focuses on perceptions of women’s suitability for leadership. To what extent do influential cross-country surveys that include such items suffer from measurement errors stemming from gender of interviewer effects? Building on the literature on social desirability, we expect that respondents are more likely to express preference for men’s suitability as political leaders with male interviewers and more likely to state support for women’s leadership when interviewed by a woman. We hypothesize that these processes are conditioned by having one’s spouse present, by age differences between respondents and interviewers, as well as by respondents’ levels of education. Analyzing Afrobarometer data, we generally find support for our claims. In addition, it seems that men are slightly more affected by such effects than women are. These gender of interviewer effects persist when analyzing alternative survey rounds and are insensitive to various fixed effects specifications and robustness tests. For the analysis of survey data, we suggest that researchers using gender-related items should control for gender of interviewer effects. We propose that comparative survey programs pay even more attention to interviewer characteristics and the interview situation in their protocols.
Article
Full-text available
In surveys, nonsampling errors are harder to quantify, when compared to sampling errors, due to their complex nature. Avoiding to release them, however, will result in bias on the survey estimates. Previously, we devise the best interviewer allocation technique by using nested experimental design to study the response error estimation. In this study, to shed new light on the success of this methodology, we apply it to an interview-reinterview surveys for inquiring about the time use and life satisfaction of academicians at Middle East Technical University, Turkey. Analysis of the pilot survey data reveals that only half of the data is reliable; interviewer effects are observed on the other half. Before the main survey, interviewers are trained particularly emphasizing on these questions, and thus, the associated response variances accounted for the total variance and data unreliability are reduced considerably, and hence, the quality of the survey is increased.
Article
Full-text available
Telephone surveys have been a ubiquitous method of collecting survey data, but the environment for telephone surveys is changing. Many surveys are transitioning from telephone to self-administration or combinations of modes for both recruitment and survey administration. Survey organizations are conducting these transitions from telephone to mixed modes with only limited guidance from existing empirical literature and best practices. This article summarizes findings by an AAPOR Task Force on how these transitions have occurred for surveys and research organizations in general. We find that transitions from a telephone to a self-administered or mixed-mode survey are motivated by a desire to control costs, to maintain or improve data quality, or both. The most common mode to recruit respondents when transitioning is mail, but recent mixed-mode studies use only web or mail and web together as survey administration modes. Although early studies found that telephone response rates met or exceeded response rates to the self-administered or mixed modes, after about 2013, response rates to the self-administered or mixed modes tended to exceed those for the telephone mode, largely because of a decline in the telephone mode response rates. Transitioning offers opportunities related to improved frame coverage and geographic targeting, delivery of incentives, visual design of an instrument, and cost savings, but challenges exist related to selecting a respondent within a household, length of a questionnaire, differences across modes in use of computerization to facilitate skip patterns and other questionnaire design features, and lack of an interviewer for respondent motivation and clarification. Other challenges related to surveying youth, conducting surveys in multiple languages, collecting nonsurvey data such as biomeasures or consent to link to administrative data, and estimation with multiple modes are also prominent.
Article
Full-text available
Why do telephone interviews last longer on cell phones than landline phones? Common explanations for this phenomenon include differential selection into subsets of questions, activities outside the question-answer sequence (such as collecting contact information for cell-minute reimbursement), respondent characteristics, behaviors indicating disruption to respondents’ perception and comprehension, and behaviors indicating interviewer reactions to disruption. We find that the time difference persists even when we focus only on the question-answer portion of the interview and only on shared questions (i.e., eliminating the first two explanations above). To learn why the difference persists, we use behavior codes from the U.S./Japan Newspaper Opinion Poll, a dual-frame telephone survey of US adults, to examine indicators of satisficing, line-quality issues, and distraction. Overall, we find that respondents on cell phones are more disrupted, and that the difference in interview duration occurs because cell phone respondents take longer to provide acceptable answers. Interviewers also slow their speed of speech when asking questions. A slower speaking rate from both actors results in a longer and more expensive interview when respondents use cell phones.
Article
Full-text available
Respondents, interviewers, and questions are all sources of potential error in responses to survey questions. Using behavior coding to measure respondent problems and interviewer errors, we assess whether question characteristics influences these behaviors. We also assess whether interviewer errors may help to explain the impact of question characteristics on respondent difficulties. Specifically, we assess whether more difficult questions (e.g., longer or more difficult to read) lead to more respondent problems (e.g., answers that do not fit the requirements of the question) at least in part because more difficult questions also lead to interviewer errors (e.g., deviating from the reading the question as written or failing to probe correctly).
Article
Although researchers have used phone surveys for decades, the lack of an accurate picture of the call opening reduces our ability to train interviewers to succeed. Sample members decide about participation quickly. We predict participation using the earliest moments of the call; to do this, we analyze matched pairs of acceptances and declinations from the Wisconsin Longitudinal Study using a case-control design and conditional logistic regression. We focus on components of the first speaking turns: acoustic-prosodic components and interviewer's actions. The sample member's "hello" is external to the causal processes within the call and may carry information about the propensity to respond. As predicted by Pillet-Shore (2012), we find that when the pitch span of the sample member's "hello" is greater the odds of participation are higher, but in contradiction to her prediction, the (less reliably measured) pitch pattern of the greeting does not predict participation. The structure of actions in the interviewer's first turn has a large impact. The large majority of calls in our analysis begin with either an "efficient" or "canonical" turn. In an efficient first turn, the interviewer delays identifying themselves (and thereby suggesting the purpose of the call) until they are sure they are speaking to the sample member, with the resulting efficiency that they introduce themselves only once. In a canonical turn, the interviewer introduces themselves and asks to speak to the sample member, but risks having to introduce themselves twice if the answerer is not the sample member. The odds of participation are substantially and significantly lower for an efficient turn compared to a canonical turn. It appears that how interviewers handle identification in their first turn has consequences for participation; an analysis of actions could facilitate experiments to design first interviewer turns for different target populations, study designs, and calling technologies.
Article
An Introduction to Survey Participation. A Conceptual Framework for Survey Participation. Data Resources for Testing Theories of Survey Participation. Influences on the Likelihood of Contact. Influences of Household Characteristics on Survey Cooperation. Social Environmental Influences on Survey Participation. Influences of the Interviewers. When Interviewers Meet Householders: The Nature of Initial Interactions. Influences of Householder-Interviewer Interactions on Survey Cooperation. How Survey Design Features Affect Participation. Practical Survey Design Acknowledging Nonresponse. References. Index.
Chapter
This chapter reviews some recent developments in computer-assisted recorded interviewing (CARI), mobile technology, geographic information systems (GIS), and field management informed by the total survey error (TSE) paradigm. Survey operations are widely understood to include data collection activities, though the scope of this can be ambiguous in some modes, such as the web. In the survey life cycle, survey operations can be defined as everything that happens on a survey between design and final report creation or data delivery. The chapter describes the CARI coding process at Westat and discusses the results of a recent experiment that demonstrates how providing feedback to interviewers may help to improve the quality of survey interviews. It presents an approach to integrating paradata and quality metrics for continuous improvement, and suggests that TSE concepts can be effective when applied during survey operations by supervisors who have direct responsibility for the data collection activities.
Article
A rich and diverse literature exists on the effects that human interviewers can have on different aspects of the survey data collection process. This research synthesis uses the Total Survey Error (TSE) framework to highlight important historical developments and advances in the study of interviewer effects on a variety of important survey process outcomes, including sample frame coverage, contact and recruitment of potential respondents, survey measurement, and data processing. Included in the scope of the synthesis is research literature that has focused on explaining variability among interviewers in these effects and the different types of variable errors that they can introduce, which can ultimately affect the efficiency of survey estimates. We first consider common tasks with which human interviewers are often charged and then use the TSE framework to organize and synthesize the literature discussing the variable errors that interviewers can introduce when attempting to execute each task. Based on our synthesis, we identify key gaps in knowledge and then use these gaps to motivate an organizing model for future research investigating explanations for interviewer effects on different aspects of the survey data collection process.