Mick P. Couper’s research while affiliated with Concordia University Ann Arbor and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (291)


Comparing Human and GPT-4o coding with temperature 0 or 0.7, respectively (Block 1)
% Agreement and Cohen's Kappa (Block 1)
% Agreement and Cohen's Kappa (Block 2)
Transcribing and Coding Voice Answers Obtained in Web Surveys: Comparing Three Leading Automatic Speech Recognition Tools and Human versus LLM-based Coding
  • Preprint
  • File available

March 2025

·

79 Reads

·

·

·

Mick P Couper

With the rise of smartphone use in web surveys, voice or oral answers have become a promising methodology for collecting rich data. Voice answers not only facilitate broader and more detailed narratives but also include additional metadata, such as voice amplitude and pitch, to assess respondent engagement. Despite these advantages, challenges persist, including high item non-response rates, mixed respondent preferences for voice input, and labor-intensive, manual answer transcription and coding. This study addresses these last two challenges by evaluating two critical aspects of processing voice answers. First, it compares the transcription performance of three leading Automatic Speech Recognition (ASR) tools—Google Cloud Speech-to-Text API, OpenAI Whisper, and Vosk—using voice answers collected from an open-ended question on nursing home transparency that was administered in an opt-in online panel in Spain. Second, it evaluates the efficiency and quality of coding these transcriptions using human coders and GPT-4o, a Large Language Model (LLM) developed by OpenAI. We found that each of the ASR tools has distinct merits and limits. Google sometimes fails to provide transcriptions, Whisper produces hallucinations (false transcriptions), and Vosk has clarity issues and high rates of incorrect words. Human and LLM-based coding also differ significantly. Thus, we recommend using several ASR tools for voice answer transcription and implementing human as well as LLM-based coding, as the latter offers additional information at minimal added cost.

Download

Toward a New Approach to Creating Population-Representative Data for Demographic Research

December 2024

·

9 Reads

·

1 Citation

Demography

Brady T West

·

Mick P Couper

·

William G Axinn

·

[...]

·

The evaluation of innovative web-based data collection methods that are convenient for the general public and that yield high-quality scientific information for demographic researchers has become critical. Web-based methods are crucial for researchers with nationally representative research objectives but without the resources of larger organizations. The web mode is appealing because it is inexpensive relative to in-person and telephone modes, and it affords a high level of privacy. We evaluate a sequential mixed-mode web/mail data collection, conducted with a national probability sample of U.S. adults from 2020 to 2022. The survey topics focus on reproductive health and family formation. We compare estimates from this survey to those obtained from a face-to-face national survey of population reproductive health: the 2017–2019 National Survey of Family Growth (NSFG). This comparison allows for maximum design complexity, including a complex household screening operation (to identify households with persons aged 18–49). We evaluate the ability of this national web/mail data collection approach to (1) recruit a representative sample of U.S. persons aged 18–49; (2) replicate key survey estimates based on the NSFG, considering expected effects of the COVID-19 pandemic lockdowns and the alternative modes on the estimates; (3) reduce complex sample design effects relative to the NSFG; and (4) reduce the costs per completed survey.


Improving follow-up survey completion rates through pilot interventions in the All of Us Research Program: Results from a non-randomized intervention study

October 2024

·

18 Reads

Objective Retention to complete follow-up surveys in extensive longitudinal epidemiological cohort studies is vital yet challenging. All of Us developed pilot interventions to improve response rates for follow-up surveys. Study design and setting The pilot interventions occurred from April 27, 2020, to August 3, 2020. The three arms were: (1) telephone appointment [staff members calling participants offering appointments to complete surveys over phone] (2) postal [mail reminder to complete surveys through U.S. Postal Service], and (3) combination of telephone appointment and postal. Controls received digital-only reminders [program-level digital recontact via email or through the participant portal]. Study sites chose their study arm and participants were not randomized. Results A total of 50 sites piloted interventions with 17,593 participants, while 47,832 participants comprised controls during the same period. Of all participants, 6,828 (10.4%) completed any follow-up surveys (1448: telephone; 522: postal; 486: combination; 4372: controls). Follow-up survey completions were 24% higher in the telephone appointment arm than in controls in bivariate analyses. When controlling for confounders, telephone appointment and combination arms increased rates of completion similarly compared to controls, while the postal arm had no significant effect (odds ratio [95% Confidence Interval], telephone appointment:2.01[1.81–2.23]; combination:1.91[1.66–2.20]; postal:0.92[0.79–1.07]). Although the effects of the telephone appointment and combination arms were similar, differential effects were observed across sub-populations. Conclusion Telephone appointments appeared to be the most successful intervention in our study. Lessons learned about retention interventions, and improvement in follow-up survey completion rates provide generalizable knowledge for similar cohort studies and demonstrate the potential value of precision reminders and engagement with sub-populations of a cohort.


A National RDD Smartphone Web Survey: Comparison With a Large-Scale CAPI Survey

The most important national surveys of the general population for creating official statistics or public policymaking in many countries, including South Korea, are still conducted using face-to-face interviews with household members. Recently face-to-face surveys have faced threats to data quality from decreasing response rates and rising costs of in-person visits. The COVID-19 pandemic and associated lockdown exacerbated the situation for face-to-face surveys. Survey organizations suspended fieldwork or began to explore alternate means of collecting data. One alternative was a shift to telephone surveys; however, telephone interviews have encountered similar difficulties with declining response rates and increasing costs. Could a self-administered web survey be a viable alternative to interviewer-administered modes such as telephone interviews (CATI) or face-to-face interviews (CAPI)? Smartphones may offer opportunities not offered by other modes. We conducted a smartphone web survey using SMS invitations where a sample of cell phone numbers was selected by random digit dialing (RDD) and compared it with a large-scale national face-to-face survey (CAPI) where a sample of households was selected by stratified cluster sampling. The two surveys were conducted during the COVID pandemic in the second half of 2020. The coverage and sample representation of the smartphone web survey were comparable to that of the face-to-face survey. Despite the relatively small number of respondents, the quality of the smartphone web survey was sufficient to provide accurate data and compared favorably with the CAPI survey. The smartphone web survey yielded more reports of socially undesirable attitudes and behavior than the CAPI survey. The findings will guide researchers to explore new opportunities in establishing a web survey methodology that obtains data more conveniently, efficiently, accurately, and with less cost.


Measuring Expenditure with a Mobile App: Do Probability-Based and Nonprobability Panels Differ?

June 2024

·

18 Reads

Journal of Survey Statistics and Methodology

In this case study, we examine a novel aspect of data collected in a typical probability and a typical nonprobability panel: mobile app data. The data were collected in Great Britain in 2018, using the Innovation Panel of the UK Household Longitudinal Study and the Lightspeed online access panel. Respondents in each panel were invited to participate in a month-long study, reporting all their daily expenditures in the app. In line with most of the research on nonprobability and probability-based panel data, our results indicate differences in the data gathered from these data sources. For example, more female, middle-aged, and highly educated people with higher digital skills and a greater interest in their finances participated in the nonprobability app study. Our findings also show that resulting differences in the app spending data are difficult to eliminate by weighting. The only data quality aspect for which we do not find evidence of differences between the nonprobability and probability-based panel is behavior in using the spending app. This finding is contrary to the argument that nonprobability online panel participants try to maximize their monetary incentive at the expense of data quality. However, this finding is in line with some of the scarce existing literature on response behavior in surveys, which is inconclusive regarding the question of whether nonprobability online panel participants answer questions less conscientiously than probability-based panel respondents. Since the two panels in our case study differ in more aspects than the sample selection procedure, more research in different contexts is necessary to establish generalizability and causality.


The Effects of Placement and Order on Consent to Data Linkage in a Web survey

February 2024

·

19 Reads

Journal of Survey Statistics and Methodology

We report on an experiment in a supplemental web survey as part of a longitudinal study in the United Kingdom where we ask survey respondents to consent to two forms of data linkage to health records and to consent to be mailed a serology kit. We varied the placement (early, early in context, or late in the survey) and order (linkage first or serology first) of the consent requests. We also examine reasons for consent or non-consent. We find that order of the requests does not make much difference, but making the requests early in the survey significantly increases consent rates over asking them after a series of content-related questions (by 3.4 percentage points) or later in the survey (by 7.2 percentage points). This is consistent with previous research showing that early requests for consent in a survey have a positive effect. The main reason chosen for not consenting related to the personal nature of the information requested.


Understanding Society: minimising selection biases in data collection using mobile apps

December 2023

·

27 Reads

·

3 Citations

Fiscal Studies

Understanding Society, the UK Household Longitudinal Study has a programme of research and development that underpins innovations in data collection methods. One of our current focuses is on using mobile applications to collect additional data that supplement data collected in annual interviews. To date, we have used mobile apps to collect data on consumer expenditure, well‐being, anthropometrics and cognition. In this paper, we review the potential barriers to data collection using mobile apps and experimental evidence collected with the Understanding Society Innovation Panel, on what can be done to reduce these barriers.


Screenshot examples of mobile grids and four item-by-item layout alternatives on mobile devices.
Differences in RQIs Between a Specific Layout and the Corresponding Five-Layout Average.
Differences in RQIs Between PC and SP Respondents Across the Five Layouts.
Differences in Survey Estimates Between the Layout and Corresponding Five-Layout Averages.
Differences in Survey Estimates Between PC and SP Respondents Across the Five Layouts.
Alternative Layouts for Grid Questions in PC and Mobile Web Surveys: An Experimental Evaluation Using Response Quality Indicators and Survey Estimates

December 2023

·

152 Reads

·

4 Citations

The grid question refers to a table layout for a series of survey question items (i.e., sub-questions) with the same introduction and identical response categories. Because of their complexity, concerns have already been raised about grids in web surveys on PCs, and these concerns have heightened regarding mobile devices. Some studies suggest decomposing grids into item-by-item layouts, while others argue that this is unnecessary. To address this challenge, this paper provides a comprehensive evaluation of the grid layout and four item-by-item alternatives, using 10 response quality indicators and 20 survey estimates. Results from the experimental web survey (n = 4644) suggest that item-by-item layouts (unfolding or scrolling) should be used instead of grids, not only on mobile devices but also on PCs. While the former justifies the already increasing use of item-by-item layouts on mobile devices in survey practice, the latter implies that the prevailing routine of using grids on PCs should be reconsidered.


What Predicts Willingness to Participate in a Follow-up Panel Study among Respondents to a National Web/Mail Survey?

August 2023

·

22 Reads

Field Methods

The American Family Health Study (AFHS) collected family health and fertility data from a national probability sample of persons aged 18–49 between September 2021 and May 2022, using web and mail exclusively. In July 2022, we surveyed AFHS respondents and gauged their willingness to become part of a national web panel that would create novel longitudinal data on these topics. We focus on predictors of willingness to participate, identifying the potential selection bias that this type of approach may introduce. We found that efforts of this type to create a national web panel may introduce potential selection bias in estimates based on the panel respondents, with individuals having higher socioeconomic status being more cooperative. Thus, alternative recruitment strategies and re-weighting of the subsample may be needed to further reduce selection bias. We present methodological implications of our results, limitations of our approach, and suggestions for further research on this topic.


What Predicts Willingness to Participate in a Follow-up Panel Study among Respondents to a National Web/Mail Survey?

August 2023

·

2 Reads

Field Methods

The American Family Health Study (AFHS) collected family health and fertility data from a national probability sample of persons aged 18–49 between September 2021 and May 2022, using web and mail exclusively. In July 2022, we surveyed AFHS respondents and gauged their willingness to become part of a national web panel that would create novel longitudinal data on these topics. We focus on predictors of willingness to participate, identifying the potential selection bias that this type of approach may introduce. We found that efforts of this type to create a national web panel may introduce potential selection bias in estimates based on the panel respondents, with individuals having higher socioeconomic status being more cooperative. Thus, alternative recruitment strategies and re-weighting of the subsample may be needed to further reduce selection bias. We present methodological implications of our results, limitations of our approach, and suggestions for further research on this topic.


Citations (81)


... Finally, apps allow researchers to solicit the collection of longitudinal data through mobile notifications rather than through recontact calls or emails. However, many challenges remain: the target population needs to have a mobile device; the mobile app should be designed to run on different operating systems; respondents' reluctance to download a mobile app for privacy or data security concerns; a demanding task, such as daily activities or a long period of the study (Jacobsen and Kühne, 2021;Jäckle et al., 2022Jäckle et al., , 2023. ...

Reference:

Collecting data on migrants’ health status and access to health services: the experience of the mobile app “ComeStai”
Understanding Society: minimising selection biases in data collection using mobile apps

Fiscal Studies

... The study sample was drawn from a larger general US population sample of internet panelists who regularly participated in online surveys. Prior research has shown that the samples recruited from these panels are not necessarily representative of the general population (Stone et al., 2023b) and the observed effects may be different in other (clinical) populations which may differ in the patterns of delayed EMA responses and in their motivation for timely assessment completion. As EMA is increasingly applied in clinical populations, further research on the occurrence and impact of delayed responses in clinical samples is needed. ...

Shedding light on participant selection bias in Ecological Momentary Assessment (EMA) studies: Findings from an internet panel study

... Female gender and higher income are associated with willingness to participate, and when the survey is computer-assisted, computer proficiency is also an important factor (Stone et al., 2023). Further, Dollinger andLeong (1993) found that individuals higher on Agreeableness, Openness, and Extraversion are more prone to take part in surveys, including online research (Buchanan, 2018;Stone et al., 2023). Such selfselection can lead to range restrictions, impact research findings, and thus limit their generalizability. ...

A population-based investigation of participation rate and self-selection bias in momentary data capture and survey studies

Current Psychology

... Due to these data sensitivities, a range of privacy safeguards and governance processes exist around the transfer and use of health data. These include legal and contractual arrangements, policies governing data access [15][16][17], information security requirements [18,19], and a specific privacy-preserving data flow known as the 'separation principle' [20]. Under this principle, PII is split from clinical information, with only PII (no clinical data) provided to linkage units and clinical information (no PII) provided to the end-user [9,21]. ...

How and why does the mode of data collection affect consent to data linkage?

... Dobiveni je udio na razini ukupnoga broja ispitanika bio izuzetno mali (samo 0,2 %). Tako mali udio nedostajućih odgovora može se objasniti time što je upitnik izrađen tako da je svako pitanje iz upitnika bilo na zasebnoj web-stranici, što, poznato je, u mnogome smanjuje sklonost ispitanika da preskoči pitanje odnosno odgovor (Vehovar et al., 2022). Također, i taj je mali udio neodgovora podjednako raspoređen među grupama koje nas u ovom radu zanimaju. ...

Alternative Layouts for Grid Questions in PC and Mobile Web Surveys: An Experimental Evaluation Using Response Quality Indicators and Survey Estimates

... Messer and Dillman (2011) found that following up an initial request that included a $5 pre-paid incentive with a second $5 pre-paid incentive via priority mail to nonresponders increased the response rate of a state-wide web survey from 59% to 68%. In the context of a different mail-web survey, Wagner, West, Couper, Zhang, Gatward, Nishimura, and Saw (2023) found that following up a $2 pre-paid incentive included in the initial mailing with an additional $5 sent via priority mail considerably increased response rates. However, in both of these studies the effect of the second incentive cannot be disentangled from the effect of the priority mailing. ...

An Experimental Evaluation of Two Approaches for Improving Response to Household Screening Efforts in National Mail/Web Surveys
  • Citing Article
  • July 2022

Journal of Survey Statistics and Methodology

... McCabe et al. (2006) found few significant differences between survey modes: their results suggest that web and mail surveys provide comparable estimates of alcohol use in a non-randomised mixed mode design. Patrick et al. (2022) found that while response rates and substance use estimates were not significantly affected, the mode of response was influenced by sociodemographic factors such as race, smoking habits, marital status, and education level. Reported substance use prevalence did not significantly differ according to survey mode after adjusting for sociodemographic characteristics. ...

Comparison of a web-push vs. mailed survey protocol in the Monitoring the Future panel study among adults ages 35 to 60

Drug and Alcohol Dependence Reports

... While there are many other experimental comparisons of mode effects between web and faceto-face surveys, most of these involve pre-recruited samples or panel studies, or mixed-mode designs (e.g., Bianchi et al., 2017;Biemer et al., 2022;Jäckle et al., 2015;McGonagle & Freedman, 2017;Ofstedal et al., 2022). We do not review these further as they do not directly pertain to our comparison. ...

Data quality and response distributions in a mixed-mode survey
  • Citing Article
  • May 2022

Longitudinal and Life Course Studies

... Finally, apps allow researchers to solicit the collection of longitudinal data through mobile notifications rather than through recontact calls or emails. However, many challenges remain: the target population needs to have a mobile device; the mobile app should be designed to run on different operating systems; respondents' reluctance to download a mobile app for privacy or data security concerns; a demanding task, such as daily activities or a long period of the study (Jacobsen and Kühne, 2021;Jäckle et al., 2022Jäckle et al., , 2023. ...

Increasing Participation in a Mobile App Study: The Effects of a Sequential Mixed-Mode Design and In-Interview Invitation

Journal of Survey Statistics and Methodology

... There is an abundance of research on optimizing consent requests across various domains, ranging from organ donation for biobanks (e.g., Johnson & Goldstein, 2003), contact tracing (e.g., Altmann et al., 2020) and transferring contact information to a third-party data-collection agency (e.g., to making multiple simultaneous requests (e.g., Beuthner et al., 2022;Walzenbach et al., 2022). In survey research, the most prominent consent requests relate to linking survey data with other data such as passively collected (e.g., GPS) data (Felderer & Blom, 2019;Keusch et al., 2019), administrative data (Bacher, 2023;Hülle, 2024;Jäckle et al., 2021;Knies et al., 2012; 1 Federal Trade Commission Act of 1914, 15 U. S. C. § § 41-58. ...

Experiments On Multiple Requests For Consent to Data Linkage in Surveys

Journal of Survey Statistics and Methodology