Merrick Osborne’s research while affiliated with Minnesota School of Business and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (13)


Advice Preferences for AI- versus Human-Generated Advice by Topic. Figure 1 displays a density ridge visualization of the distribution of preferences for artificial intelligence (AI) versus human-generated advice across the set of topics. Colors represent unique tasks, ordered by the strength of the preference for human-generated advice.
Advice Evaluations for ChatGPT- and Human-Generated Advice. The figure displays the evaluation of the advice by participants. The evaluations of ChatGPT are presented in the darker bars, with evaluations of human-generated advice in the lighter bars. The bars represent the mean, and the error bars represent the standard errors.
Advice Evaluations for ChatGPT- and Self-Generated Advice by Order. The figure displays the evaluation of the advice by participants for Study 4 (left) and Study 5 (right). These evaluations are split between those that generated and evaluated their advice prior to evaluating ChatGPT advice (right) and vice versa (left). The color represents the target of the ratings, with self-evaluations in gray, lighter bars, and ChatGPT ratings in black, darker bars. The height of the bars represents the mean, and the error bars represent the standard errors.
Me vs. the machine? Subjective evaluations of human- and AI-generated advice
  • Article
  • Full-text available

February 2025

·

27 Reads

·

1 Citation

Merrick R. Osborne

·

Erica R. Bailey

Artificial intelligence (“AI”) has the potential to vastly improve human decision-making. In line with this, researchers have increasingly sought to understand how people view AI, often documenting skepticism and even outright aversion to these tools. In the present research, we complement these findings by documenting the performance of LLMs in the personal advice domain. In addition, we shift the focus in a new direction—exploring how interacting with AI tools, specifically large language models, impacts the user’s view of themselves. In five preregistered experiments (N = 1,722), we explore evaluations of human- and ChatGPT-generated advice along three dimensions: quality, effectiveness, and authenticity. We find that ChatGPT produces superior advice relative to the average online participant even in a domain in which people strongly prefer human-generated advice (dating and relationships). We also document a bias against ChatGPT-generated advice which is present only when participants are aware the advice was generated by ChatGPT. Novel to the present investigation, we then explore how interacting with these tools impacts self-evaluations. We manipulate the order in which people interact with these tools relative to self-generation and find that generating advice before interacting with ChatGPT advice boosts the quality ratings of the ChatGPT advice. At the same time, interacting with ChatGPT-generated advice before self-generating advice decreases self-ratings of authenticity. Taken together, we document a bias towards AI in the context of personal advice. Further, we identify an important externality in the use of these tools—they can invoke social comparisons of me vs. the machine.

Download

Responsible Generative AI Use by Product Managers: Recoupling Ethical Principles and Practices

January 2025

·

39 Reads

Genevieve Smith

·

Natalia Luka

·

Merrick Osborne

·

[...]

·

Since 2022, generative AI (genAI) has rapidly become integrated into workplaces. Though organizations have made commitments to use this technology "responsibly", how organizations and their employees prioritize responsibility in their decision-making remains absent from extant management theorizing. In this paper, we examine how product managers - who often serve as gatekeepers in decision-making processes - implement responsible practices in their day-to-day work when using genAI. Using Institutional Theory, we illuminate the factors that constrain or support proactive responsible development and usage of genAI technologies. We employ a mixed methods research design, drawing on 25 interviews with product managers and a global survey of 300 respondents in product management-related roles. The majority of our respondents report (1) widespread uncertainty regarding what "responsibility" means or looks like, (2) diffused responsibility given assumed ethical actions by other teams, (3) lack of clear incentives and guidance within organizations, and (4) the importance of leadership buy-in and principles for navigating tensions between ethical commitments and profit motives. However, our study finds that even in highly uncertain environments, absent guidance from leadership, product managers can "recouple" ethical commitments and practices by finding responsibility "micro-moments". Product managers seek out low-risk, small-scale actions they can take without explicit buy-in from higher-level managers, such as individual or team-wide checks and reviews and safeguarding standards for data. Our research highlights how genAI poses unique challenges to organizations trying to couple ethical principles and daily practices and the role that middle-level management can play in recoupling the two.






The sins of the parents are to be laid upon the children: biased humans, biased data, biased models

November 2022

·

20 Reads

·

1 Citation

Technological innovations have become a key driver of societal advancements. Nowhere is this more evident than in the field of machine learning (ML), which has developed algorithmic models that shape our decisions, behaviors, and outcomes. These tools have widespread use, in part, because they can synthesize massive amounts of data to make seemingly objective recommendations. Yet, in the past few years, the ML community has been raising the alarm on why we should be cautious in interpreting and using these models: they are created by humans, from data generated by humans, whose psychology allows for various biases that impact how the models are developed, trained, tested and interpreted. As psychologists, we thus face a fork in the road; Down the first path, we can continue to use these models without examining and addressing these critical flaws, and rely on computer scientists to try to mitigate them. Down the second path, we can turn our expertise in bias towards this growing field, collaborating with computer scientists to mitigate the deleterious outcomes associated with these models. This paper serves to light the way down the second path by identifying how extant psychological research can help examine and mitigate bias in ML models.




Structuring local environments to avoid racial diversity: Anxiety drives Whites' geographical and institutional self-segregation preferences

July 2021

·

44 Reads

·

34 Citations

Journal of Experimental Social Psychology

The current research explores how local racial diversity affects Whites' efforts to structure their local communities to avoid incidental intergroup contact. In two experimental studies (N = 509; Studies 1a-b), we consider Whites' choices to structure a fictional, diverse city and find that Whites choose greater racial segregation around more (vs. less) self-relevant landmarks (e.g., their workplace and children's school). Specifically, the more time they expect to spend at a landmark, the more they concentrate other Whites around that landmark, thereby reducing opportunities for incidental intergroup contact. Whites also structure environments to reduce incidental intergroup contact by instituting organizational policies that disproportionately exclude non-Whites: Two large-scale archival studies (Studies 2a-b) using data from every U.S. tennis (N = 15,023) and golf (N = 10,949) facility revealed that facilities in more racially diverse communities maintain more exclusionary barriers (e.g., guest policies, monetary fees, dress codes) that shield the patrons of these historically White institutions from incidental intergroup contact. In a final experiment (N = 307; Study 3), we find that Whites' anticipated intergroup anxiety is one driver of their choices to structure environments to reduce incidental intergroup contact in more (vs. less) racially diverse communities. Our results suggest that despite increasing racial diversity, White Americans structure local environments to fuel a self-perpetuating cycle of segregation.


Citations (4)


... The high effectiveness of such a solution has already been confirmed in the studies by Butt et al. (2021), where AI was a human avatar in games, and Erengin et al. (2025) claimed that AI-generated personal and psychological added value for customers. Finally, Osborne and Bailey (2025) found that AI significantly aids human decision-making. Therefore, it seems justifiable to assume that the concept of "Me and AI" should contribute significantly to improving the efficiency of the agent's training processes in life insurance companies, although the author cautions against excessive optimism in this area, as each approach has its limitations. ...

Reference:

The Effectiveness of Life Insurance Sales Force Training: Welcome “Me and AI”
Me vs. the machine? Subjective evaluations of human- and AI-generated advice

... In spite of the resource-saving potential, we similarly urge caution and careful psychometric consideration. Decades of research in social psychology have established that judgments of human behavior are strongly subject to prejudice and bias (Greenwald & Krieger, 2006), and lessons from NLP suggest that these biases will inevitably permeate early attempts at behavioral coding algorithms (Osborne et al., 2022). In the long run, many of our recommendations for mitigating sources of biases in human coders (see "Coding system development" and "Recruiting and training coders") may be adapted for the training of machine learning algorithms. ...

The sins of the parents are to be laid upon the children: biased humans, biased data, biased models
  • Citing Preprint
  • November 2022

... This assumption predicates the outgroup homogeneity and ingroup tuning hypotheses. Homophily remains normative in many (if not most) societies due to both societal structure and personal choices (Anicich et al., 2021;Iceland et al., 2002), so on average, we might expect configurations of face space that conform to Valentine's predictions. However, individual perceivers vary with respect to their intergroup interactions. ...

Structuring local environments to avoid racial diversity: Anxiety drives Whites' geographical and institutional self-segregation preferences
  • Citing Article
  • July 2021

Journal of Experimental Social Psychology

... This enhances ecological validity and answers calls to integrate this contextual component of emotion regulation as a factor that influences emotion regulation strategies (Aldao, 2013). Such extreme cases can shed greater light on the underlying psychological processes of emotion regulation (Anicich, Foulk, Osborne, Gale, & Schaerer, 2020;Eisenhardt, 1989;Pratt, 2000), as optimal strategies for regulating emotions might be driven by event intensity and differ from conventional wisdom in the emotion regulation literature (Troy, Shallcross, & Mauss, 2013). Second, we integrate new emotion regulation theorizing (Gross, 2015) with recent calls to model the co-occurring complexity of emotion regulation (Gabriel, Diefendorff, & Grandey, 2023;Grandey & Melloy, 2017) by using a profile approach to identify emergent combinations of emotion regulation strategies used for specific hostile events. ...

Getting Back to the “New Normal”: Autonomy Restoration During a Global Pandemic