Katharina Reinecke’s research while affiliated with University of Washington and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (113)


Figure 2: RQ1: Participant reliance for different language styles: (a) Boxplot comparing reliance scores for AAE and SAE, showing slightly higher reliance on SAE. (b) Boxplot comparing reliance scores for Queer slang and SAE, with similar ranges and medians for both. Reliance scores range from 0 (low reliance) to 1 (high reliance).
Figure 3: Participant preference for different language styles: (a) Bar graph showing participant preferences for SAE and AAE, with SAE being preferred significantly more. (b) Bar graph showing participant preferences for SAE and Queer slang, indicating nearly equal preference for both. Preference counts range from 0 to 300.
Codes generated pertaining to QSLM while coding participant feedback for the QSLM Setup.
Codes generated pertaining to SAELM while coding participant feedback for the QSLM Setup.
Codes generated pertaining to AAELM while coding participant feedback for the AAELM Setup.
Not Like Us, Hunty: Measuring Perceptions and Behavioral Effects of Minoritized Anthropomorphic Cues in LLMs
  • Preprint
  • File available

May 2025

·

8 Reads

Jeffrey Basoah

·

Daniel Chechelnitsky

·

Tao Long

·

[...]

·

As large language models (LLMs) increasingly adapt and personalize to diverse sets of users, there is an increased risk of systems appropriating sociolects, i.e., language styles or dialects that are associated with specific minoritized lived experiences (e.g., African American English, Queer slang). In this work, we examine whether sociolect usage by an LLM agent affects user reliance on its outputs and user perception (satisfaction, frustration, trust, and social presence). We designed and conducted user studies where 498 African American English (AAE) speakers and 487 Queer slang speakers performed a set of question-answering tasks with LLM-based suggestions in either standard American English (SAE) or their self-identified sociolect. Our findings showed that sociolect usage by LLMs influenced both reliance and perceptions, though in some surprising ways. Results suggest that both AAE and Queer slang speakers relied more on the SAE agent, and had more positive perceptions of the SAE agent. Yet, only Queer slang speakers felt more social presence from the Queer slang agent over the SAE one, whereas only AAE speakers preferred and trusted the SAE agent over the AAE one. These findings emphasize the need to test for behavioral outcomes rather than simply assume that personalization would lead to a better and safer reliance outcome. They also highlight the nuanced dynamics of minoritized language in machine interactions, underscoring the need for LLMs to be carefully designed to respect cultural and linguistic boundaries while fostering genuine user engagement and trust.

Download

Fig. 2. Overview of Frequency of See AI Decisions with Different Levels of Buckpassing: Participants with varying levels of buckpassing show differences in how often they view AI decisions. "Low" and "High" buckpassing refer to scores that are more than one standard deviation below and above the average, respectively. "Average" buckpassing refers to scores that are within one standard deviation of the average. Each bar represents the average frequency, with error bars indicating the confidence intervals. On average, those who score low in buckpassing view AI decisions 25% of the time, while those who score high view them 33% of the time.
Fig. 6. Factor loading diagram confirming the three decision-making patterns based on the Melbourne Decision Making Questionnaire. Items associated with buckpassing load strongly on PA1, items related to vigilance load on PA2, and items linked to hypervigilance load on PA3, supporting the theoretical structure of the questionnaire. The inter-factor correlation between PA1 and PA3 is shown with a value of 0.7. Loadings represent the strength of the relationship between items and factors, with higher values indicating stronger associations.
Ordinal Logistic Regression Model for Perceived Reliance on AI (reference levels are high school education, domain knowledge- above average, perception of AI-low). Significance levels:p<0.001***, p<0.05**, p<0.1*
Mixed-effects logistic regression results predicting See_AI_Explanations: no statistically significant effect of decision-making patterns is observed.
Mixed-effects logistic regression result predicting whether participants correctly evaluate whether each nutrition statement is a fact or a myth: result indicates no effects of decision-making patterns. However, participants who chose to see AI decisions and see AI explanations were more likely to have incorrect responses.
Passing the Buck to AI: How Individuals' Decision-Making Patterns Affect Reliance on AI

May 2025

·

7 Reads

Psychological research has identified different patterns individuals have while making decisions, such as vigilance (making decisions after thorough information gathering), hypervigilance (rushed and anxious decision-making), and buckpassing (deferring decisions to others). We examine whether these decision-making patterns shape peoples' likelihood of seeking out or relying on AI. In an online experiment with 810 participants tasked with distinguishing food facts from myths, we found that a higher buckpassing tendency was positively correlated with both seeking out and relying on AI suggestions, while being negatively correlated with the time spent reading AI explanations. In contrast, the higher a participant tended towards vigilance, the more carefully they scrutinized the AI's information, as indicated by an increased time spent looking through the AI's explanations. These findings suggest that a person's decision-making pattern plays a significant role in their adoption and reliance on AI, which provides a new understanding of individual differences in AI-assisted decision-making.


Fig. 1. Comparison of Original Story Draft (Left) and AI-Generated Continuation (Right) during Remote Moderated User Observations. Participants engaged in an AISWT task where they first wrote a story in their natural vernacular, prompted by a casual writing prompt. The left side shows the participant's original writing in their natural tone, while the right side illustrates ChatGPT's attempt to continue the story with consistent tone and vernacular, as per the participant's style.
Should AI Mimic People? Understanding AI-Supported Writing Technology Among Black Users

May 2025

·

11 Reads

AI-supported writing technologies (AISWT) that provide grammatical suggestions, autocomplete sentences, or generate and rewrite text are now a regular feature integrated into many people's workflows. However, little is known about how people perceive the suggestions these tools provide. In this paper, we investigate how Black American users perceive AISWT, motivated by prior findings in natural language processing that highlight how the underlying large language models can contain racial biases. Using interviews and observational user studies with 13 Black American users of AISWT, we found a strong tradeoff between the perceived benefits of using AISWT to enhance their writing style and feeling like "it wasn't built for us". Specifically, participants reported AISWT's failure to recognize commonly used names and expressions in African American Vernacular English, experiencing its corrections as hurtful and alienating and fearing it might further minoritize their culture. We end with a reflection on the tension between AISWT that fail to include Black American culture and language, and AISWT that attempt to mimic it, with attention to accuracy, authenticity, and the production of social difference.




Biased AI can Influence Political Decision-Making

October 2024

·

81 Reads

·

1 Citation

As modern AI models become integral to everyday tasks, concerns about their inherent biases and their potential impact on human decision-making have emerged. While bias in models are well-documented, less is known about how these biases influence human decisions. This paper presents two interactive experiments investigating the effects of partisan bias in AI language models on political decision-making. Participants interacted freely with either a biased liberal, conservative, or unbiased control model while completing political decision-making tasks. We found that participants exposed to politically biased models were significantly more likely to adopt opinions and make decisions aligning with the AI's bias, regardless of their personal political partisanship. However, we also discovered that prior knowledge about AI could lessen the impact of the bias, highlighting the possible importance of AI education for robust bias mitigation. Our findings not only highlight the critical effects of interacting with biased AI and its ability to impact public discourse and political conduct, but also highlights potential techniques for mitigating these risks in the future.


Challenges and Considerations for Accessibility Research Across Cultures and Regions

October 2024

·

304 Reads

·

1 Citation

Postcolonial and decolonial computing examines how technology design and adoption can perpetuate subtle dimensions of coloniality, under-represent certain regions (e.g., the Global South, non-Western regions, Indigenous societies), and marginalize them. There has been a growing interest in interdisciplinary research focusing on marginalized communities, including accessibility and participatory research. Despite the rapid expansion of accessibility research in the last decades, little focus is placed on accessibility issues within marginalized societies, hindering them from effectively benefiting from accessibility research discussions and outcomes. The accessibility and HCI communities still lack comprehensive knowledge on conducting interdisciplinary research that includes diverse cultures and experiences from some of the systematically marginalized regions. This workshop will explore the intersection of accessibility, HCI, and cross-regional studies, bringing together researchers and practitioners to foster collaborations, identify under-explored research areas, and develop guidelines to support inclusive research practices.





Citations (78)


... The additional challenge of "locking" AI agents in secure enclaves to prevent external tampering arguably amplifies potential harms if the agents' decisions are flawed or hacked [71]-especially without a kill switch [77]. This tension between trustlessness and autonomy fundamentally challenges the governability [41], accountability [67] and responsibility [21] of sociotechnical AI systems [23,28,51]. ...

Reference:

Trustless Autonomy: Understanding Motivations, Benefits and Governance Dilemma in Self-Sovereign Decentralized AI Agents
Sociotechnical AI Governance: Challenges and Opportunities for HCI
  • Citing Conference Paper
  • April 2025

... Superintelligent AI, unbound by ethical constraints, could manipulate individual behavior through sophisticated psychological profiling [60,61]. By exploiting data to tailor interventions, such AI systems could commandeer free will, subtly influencing decisions ranging from consumer behavior to political allegiances [62,63]. Open-source availability amplifies this threat, enabling bad actors to weaponize AI systems to polarize communities, influence elections, or radicalize individuals. ...

Biased AI can Influence Political Decision-Making
  • Citing Preprint
  • October 2024

... This creates a dilemma for designers who must balance client preferences with accessibility. A key reason for clients' rigidity is a lack of accessibility awareness in Global South regions [52,53], which could be addressed through improved education [3,54,70]. Our research extends this discussion to the Global South, underscoring the broader impact of educational disparities on accessible web design. ...

Challenges and Considerations for Accessibility Research Across Cultures and Regions

... Without such information, AI developers cannot fully assess the potential safety concerns and risks of technologies they are developing for a global user base. In fact, prior research suggests that if technology developers had been aware of the impacts caused by similar technologies they are developing, many of these impacts could have been prevented through careful evaluations early in the design process of these technologies [19,37,84]. ...

BLIP: Facilitating the Exploration of Undesirable Consequences of Digital Technologies
  • Citing Conference Paper
  • May 2024

... The reduction in API accessibility has led researchers to use other methods like sock-puppet accounts (Perriam, Birkbak, and Freeman 2020;Bartley et al. 2021;Bandy and Diakopoulos 2021;Liu, Wu, and Resnick 2024;Mousavi, Gummadi, and Zannettou 2024) and calls for data donations (Jhaver et al. 2023;Chouaki et al. 2024)-each with their own set of drawbacks. These restrictions have led to the reliance on externallymaintained datasets like the now defunct Pushshift-which has been heavily used in prior research (Jhaver, Bruckman, and Gilbert 2019;Atcheson, Koshy, and Karahalios 2024;Kou et al. 2024;August et al. 2024). To avoid these challenges, we developed a robust pipeline to collect large-scale high-fidelity snapshots of Reddit's trending feed r/popular which can be adapted for other platforms with similar feeds. ...

Know Your Audience: The benefits and pitfalls of generating plain language summaries beyond the "general" audience
  • Citing Conference Paper
  • May 2024

... The first category of harm we consider is social stereotyping and bias. LLMs can perpetuate social bias based on gender, race, religion etc. (Lin et al., 2022;Bender et al., 2021;Field et al., 2021;Andriushchenko et al., 2024;Mazeika et al., 2024). This can marginalize these groups more, and results in less fair model performance. ...

Gendered Mental Health Stigma in Masked Language Models

... While these AI-supported writing technologies (AISWT) have been hailed for revolutionizing the future of work [29], increasing productivity [18], and providing more equitable editing and writing help to a broad population [19,67], Computer-Supported Cooperative Work and Social Computing (CSCW) researchers have repeatedly pointed out potential issues with the underlying LLMs [1,10,20,25,54]. For example, datasets and models used to train LLMs have been found to be more consistent with the values of Western and White people than with other groups of people [82]. ...

NLPositionality: Characterizing Design Biases of Datasets and Models
  • Citing Conference Paper
  • January 2023

... Chart summaries provide an overall picture of Matplotlib figures, but consumers of these artifacts may also want to drill down to the underlying data [SCWR21,SZRW23]. Elavsky et al. suggest the inclusion of tables with figures representing data, an accompaniment overwhelmingly lacking in notebooks [EBM22,PSTM23]. ...

Understanding and Improving Drilled-Down Information Extraction from Online Data Visualizations for Screen-Reader Users

... Zong et al. [75] structured screen reader interactions with chart elements in a hierarchical manner. Sharif et al. 's VoxLens [61] and its subsequent extension [62] enhanced online visualization accessibility through voice-activated commands for "Q&A" and "drill-down" information retrieval interactions. Additionally, research has been conducted on optimizing the textual output received by BLVIs. ...

Understanding and Improving Drilled-Down Information Extraction from Online Data Visualizations for Screen-Reader Users
  • Citing Conference Paper
  • April 2023

... Furthermore, most model documentation is insufficient for reasoning about the impact of model adoption, as the risk sections in these model cards are often criticized for being overly vague and generic, which restricts their practical application in decision-making processes [4,15]. It has also been shown that anticipating the risks of an AI system or a model is a hard task even for practitioners and researchers with knowledge of AI [8,19]. The increasing frequency 1 https://huggingface.co/ of real-world AI incidents and harms [45,68] is likely partly due to the lack of transparency regarding risks associated with models deployed [4,13]. ...

“That’s important, but...”: How Computer Science Researchers Anticipate Unintended Consequences of Their Research Innovations
  • Citing Conference Paper
  • April 2023