Aayushi Dangol’s research while affiliated with University of Mary Washington and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (10)


Dataset Scale and Societal Consistency Mediate Facial Impression Bias in Vision-Language AI
  • Article

October 2024

·

2 Reads

·

Aayushi Dangol

·

·

Bill Howe

Multimodal AI models capable of associating images and text hold promise for numerous domains, ranging from automated image captioning to accessibility applications for blind and low-vision users. However, uncertainty about bias has in some cases limited their adoption and availability. In the present work, we study 43 CLIP vision-language models to determine whether they learn human-like facial impression biases, and we find evidence that such biases are reflected across three distinct CLIP model families. We show for the first time that the the degree to which a bias is shared across a society predicts the degree to which it is reflected in a CLIP model. Human-like impressions of visually unobservable attributes, like trustworthiness and sexuality, emerge only in models trained on the largest dataset, indicating that a better fit to uncurated cultural data results in the reproduction of increasingly subtle social biases. Moreover, we use a hierarchical clustering approach to show that dataset size predicts the extent to which the underlying structure of facial impression bias resembles that of facial impression bias in humans. Finally, we show that Stable Diffusion models employing CLIP as a text encoder learn facial impression biases, and that these biases intersect with racial biases in Stable Diffusion XL-Turbo. While pretrained CLIP models may prove useful for scientific studies of bias, they will also require significant dataset curation when intended for use as general-purpose models in a zero-shot setting.


Representation Bias of Adolescents in AI: A Bilingual, Bicultural Study

October 2024

·

4 Reads

·

1 Citation

Popular and news media often portray teenagers with sensationalism, as both a risk to society and at risk from society. As AI begins to absorb some of the epistemic functions of traditional media, we study how teenagers in two countries speaking two languages: 1) are depicted by AI, and 2) how they would prefer to be depicted. Specifically, we study the biases about teenagers learned by static word embeddings (SWEs) and generative language models (GLMs), comparing these with the perspectives of adolescents living in the U.S. and Nepal. We find English-language SWEs associate teenagers with societal problems, and more than 50% of the 1,000 words most associated with teenagers in the pretrained GloVe SWE reflect such problems. Given prompts about teenagers, 30% of outputs from GPT2-XL and 29% from LLaMA-2-7B GLMs discuss societal problems, most commonly violence, but also drug use, mental illness, and sexual taboo. Nepali models, while not free of such associations, are less dominated by social problems. Data from workshops with N=13 U.S. adolescents and N=18 Nepalese adolescents show that AI presentations are disconnected from teenage life, which revolves around activities like school and friendship. Participant ratings of how well 20 trait words describe teens are decorrelated from SWE associations, with Pearson's rho=.02, n.s. in English FastText and rho=.06, n.s. GloVe; and rho=.06, n.s. in Nepali FastText and rho=-.23, n.s. in GloVe. U.S. participants suggested AI could fairly present teens by highlighting diversity, while Nepalese participants centered positivity. Participants were optimistic that, if it learned from adolescents, rather than media sources, AI could help mitigate stereotypes. Our work offers an understanding of the ways SWEs and GLMs misrepresent a developmentally vulnerable group and provides a template for less sensationalized characterization.


Figure 2: Examples from the OMI dataset repository at https://github.com/jcpeterson/omi, used as stimuli in our research.
Figure 3: The similarity of CLIP bias to human bias is strongly correlated with human IRR, indicating that the societal consistency of a bias plays a significant role in whether a model learns it during semi-supervised pretraining.
Figure 4: CLIP models exhibit significant Spearman's ρ between Mean Model-Human Similarity and OMI IRR.
Figure 6: Scaling-2B CLIP models exhibit the greatest structural similarity to human facial impression biases.
Figure 8: The full cross-correlation matrix for OpenAI CLIP-ViT-L-14, the most commonly used CLIP model on the Huggingface Hub (Wolf et al. 2020) and the model with the highest Jaccard similarity to the OMI dataset based on statistically significant correlations.

+2

Dataset Scale and Societal Consistency Mediate Facial Impression Bias in Vision-Language AI
  • Preprint
  • File available

August 2024

·

24 Reads

Multimodal AI models capable of associating images and text hold promise for numerous domains, ranging from automated image captioning to accessibility applications for blind and low-vision users. However, uncertainty about bias has in some cases limited their adoption and availability. In the present work, we study 43 CLIP vision-language models to determine whether they learn human-like facial impression biases, and we find evidence that such biases are reflected across three distinct CLIP model families. We show for the first time that the the degree to which a bias is shared across a society predicts the degree to which it is reflected in a CLIP model. Human-like impressions of visually unobservable attributes, like trustworthiness and sexuality, emerge only in models trained on the largest dataset, indicating that a better fit to uncurated cultural data results in the reproduction of increasingly subtle social biases. Moreover, we use a hierarchical clustering approach to show that dataset size predicts the extent to which the underlying structure of facial impression bias resembles that of facial impression bias in humans. Finally, we show that Stable Diffusion models employing CLIP as a text encoder learn facial impression biases, and that these biases intersect with racial biases in Stable Diffusion XL-Turbo. While pretrained CLIP models may prove useful for scientific studies of bias, they will also require significant dataset curation when intended for use as general-purpose models in a zero-shot setting.

Download

Figure 3: Word associations with "teenager" in FastText are decorrelated from U.S. teens' ratings of their similarity to "teenager."
Representation Bias of Adolescents in AI: A Bilingual, Bicultural Study

August 2024

·

28 Reads

Popular and news media often portray teenagers with sensationalism, as both a risk to society and at risk from society. As AI begins to absorb some of the epistemic functions of traditional media, we study how teenagers in two countries speaking two languages: 1) are depicted by AI, and 2) how they would prefer to be depicted. Specifically, we study the biases about teenagers learned by static word embeddings (SWEs) and generative language models (GLMs), comparing these with the perspectives of adolescents living in the U.S. and Nepal. We find English-language SWEs associate teenagers with societal problems, and more than 50% of the 1,000 words most associated with teenagers in the pretrained GloVe SWE reflect such problems. Given prompts about teenagers, 30% of outputs from GPT2-XL and 29% from LLaMA-2-7B GLMs discuss societal problems, most commonly violence, but also drug use, mental illness, and sexual taboo. Nepali models, while not free of such associations, are less dominated by social problems. Data from workshops with N=13 U.S. adolescents and N=18 Nepalese adolescents show that AI presentations are disconnected from teenage life, which revolves around activities like school and friendship. Participant ratings of how well 20 trait words describe teens are decorrelated from SWE associations, with Pearson's r=.02, n.s. in English FastText and r=.06, n.s. in GloVe; and r=.06, n.s. in Nepali FastText and r=-.23, n.s. in GloVe. U.S. participants suggested AI could fairly present teens by highlighting diversity, while Nepalese participants centered positivity. Participants were optimistic that, if it learned from adolescents, rather than media sources, AI could help mitigate stereotypes. Our work offers an understanding of the ways SWEs and GLMs misrepresent a developmentally vulnerable group and provides a template for less sensationalized characterization.







Citations (4)


... There is an argument for including youth in RAI processes throughout the AI lifecycle "not in spite of [their] age, but specifically because of it, " one teen suggested in an interview with Time [9]. Due to youth 1) being early adopters of AI (i.e., among the first users and ways of using AI different from adults) [42], 2) having experiences growing up with AI-driven systems that are unique to this time of innovation (i.e., adults have had different experiences with technologies and do not have the same insight as current youth) [2,69], and 3) expressing an interest in contributing to AI fairness (i.e., wanting to engage with design and evaluation of AI) [62,64,66,68], youth are a core underexplored stakeholder in participatory RAI. Furthermore, youth have demonstrated great potential to engage in taking action toward more ethical AI. ...

Reference:

Investigating Youth AI Auditing
Representation Bias of Adolescents in AI: A Bilingual, Bicultural Study
  • Citing Article
  • October 2024

... This overtrust highlights the importance of fostering critical AI literacy-helping youth develop the skills to interrogate the sociotechnical implications of AI systems, alongside the technical capabilities. Emerging approaches to AI education, particularly for systemically marginalized learners, further emphasize empowerment and encourage the development of "techno-social change agency," such that youth are positioned to engage with and innovate toward equitable computing technologies [16,17,39,55,56,62]. ...

Mediating Culture: Cultivating Socio-cultural Understanding of AI in Children through Participatory Design
  • Citing Conference Paper
  • July 2024

... Each activity was designed to take approximately 10-20 minutes each to accommodate the schedules of the participants. The activities asked for recall and generative responses similar to previous ARC studies [23]. ...

Opportunities and Challenges for AI-Based Support for Speech-Language Pathologists
  • Citing Conference Paper
  • June 2024

... We also highlight qualitative studies of troubling data models with similar aims to ours, including grounded theory investigations of Michael Muller et al. on how data scientists work [30], and the recent ethnographic "autospeculation" method of Brian Kinnee et al. [23]. Finally, we note the broad literature on critical data literacy, which aims in part to help students understand the complex social factors behind data collections [12]. Our work is complementary with several of the aims of this area of scholarship. ...

Constructionist approaches to critical data literacy: A review
  • Citing Conference Paper
  • June 2023