Kurt Thomas’s research while affiliated with Mountain View College and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (53)


Help-seeking and Coping Strategies for Technology-facilitated Abuse Experienced by Youth
  • Article

May 2025

·

3 Reads

·

1 Citation

Proceedings of the ACM on Human-Computer Interaction

Diana Freed

·

Sunny Consolvo

·

Dan Cosley

·

[...]

·

Natalie N. Bazarova

Technology provides youth (ages 10--17) with near-constant opportunities for learning, communication, and self-expression. It can also expose them to technology-facilitated abuse: harassment, coercion, fraud, and more. The ability of youth to navigate such abuse is crucial for their well-being and development. A recent advisory by the U.S. Surgeon General called for better support of youth, including that youth should ''reach out for help.'' However, little is known about how youth seek help or otherwise cope with technology-facilitated abuse. Through a qualitative study in the U.S., we examine how youth engage in self-reliance, seek help from others, and how others seek help on a youth's behalf. We discuss these strategies and outline opportunities for how the HCI community can better support youth who experience technology-facilitated abuse.



Supporting the Digital Safety of At-Risk Users: Lessons Learned from 9+ Years of Research & Training

February 2025

·

19 Reads

ACM Transactions on Computer-Human Interaction

Creating information technologies intended for broad use that allow everyone to participate safely online—which we refer to as inclusive digital safety —requires understanding and addressing the digital-safety needs of a diverse range of users who face elevated risk of technology-facilitated attacks or disproportionate harm from such attacks—i.e., at-risk users . This paper draws from more than nine years of our work at Google to understand and support the digital safety of at-risk users—including survivors of intimate partner abuse, people involved with political campaigns, content creators, youth, and more—in technology intended for broad use. Among our learnings is that designing for inclusive digital safety across widely varied user needs and dynamic contexts is a wicked problem with no ‘correct’ solution. Given this, we describe frameworks and design principles we have developed to help make at-risk research findings practically applicable to technologies intended for broad use and lessons we have learned about communicating them to practitioners.



Fig. 3: CDF of the sample sizes in our benchmark dataset.
Fig. 4: Architecture of MAGIKA. The input and the output are depicted in blue and green, respectively. The model's layers are in yellow. The layers in purple are used only in training. The numbers next to the layers' names indicate the size of their outputs.
Fig. 5: Validation loss and validation accuracy as the training progresses in terms of the number of epochs. We find accuracy increases up to around 30 epochs.
Fig. 6: Average F1 Score (after one epoch of training) with increasing number of samples per content type, across binary, text, overall content types.
Fig. 7: Precision-recall curve for a fixed threshold Θ applied to all content type predictions. Note that the truncated scale for precision and recall is different.

+1

Magika: AI-Powered Content-Type Detection
  • Preprint
  • File available

September 2024

·

91 Reads

The task of content-type detection -- which entails identifying the data encoded in an arbitrary byte sequence -- is critical for operating systems, development, reverse engineering environments, and a variety of security applications. In this paper, we introduce Magika, a novel AI-powered content-type detection tool. Under the hood, Magika employs a deep learning model that can execute on a single CPU with just 1MB of memory to store the model's weights. We show that Magika achieves an average F1 score of 99% across over a hundred content types and a test set of more than 1M files, outperforming all existing content-type detection tools today. In order to foster adoption and improvements, we open source Magika under an Apache 2 license on GitHub and make our model and training pipeline publicly available. Our tool has already seen adoption by the Gmail email provider for attachment scanning, and it has been integrated with VirusTotal to aid with malware analysis. We note that this paper discusses the first iteration of Magika, and a more recent version already supports more than 200 content types. The interested reader can see the latest development on the Magika GitHub repository, available at https://github.com/google/magika.

Download

Privacy Risks of General-Purpose AI Systems: A Foundation for Investigating Practitioner Perspectives

July 2024

·

57 Reads

The rise of powerful AI models, more formally General-Purpose AI Systems\textit{General-Purpose AI Systems} (GPAIS), has led to impressive leaps in performance across a wide range of tasks. At the same time, researchers and practitioners alike have raised a number of privacy concerns, resulting in a wealth of literature covering various privacy risks and vulnerabilities of AI models. Works surveying such risks provide differing focuses, leading to disparate sets of privacy risks with no clear unifying taxonomy. We conduct a systematic review of these survey papers to provide a concise and usable overview of privacy risks in GPAIS, as well as proposed mitigation strategies. The developed privacy framework strives to unify the identified privacy risks and mitigations at a technical level that is accessible to non-experts. This serves as the basis for a practitioner-focused interview study to assess technical stakeholder perceptions of privacy risks and mitigations in GPAIS.


Figure 3: Initial prompt derived from Google's public policy around census and election misinformation. Across all our experiments, this policy language remains static.
Figure 4: A few-shot prompt variant that includes both an example policy-relevant comment and answer, and keyword context. The comment under evaluation appears after all examples.
Figure 5: Accuracy of text-unicorn for our hand-picked fewshot prompt variant. We segment our evaluation corpus into buckets of 0-9 characters, 10-19 characters, and so on up to 100+ characters. For each sample, we display error margins for a confidence level of 95%. (Note the truncated Y-axis.)
Figure 6: Precision vs. recall for our adaptive few-shot example policy prompt and text-unicorn. Using the probability of tokens predicted by an LLM, we can flexibly alter whether a prompt favors precision or recall.
Supporting Human Raters with the Detection of Harmful Content using Large Language Models

June 2024

·

46 Reads

In this paper, we explore the feasibility of leveraging large language models (LLMs) to automate or otherwise assist human raters with identifying harmful content including hate speech, harassment, violent extremism, and election misinformation. Using a dataset of 50,000 comments, we demonstrate that LLMs can achieve 90% accuracy when compared to human verdicts. We explore how to best leverage these capabilities, proposing five design patterns that integrate LLMs with human rating, such as pre-filtering non-violative content, detecting potential errors in human rating, or surfacing critical context to support human rating. We outline how to support all of these design patterns using a single, optimized prompt. Beyond these synthetic experiments, we share how piloting our proposed techniques in a real-world review queue yielded a 41.5% improvement in optimizing available human rater capacity, and a 9--11% increase (absolute) in precision and recall for detecting violative content.


Strategies attempted before posting on Reddit.
Understanding Help-Seeking and Help-Giving on Social Media for Image-Based Sexual Abuse

June 2024

·

130 Reads

Image-based sexual abuse (IBSA), like other forms of technology-facilitated abuse, is a growing threat to people's digital safety. Attacks include unwanted solicitations for sexually explicit images, extorting people under threat of leaking their images, or purposefully leaking images to enact revenge or exert control. In this paper, we explore how people seek and receive help for IBSA on social media. Specifically, we identify over 100,000 Reddit posts that engage relationship and advice communities for help related to IBSA. We draw on a stratified sample of 261 posts to qualitatively examine how various types of IBSA unfold, including the mapping of gender, relationship dynamics, and technology involvement to different types of IBSA. We also explore the support needs of victim-survivors experiencing IBSA and how communities help victim-survivors navigate their abuse through technical, emotional, and relationship advice. Finally, we highlight sociotechnical gaps in connecting victim-survivors with important care, regardless of whom they turn to for help.


Figure 1: Example giveaway scam landing pages promoted via Twitter. Scammers impersonate popular personalities including Brad Garlinghouse (the Ripple CEO) and Elon Musk.
Figure 2: Example livestream containing a giveaway scam. The video playing is of Brad Garlinghouse and the scam website is linked to in both the chat and the embedded QR code.
Figure 5: Effectiveness of keywords
Give and Take: An End-To-End Investigation of Giveaway Scam Conversion Rates

May 2024

·

168 Reads

Scams -- fraudulent schemes designed to swindle money from victims -- have existed for as long as recorded history. However, the Internet's combination of low communication cost, global reach, and functional anonymity has allowed scam volumes to reach new heights. Designing effective interventions requires first understanding the context: how scammers reach potential victims, the earnings they make, and any potential bottlenecks for durable interventions. In this short paper, we focus on these questions in the context of cryptocurrency giveaway scams, where victims are tricked into irreversibly transferring funds to scammers under the pretense of even greater returns. Combining data from Twitter, YouTube and Twitch livestreams, landing pages, and cryptocurrency blockchains, we measure how giveaway scams operate at scale. We find that 1 in 1000 scam tweets, and 4 in 100,000 livestream views, net a victim, and that scammers managed to extract nearly \$4.62 million from just hundreds of victims during our measurement window.



Citations (38)


... YouTube has been used to study a wide array of crucial societal problems like online hate [1,17], accessibility [13], pseudoscientific misinformation [18], online scams [10], and child exposure to inappropriate content [5,9,16]. Much of this work has made use of the YouTube Data API, which offers several endpoints. ...

Reference:

I'm Sorry Dave, I'm Afraid I Can't Return That: On YouTube Search API Use in Research
Give and Take: An End-To-End Investigation of Giveaway Scam Conversion Rates
  • Citing Conference Paper
  • November 2024

... Therefore, before post-processing, we sampled 100,000 tweets from the 21M tweets and used Google's Perspective API 2 to identify potentially offensive tweets to annotate. Researchers have used Perspective API to detect toxic comments (Wulczyn, Thain, and Dixon 2017) in YouTube (Obadimu et al. 2019), to understand behaviors of toxic account on Reddit (Kumar et al. 2023), and to filter potentially offensive tweets in COVID-19 dataset (Liao et al. 2023). Following (Liao et al. 2023), we use the Perspective API to filter potentially offensive tweets for labeling. ...

Understanding the Behaviors of Toxic Accounts on Reddit
  • Citing Conference Paper
  • April 2023

... Recent evidence highlights that adolescents facing emotional dysregulation, impulsivity, or alexithymia show increased vulnerability to externalizing behaviors and non-suicidal self-injury (36)(37)(38)(39)(40). These traits tend to amplify under conditions of peer instability and poor affective scaffolding (41)(42)(43)(44)(45). Personality traits such as sensation seeking and low harm avoidance have been associated with a preference for highstimulation environments, which often include risky group dynamics and nonconforming behaviors (46)(47)(48)(49). ...

Understanding Digital-Safety Experiences of Youth in the U.S.
  • Citing Conference Paper
  • April 2023

... Hate Speech and Hate Campaigns on Web Communities. Hate speech towards different target identity groups such as race, ethnicity, gender, religion, disability, and sexual orientation has a long-standing history on the Internet [11,12,50,61,70,72,74,82,84,85]. According to a report by the Anti-Defamation League (ADL), 33% of adults experienced hate and harassment in 2023, up from 23% in 2022 [11]. ...

“There’s so much responsibility on users right now:” Expert Advice for Staying Safer From Hate and Harassment
  • Citing Conference Paper
  • April 2023

... Mirai, a family of malware that infects IoT devices and turns these devices into a Distributed Denial-of-Service (DDoS) botnet, was publicly released on September 30, 2016, and grew to a peak of 600 k infections, with massive DDoS attacks against several high-pro¯le targets. 2 In October 2016, hackers infected numerous IoT devices to launch DDoS attacks on Dyn, a DNS service provider, leaving much of the Internet inaccessible. 7 The Mirai botnet has evolved and continues with numerous infamous variants such as Okiru, Satori, Masuta, and PureMasuta emerging over the years. ...

Understanding the Mirai Botnet Understanding the Mirai Botnet

... In the information-security literature, activists are commonly referred to as a 'high-risk' or 'at-risk' population, along with other groups such as, refugees, LGBTQAI+ and survivors of intimate partner abuse [78]. However, the emerging body of research on activism in security-related scholarship points to the particularities -and temporalities -of specific activist contexts. ...

SoK: A Framework for Unifying At-Risk User Research
  • Citing Conference Paper
  • May 2022

... We speculate that digital skills facilitate content creation, increasing visibility to both peers and strangers and, consequently, exposure to unexpected sexts-explaining the lack of association in infrequent cases but a positive relationship for monthly occurrences. A U.S. survey of 135 content creators found that ~70% experienced sexual harassment, bullying, or identity attacks more than rarely (Thomas et al., 2022). ...

“It’s common and a part of being a content creator”: Understanding How Creators Experience and Cope with Hate and Harassment Online
  • Citing Conference Paper
  • April 2022

... While social media enables people around the world to communicate with each other, these interactions can also lead to interpersonal conflicts [58,77]. Sometimes these conflicts can escalate into online harassment and cause great emotional harm [125]. A 2021 survey found that 41% of Americans reported personally experiencing harassment or bullying online [74]. ...

SoK: Hate, Harassment, and the Changing Landscape of Online Abuse
  • Citing Conference Paper
  • May 2021

... Despite several weaknesses of the protocol, billions of users regularly use email messages for business and personal exchange [10]. Due to its popularity, email is a constant magnet for cybercrime, serving as a vehicle for transporting unsolicited, fraudulent and malicious content, which ranges from spam and phishing attempts to targeted attacks and malware distribution [e.g., 8,12,14,30]. These activities benefit from the lack of security mechanisms in the original protocols that cannot establish the authenticity of senders and content by itself. ...

Who is targeted by email-based phishing and malware?: Measuring factors that differentiate risk
  • Citing Conference Paper
  • October 2020

... The clickable link guides the victim of the PSC campaign to a prepared landing page, where user behavior is tracked. These landing pages are typically login pages akin to the ones real attackers would use [40,49]. Measuring whether a user submits login data to a login field on these landing pages is especially interesting. ...

Sunrise to Sunset: Analyzing the End-to-end Life Cycle and E ectiveness of Phishing Attacks at Scale