Emilio Ferrara’s research while affiliated with University of Southern California and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (382)


Figure 2: Comparison of different metrics (precision, recall, F1 score, and ROC AUC) across various (Llama) model sizes. The graph illustrates how each metric changes as the number of model parameters increases. The exact numerical results are also available in Table 7.
Statistics of the dataset.
Main evaluation results of different baseline models and framework variants. The best results for each metric are highlighted in bold. The horizontal line distinguishes the baseline models from the framework variants. The best results for each metric are highlighted in bold.
Evaluation results of a framework variant with re- spect to different temperature values ranging from 0 to 1. The best results for each metric are highlighted in bold. Llama 3.1 8B was used in these experiments.
Evaluation results of a framework variant with re- spect to different numbers of prompting examples ranging from 1 to 8. The best results for each metric are highlighted in bold. Llama 3.1 8B was used in these experiments.

+2

Network-informed Prompt Engineering against Organized Astroturf Campaigns under Extreme Class Imbalance
  • Preprint
  • File available

January 2025

·

2 Reads

·

Heng Ping

·

·

[...]

·

Detecting organized political campaigns is of paramount importance in fighting against disinformation on social media. Existing approaches for the identification of such organized actions employ techniques mostly from network science, graph machine learning and natural language processing. Their ultimate goal is to analyze the relationships and interactions (e.g. re-posting) among users and the textual similarities of their posts. Despite their effectiveness in recognizing astroturf campaigns, these methods face significant challenges, notably the class imbalance in available training datasets. To mitigate this issue, recent methods usually resort to data augmentation or increasing the number of positive samples, which may not always be feasible or sufficient in real-world settings. Following a different path, in this paper, we propose a novel framework for identifying astroturf campaigns based solely on large language models (LLMs), introducing a Balanced Retrieval-Augmented Generation (Balanced RAG) component. Our approach first gives both textual information concerning the posts (in our case tweets) and the user interactions of the social network as input to a language model. Then, through prompt engineering and the proposed Balanced RAG method, it effectively detects coordinated disinformation campaigns on X (Twitter). The proposed framework does not require any training or fine-tuning of the language model. Instead, by strategically harnessing the strengths of prompt engineering and Balanced RAG, it facilitates LLMs to overcome the effects of class imbalance and effectively identify coordinated political campaigns. The experimental results demonstrate that by incorporating the proposed prompt engineering and Balanced RAG methods, our framework outperforms the traditional graph-based baselines, achieving 2x-3x improvements in terms of precision, recall and F1 scores.

Download





Figure 2: Hashtag Co-Occurrence Graph
Figure 3: Hashtag usage trends by category from November 2023 to October 2024, showing peak activity for Republicans, Democrats, Neutral, and Third-Party hashtags over time.
Summary statistics of the dataset.
Tracking the 2024 US Presidential Election Chatter on TikTok: A Public Multimodal Dataset

December 2024

·

35 Reads

This paper presents the TikTok 2024 U.S. Presidential Election Dataset, a large-scale, resource designed to advance research into political communication and social media dynamics. The dataset comprises 3.14 million videos published on TikTok between November 1, 2023, and October 16, 2024, encompassing video ids and transcripts. Data collection was conducted using the TikTok Research API with a comprehensive set of election-related keywords and hashtags, supplemented by third-party tools to address API limitations and expand content coverage, enabling analysis of hashtag co-occurrence networks that reveal politically aligned hashtags based on ideological affiliations, the evolution of top hashtags over time, and summary statistics that highlight the dataset's scale and richness. This dataset offers insights into TikTok's role in shaping electoral discourse by providing a multimodal view of election-related content. It enables researchers to explore critical topics such as coordinated messaging, misinformation spread, audience engagement, and linguistic trends. The TikTok 2024 U.S. Presidential Election Dataset is publicly available and aims to contribute to the broader understanding of social media's impact on democracy and public opinion.


Safe Spaces or Toxic Places? Content Moderation and Social Dynamics of Online Eating Disorder Communities

December 2024

·

11 Reads

Social media platforms have become critical spaces for discussing mental health concerns, including eating disorders. While these platforms can provide valuable support networks, they may also amplify harmful content that glorifies disordered cognition and self-destructive behaviors. While social media platforms have implemented various content moderation strategies, from stringent to laissez-faire approaches, we lack a comprehensive understanding of how these different moderation practices interact with user engagement in online communities around these sensitive mental health topics. This study addresses this knowledge gap through a comparative analysis of eating disorder discussions across Twitter/X, Reddit, and TikTok. Our findings reveal that while users across all platforms engage similarly in expressing concerns and seeking support, platforms with weaker moderation (like Twitter/X) enable the formation of toxic echo chambers that amplify pro-anorexia rhetoric. These results demonstrate how moderation strategies significantly influence the development and impact of online communities, particularly in contexts involving mental health and self-harm.


Results of IOHunter in terms of Macro-F1 for the cross-country detection task. For each country c, IOHunter is pretrained on data from all other countries (i.e., Only PreTraining) and then tested on c's test set. IOHunter can be further fine-tuned on a tiny percentage (0.1%) of c training set. We compare the model trained on the 0.1% of c's training set.
IOHunter: Graph Foundation Model to Uncover Online Information Operations

December 2024

·

5 Reads

Social media platforms have become vital spaces for public discourse, serving as modern agor\'as where a wide range of voices influence societal narratives. However, their open nature also makes them vulnerable to exploitation by malicious actors, including state-sponsored entities, who can conduct information operations (IOs) to manipulate public opinion. The spread of misinformation, false news, and misleading claims threatens democratic processes and societal cohesion, making it crucial to develop methods for the timely detection of inauthentic activity to protect the integrity of online discourse. In this work, we introduce a methodology designed to identify users orchestrating information operations, a.k.a. \textit{IO drivers}, across various influence campaigns. Our framework, named \texttt{IOHunter}, leverages the combined strengths of Language Models and Graph Neural Networks to improve generalization in \emph{supervised}, \emph{scarcely-supervised}, and \emph{cross-IO} contexts. Our approach achieves state-of-the-art performance across multiple sets of IOs originating from six countries, significantly surpassing existing approaches. This research marks a step toward developing Graph Foundation Models specifically tailored for the task of IO detection on social media platforms.


Hybrid Forecasting of Geopolitical Events

December 2024

·

55 Reads

Sound decision-making relies on accurate prediction for tangible outcomes ranging from military conflict to disease outbreaks. To improve crowdsourced forecasting accuracy, we developed SAGE, a hybrid forecasting system that combines human and machine generated forecasts. The system provides a platform where users can interact with machine models and thus anchor their judgments on an objective benchmark. The system also aggregates human and machine forecasts weighting both for propinquity and based on assessed skill while adjusting for overconfidence. We present results from the Hybrid Forecasting Competition (HFC) - larger than comparable forecasting tournaments - including 1085 users forecasting 398 real-world forecasting problems over eight months. Our main result is that the hybrid system generated more accurate forecasts compared to a human-only baseline which had no machine generated predictions. We found that skilled forecasters who had access to machine-generated forecasts outperformed those who only viewed historical data. We also demonstrated the inclusion of machine-generated forecasts in our aggregation algorithms improved performance, both in terms of accuracy and scalability. This suggests that hybrid forecasting systems, which potentially require fewer human resources, can be a viable approach for maintaining a competitive level of accuracy over a larger number of forecasting questions.


Word embedding for social sciences: an interdisciplinary survey

December 2024

·

17 Reads

Machine learning models learn low-dimensional representations from complex high-dimensional data. Not only computer science but also social science has benefited from the advancement of these powerful tools. Within such tools, word embedding is one of the most popular methods in the literature. However, we have no particular documentation of this emerging trend because this trend overlaps different social science fields. To well compile this fragmented knowledge, we survey recent studies that apply word embedding models to human behavior mining. Our taxonomy built on the surveyed article provides a concise but comprehensive overview of this emerging trend of intersection between computer science and social science and guides scholars who are going to navigate the use of word embedding algorithms in their voyage of social science research.


Citations (52)


... In Europe, and specifically Germany, Telegram became the platform for spreading false information about pandemics, conspiracy theories and The case of romantic relationships anti-European thinking (Kiess 2023;Zehring and Domahidi 2023). In the US, Telegram became popular in extremist circles (both far-right and far-left) (Blas et al. 2024;Davey and Weinberg 2021;Walther and McCoy 2021) as they were banned from established social media platforms (Rogers 2020). ...

Reference:

The case of romantic relationships: analysis of the use of metaphorical frames with ‘traditional family’ and related terms in political Telegram posts in three countries and three languages
Unearthing a Billion Telegram Posts about the 2024 U.S. Presidential Election: Development of a Public Dataset
  • Citing Preprint
  • January 2025

... Of the kinds of sources used in news articles, the majority are either Quotes, 40%, (i.e., information derived directly from people the reporter spoke to), or Press Reports, 23% (i.e., information from other news articles). We obtain these labels by scoring our documents using models trained and described by Spangher et al. (2024a). As shown in Table 5, the use of Quotes, or person-derived information, is correlated more with Contradictory articles. ...

Explaining Mixtures of Sources in News Articles
  • Citing Conference Paper
  • January 2024

... The integrity of these online spaces is paramount, given their significant role in shaping public opinion and influencing societal outcomes, such as elections or public health interventions (Starbird 2019;Ferrara 2015;Nogara et al. 2022). However, these platforms are increasingly vulnerable to state-sponsored Information Operations (IOs), which seek to manipulate narratives, spread disinformation, and foster division through the promotion of hate speech and other harmful content (Badawy, Ferrara, and Lerman 2018;Zannettou et al. 2019;Suresh et al. 2024;Minici et al. 2024). The proliferation of such campaigns poses a significant threat to democratic processes, highlighting the urgent need for robust methods to Copyright © 2025, Association for the Advancement of Artificial Intelligence (www.aaai.org). ...

Uncovering coordinated cross-platform information operations: Threatening the integrity of the 2024 U.S. presidential election

First Monday

... However, maintaining safe, welcoming, and inclusive online environments has proven challenging. Online communities that indoctrinate users into extreme ideologies [2,3,4,5,6] or glorify self-harm [7] are especially challenging to moderate, as users often evade moderation through the use of coded language, insider jargon, and intentional misspellings [8,9,10]. While some social media platforms have made progress in identifying and moderating online speech that violates community norms, such as hate speech and personal attacks [11,12], others have adopted a more laissez-faire approach. ...

Attraction to politically extreme users on social media

PNAS Nexus

... Using human-decision making as a basis of comparison for LLMs is standard, even in creative, open-ended tasks: e.g. story-planning (Mostafazadeh et al., 2016), computational journalism (Spangher et al., 2024b(Spangher et al., , 2023(Spangher et al., , 2022 and others (Tian et al., 2023a). If this problem were unlearnable (e.g. ...

Tracking the Newsworthiness of Public Documents
  • Citing Conference Paper
  • January 2024

... Previous research has tried converting personal health information of older adults into a narrative format, but is often limited by rigid, rule-based systems that lack the ability to perform complex, dynamic sensemaking (Metaxas et al. 2007;Chung, Ozkaynak, and Demiris 2017;Shahid, Saguna, andÅhlund 2022). With the advent of Large Language Models (LLMs), there is potential to leverage their commonsense capabilities to generate insights from sensing data (Englhardt et al. 2024;Yang et al. 2024;Ferrara 2024;Ji, Zheng, and Wu 2024;Fang et al. 2024). However, most research has concentrated on generating simple classification (Ji, Zheng, and Wu 2024;Englhardt et al. 2024), rather than providing a holistic and qualitative overview of an older adult's day that truly reflects the needs of caregivers and other stakeholders. ...

Large Language Models for Wearable Sensor-Based Human Activity Recognition, Health Monitoring, and Behavioral Modeling: A Survey of Early Trends, Datasets, and Challenges

Sensors

... • LLMs for Fact-Checking: Large language models have shown promise in generating explanations and summaries relevant to fact-checking, but struggle with reliability. Researchers are exploring fine-tuning techniques and ways to constrain LLM output to improve factual accuracy [6,8,19]. ...

FACT-GPT: Fact-Checking Augmentation via Claim Matching with LLMs
  • Citing Article
  • January 2024

SSRN Electronic Journal

... Moreover, the widespread nature of online toxicity makes manual counterspeech highly impractical. To overcome these limitations, automated counterspeech systems reference counterspeech persuasion personalization [10,40] ✓ ✓ [8,20] ✓ ✓ [4,37,46,50 leveraging generative AI technologies-such as large language models (LLMs)-have been developed [5,66]. Yet, evaluating their effectiveness remains challenging [28]. ...

Can Language Model Moderators Improve the Health of Online Discourse?
  • Citing Conference Paper
  • January 2024

... IOs have also been linked to global political events, such as the Ukraine-Russia war (Chen and Ferrara. 2023;Pierri, et al., 2023;Soares, et al., 2023), the conflict in Gaza (Dey, et al., 2024), and protests in Spain, Turkey, Serbia, and Venezuela (Guo and Vosoughi, 2022;Stella, et al., 2018). The need to detect IOs on social media has thus become a critical research priority for preserving the integrity of online discourse and democratic processes. ...

Coordinated Activity Modulates the Behavior and Emotions of Organic Users: A Case Study on Tweets about the Gaza Conflict
  • Citing Conference Paper
  • May 2024

... We utilize open source Small Language Models (SLMs) and achieve comparable results on our evaluation dataset, presenting more quantitative findings. Another study by Eun et al. [35] employs LoRa for fine-tuning LLMs and evaluates claim accuracy across various claim styles by adjusting the temperature setting in the model. However, it majorly only focuses on combating fake news related to the coronavirus period, raising concerns about the generalizability of these claims. ...

FACT-GPT: Fact-Checking Augmentation via Claim Matching with LLMs
  • Citing Conference Paper
  • May 2024