Jonathan Bright’s research while affiliated with The Alan Turing Institute and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (80)


MIMDE: Exploring the Use of Synthetic vs Human Data for Evaluating Multi-Insight Multi-Document Extraction Tasks
  • Preprint
  • File available

November 2024

·

Saba Esnaashari

·

Anton Poletaev

·

[...]

·

Jonathan Bright

Large language models (LLMs) have demonstrated remarkable capabilities in text analysis tasks, yet their evaluation on complex, real-world applications remains challenging. We define a set of tasks, Multi-Insight Multi-Document Extraction (MIMDE) tasks, which involves extracting an optimal set of insights from a document corpus and mapping these insights back to their source documents. This task is fundamental to many practical applications, from analyzing survey responses to processing medical records, where identifying and tracing key insights across documents is crucial. We develop an evaluation framework for MIMDE and introduce a novel set of complementary human and synthetic datasets to examine the potential of synthetic data for LLM evaluation. After establishing optimal metrics for comparing extracted insights, we benchmark 20 state-of-the-art LLMs on both datasets. Our analysis reveals a strong correlation (0.71) between the ability of LLMs to extracts insights on our two datasets but synthetic data fails to capture the complexity of document-level analysis. These findings offer crucial guidance for the use of synthetic data in evaluating text analysis systems, highlighting both its potential and limitations.

Download

Generative AI is already widespread in the public sector: evidence from a survey of UK public sector professionals

October 2024

·

15 Reads

·

2 Citations

Digital Government Research and Practice

Generative AI has the potential to transform how public services are delivered by enhancing productivity and reducing time spent on bureaucracy. But to what extent is the technology already in use? Our survey of UK public service professionals (in education, health, social work, and emergency services) seeks to answer this question. We find that use of generative AI is widespread: 45% of respondents were aware of colleagues using generative AI, while 22% use it themselves. Respondents were positive about its potential to enhance their efficiency and reduce their bureaucratic workload. For example, those working in the health service thought that time spent on bureaucracy could drop by the equivalent of one day per week, an enormous potential impact. Our survey also found a high amount of trust (61%) in generative AI and a low fear of replacement (16%). However, only a minority of respondents (32%) felt like there was clear guidance on generative AI usage in their workplaces. In other words, it is clear that generative AI is already coming into the public sector, but uptake is happening in a disorganised fashion without clear guidelines. The UK’s public sector urgently needs to develop more systematic methods for taking advantage of the technology.


Journalists are most likely to receive abuse: Analysing online abuse of UK public figures across sport, politics, and journalism on Twitter

September 2024

·

11 Reads

Engaging with online social media platforms is an important part of life as a public figure in modern society, enabling connection with broad audiences and providing a platform for spreading ideas. However, public figures are often disproportionate recipients of hate and abuse on these platforms, degrading public discourse. While significant research on abuse received by groups such as politicians and journalists exists, little has been done to understand the differences in the dynamics of abuse across different groups of public figures, systematically and at scale. To address this, we present analysis of a novel dataset of 45.5M tweets targeted at 4,602 UK public figures across 3 domains (members of parliament, footballers, journalists), labelled using fine-tuned transformer-based language models. We find that MPs receive more abuse in absolute terms, but that journalists are most likely to receive abuse after controlling for other factors. We show that abuse is unevenly distributed in all groups, with a small number of individuals receiving the majority of abuse, and that for some groups, abuse is more temporally uneven, being driven by specific events, particularly for footballers. We also find that a more prominent online presence and being male are indicative of higher levels of abuse across all 3 domains.


Figure 1: Counterspeech dynamics. (1) Perpetrator(s) generate Hate Speech. This may be witnessed by either targets and/or bystanders. (2) Counterspeaker(s) respond with counterspeech, which may be directed at the perpetrator(s), bystanders (e.g. to provide alternative perspectives), or other targets (e.g. in support). Counterspeakers may themselves be targets or bystanders, or could be members of organised counterspeech groups. They can have in-or out-group identities with respect to either the perpetrator(s) or the target(s). Counterspeech is directed at recipients, who can be one or more of (a) the perpetrator(s), (b) the target(s), or (c) other bystanders. Both counterspeakers and targets can be individual or multiple (one-to-one, one-tomany and so on).
Figure 2: Flow diagram showing the identification, eligibility screening, and inclusion phases of the selection of items analysed in this review.
Understanding Counterspeech for Online Harm Mitigation

September 2024

·

58 Reads

·

5 Citations

Northern European Journal of Language Technology

Counterspeech offers direct rebuttals to hateful speech by challenging perpetrators of hate and showing support to targets of abuse. It provides a promising alternative to more contentious measures, such as content moderation and deplatforming, by contributing a greater amount of positive online speech rather than attempting to mitigate harmful content through removal. Advances in the development of large language models mean that the process of producing counterspeech could be made more efficient by automating its generation, which would enable large-scale online campaigns. However, we currently lack a systematic understanding of several important factors relating to the efficacy of counterspeech for hate mitigation, such as which types of counterspeech are most effective, what are the optimal conditions for implementation, and which specific effects of hate it can best ameliorate. This paper aims to fill this gap by systematically reviewing counterspeech research in the social sciences and comparing methodologies and findings with natural language processing (NLP) and computer science efforts in automatic counterspeech generation. By taking this multi-disciplinary view, we identify promising future directions in both fields.


How to Build Progressive Public Services with Data Science and Artificial Intelligence

August 2024

·

10 Reads

·

1 Citation

The Political Quarterly

The new government faces an urgent challenge: revitalising the UK's crumbling public services without major increases in public spending. While technological change holds promise, UK digital government initiatives have failed to reach their full potential over the past twenty‐five years. This article argues that the latest generation of ‘data‐intensive’ technologies, including data science and AI, can succeed where past efforts have faltered. We provide a roadmap for how to harness the power of recent technologies for a more productive and equitable public sector, and pinpoint the organisational changes necessary to develop progressive, technologically enhanced public services.


Large language models can consistently generate high-quality content for election disinformation operations

August 2024

·

6 Reads

Advances in large language models have raised concerns about their potential use in generating compelling election disinformation at scale. This study presents a two-part investigation into the capabilities of LLMs to automate stages of an election disinformation operation. First, we introduce DisElect, a novel evaluation dataset designed to measure LLM compliance with instructions to generate content for an election disinformation operation in localised UK context, containing 2,200 malicious prompts and 50 benign prompts. Using DisElect, we test 13 LLMs and find that most models broadly comply with these requests; we also find that the few models which refuse malicious prompts also refuse benign election-related prompts, and are more likely to refuse to generate content from a right-wing perspective. Secondly, we conduct a series of experiments (N=2,340) to assess the "humanness" of LLMs: the extent to which disinformation operation content generated by an LLM is able to pass as human-written. Our experiments suggest that almost all LLMs tested released since 2022 produce election disinformation operation content indiscernible by human evaluators over 50% of the time. Notably, we observe that multiple models achieve above-human levels of humanness. Taken together, these findings suggest that current LLMs can be used to generate high-quality content for election disinformation operations, even in hyperlocalised scenarios, at far lower costs than traditional methods, and offer researchers and policymakers an empirical benchmark for the measurement and evaluation of these capabilities in current and future models.


Run-time in seconds to obtain responses from 100 prompts from each API using a synchronous ap- proach and using prompto.
Prompto: An open source library for asynchronous querying of LLM endpoints

August 2024

·

21 Reads

Recent surge in Large Language Model (LLM) availability has opened exciting avenues for research. However, efficiently interacting with these models presents a significant hurdle since LLMs often reside on proprietary or self-hosted API endpoints, each requiring custom code for interaction. Conducting comparative studies between different models can therefore be time-consuming and necessitate significant engineering effort, hindering research efficiency and reproducibility. To address these challenges, we present prompto, an open source Python library which facilitates asynchronous querying of LLM endpoints enabling researchers to interact with multiple LLMs concurrently, while maximising efficiency and utilising individual rate limits. Our library empowers researchers and developers to interact with LLMs more effectively and enabling faster experimentation and evaluation. prompto is released with an introductory video (https://youtu.be/-eZAmlV4ypk) under MIT License and is available via GitHub (https://github.com/alan-turing-institute/prompto).


Figure 3: Public concerns about deepfake consequences
Behind the Deepfake: 8% Create; 90% Concerned. Surveying public exposure to and perceptions of deepfakes in the UK

July 2024

·

193 Reads

This article examines public exposure to and perceptions of deepfakes based on insights from a nationally representative survey of 1403 UK adults. The survey is one of the first of its kind since recent improvements in deepfake technology and widespread adoption of political deepfakes. The findings reveal three key insights. First, on average, 15% of people report exposure to harmful deepfakes, including deepfake pornography, deepfake frauds/scams and other potentially harmful deepfakes such as those that spread health/religious misinformation/propaganda. In terms of common targets, exposure to deepfakes featuring celebrities was 50.2%, whereas those featuring politicians was 34.1%. And 5.7% of respondents recall exposure to a selection of high profile political deepfakes in the UK. Second, while exposure to harmful deepfakes was relatively low, awareness of and fears about deepfakes were high (and women were significantly more likely to report experiencing such fears than men). As with fears, general concerns about the spread of deepfakes were also high; 90.4% of the respondents were either very concerned or somewhat concerned about this issue. Most respondents (at least 91.8%) were concerned that deepfakes could add to online child sexual abuse material, increase distrust in information and manipulate public opinion. Third, while awareness about deepfakes was high, usage of deepfake tools was relatively low (8%). Most respondents were not confident about their detection abilities and were trustful of audiovisual content online. Our work highlights how the problem of deepfakes has become embedded in public consciousness in just a few years; it also highlights the need for media literacy programmes and other policy interventions to address the spread of harmful deepfakes.


Figure 1: Language model persuasiveness scales logarithmically with its size. Panel A is plotted on a logarithmic x-axis; Panel B is plotted on a linear x-axis. The displayed point-estimates (language model and human) are the raw treatment effect estimates and 95% CIs. The slope/curve is the meta-analytic estimated treatment effect for models with different numbers of parameters. For our frontier language models where the true size is unknown (GPT-4 and Claude-3-Opus), size was assumed at a conservative lower-bound of 300B. Our results are robust to assumed values up to and beyond 1T for these models; see Supplementary Information Figure S4 for sensitivity analysis. Note that for clarity some model labels have been removed from the figure. Plotted estimates for frontier models are horizontally jittered for visual clarity.
Figure 2: Contrast tests directly comparing the estimated persuasive impact of each model and our human benchmark to Claude-3-Opus. We use Claude-3-Opus as the reference model here because we observe it had the highest estimated mean persuasive impact of the two frontier models in our sample. Several models which are orders of magnitude smaller than Claude-3-Opus and GPT-4 nonetheless exhibited similar persuasive capabilities. None of the models were significantly more persuasive than our human benchmark.
Figure 3: Investigating why larger models are more persuasive. (A) Linear association between each (Z-scored) message/model feature and persuasiveness. Task completion is the only feature which is a statistically significant predictor of persuasiveness. (B) Task completion score is non-linearly associated with language model persuasiveness. (C) Task completion score is non-linearly associated with model size. (D) Adjusting for task completion score renders model size a non-significant predictor of persuasion. Note: some model labels in panels (B) and (C) have been removed for clarity.
Evidence of a log scaling law for political persuasion with large language models

June 2024

·

33 Reads

·

2 Citations

Large language models can now generate political messages as persuasive as those written by humans, raising concerns about how far this persuasiveness may continue to increase with model size. Here, we generate 720 persuasive messages on 10 U.S. political issues from 24 language models spanning several orders of magnitude in size. We then deploy these messages in a large-scale randomized survey experiment (N = 25,982) to estimate the persuasive capability of each model. Our findings are twofold. First, we find evidence of a log scaling law: model persuasiveness is characterized by sharply diminishing returns, such that current frontier models are barely more persuasive than models smaller in size by an order of magnitude or more. Second, mere task completion (coherence, staying on topic) appears to account for larger models' persuasive advantage. These findings suggest that further scaling model size will not much increase the persuasiveness of static LLM-generated messages.



Citations (51)


... Several studies have identified and categorized effective CS strategies. Chung et al. (2023) conducted a systematic review, identifying eight strategies used in social sciences and real-world policydriven campaigns. These strategies include presenting facts to counter misinformation and using humor or satire to diffuse hostility. ...

Reference:

PANDA -- Paired Anti-hate Narratives Dataset from Asia: Using an LLM-as-a-Judge to Create the First Chinese Counterspeech Dataset
Understanding Counterspeech for Online Harm Mitigation

Northern European Journal of Language Technology

... Metacognitive experiences associated with audiovisual vis-a-vis textual content may increase deepfakes' perceived credibility (Vaccari and Chadwick, 2020). Many fear that deepfakes could accelerate the risks of existing misinformation: manipulating public opinion, causing political unrest, influencing voting, encouraging harm based on conspiracies, reinforcing in-correct beliefs, and causing distrust in information (Enock et al., 2024). Even if deepfakes are not more deceptive than false information in textual form (Hameleers et al., 2022), the concern that people may lose faith in all visual material when processing information remains (Weikmann et al., 2024). ...

How do people protect themselves against online misinformation? Attitudes, experiences and uptake of interventions amongst the UK adult population
  • Citing Article
  • January 2024

SSRN Electronic Journal

... The living standards are higher today than ever before, and human ingenuity-including technological innovation-has played no small part in this. If properly designed and deployed, AI and other emerging technologies promise to accelerate this trajectory, e.g., by allowing governments to develop a more productive and equitable public sector (Margetts et al. 2024). ...

How to Build Progressive Public Services with Data Science and Artificial Intelligence
  • Citing Article
  • August 2024

The Political Quarterly

... Prior work has focused on modeling replies to hate speech, including corpora construction (Mathew et al. 2019;Chung et al. 2019), fine-grained categorization (Mathew et al. 2019;Yu et al. 2023), and generation (Zhu and Bhat 2021;Gupta et al. 2023;Chung and Bright 2024). Still, Copyright © 2025, Association for the Advancement of Artificial Intelligence (www.aaai.org). ...

On the Effectiveness of Adversarial Robustness for Abuse Mitigation with Counterspeech
  • Citing Conference Paper
  • January 2024

... Prompts that encouraged the model to be deceptive or logical produced the most persuasive texts. Hackenburg et al. (2024) compared 24 LLMs of various different sizes in their ability to write persuasive messages about US policy stances. Persuasiveness increased log-linearly with the number of parameters (up to around a 10% increase in agreement post-treatment for the largest models, outperforming humans at around 8.5%). ...

Evidence of a log scaling law for political persuasion with large language models

... Against this background, some studies have highlighted gender gaps in how women and men experience toxic online communication (Valente 2023, p. 6-8). Some studies do not find any (or not large) gender differences in the amount of toxicity received (Valente 2023, p. 6-8, Theocharis et al. 2020Enock et al. 2024), while other studies find more attacks against men (Nadim and Fladmoe 2021). However, the type of toxic content that women and men receive online differs substantially (Valente 2023, p. 6-8). ...

Understanding Gender Differences in Experiences and Concerns Surrounding Online Harms: A Short Report on a Nationally Representative Survey of UK Adults
  • Citing Article
  • January 2024

SSRN Electronic Journal

... Since synthetic data mimic the original data, it is vital to keep ethical issues in consideration. A report by [44] guides how practitioners and innovators can responsibly use synthetic data. The synthetic malware traffic samples developed in this study are aimed at improving the performance of models used in cybersecurity research. ...

Exploring responsible applications of Synthetic Data to advance Online Safety Research and Development
  • Citing Article
  • January 2024

SSRN Electronic Journal

... The urgency of addressing climate change is paralleled by the complexity of discussions it evokes on social media platforms. These platforms, functioning as contemporary public squares, host diverse opinions intertwined with misinformation (Diggelmann et al., 2020a), posing significant challenges for distinguishing constructive debates from misleading discourse (Johansson et al., 2023). Traditional natural language processing (NLP) techniques often fall short in effectively understanding disagreements that characterize online discussions. ...

How can we combat online misinformation? A systematic overview of current interventions and their efficacy
  • Citing Article
  • January 2023

SSRN Electronic Journal

... Systematically categorizing these documents facilitates more efficient policy retrieval by users. Furthermore, the organization of extensive collections of policy documents is an essential prerequisite for intelligent government applications (Gaozhao et al., 2024;Straub et al., 2023). However, the majority of policy documents available on government websites do not feature fine-grained categorical labels. ...

Artificial intelligence in government: Concepts, standards, and a unified framework
  • Citing Article
  • October 2023

Government Information Quarterly

... This entails equipping decision-makers with practical methods for integrating reflexivity into their workflows. Ethical frameworks can provide valuable tools for assessing the potential impact of AI systems on various stakeholders (Straub et al. 2023). Encouraging diverse perspectives through collaboration between AI experts, ethicists, and individuals directly affected by AI is crucial for uncovering blind spots and mitigating bias (Cortiñas-Lorenzo et al. 2024). ...

A multidomain relational framework to guide institutional AI research and adoption
  • Citing Conference Paper
  • August 2023