Divya Siddarth’s research while affiliated with Microsoft and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (24)


Prosocial Media
  • Preprint

February 2025

·

2 Reads

E. Glen Weyl

·

Luke Thorburn

·

Emillie de Keulenaar

·

[...]

·

Audrey Tang

Social media empower distributed content creation by algorithmically harnessing "the social fabric" (explicit and implicit signals of association) to serve this content. While this overcomes the bottlenecks and biases of traditional gatekeepers, many believe it has unsustainably eroded the very social fabric it depends on by maximizing engagement for advertising revenue. This paper participates in open and ongoing considerations to translate social and political values and conventions, specifically social cohesion, into platform design. We propose an alternative platform model that the social fabric an explicit output as well as input. Citizens are members of communities defined by explicit affiliation or clusters of shared attitudes. Both have internal divisions, as citizens are members of intersecting communities, which are themselves internally diverse. Each is understood to value content that bridge (viz. achieve consensus across) and balance (viz. represent fairly) this internal diversity, consistent with the principles of the Hutchins Commission (1947). Content is labeled with social provenance, indicating for which community or citizen it is bridging or balancing. Subscription payments allow citizens and communities to increase the algorithmic weight on the content they value in the content serving algorithm. Advertisers may, with consent of citizen or community counterparties, target them in exchange for payment or increase in that party's algorithmic weight. Underserved and emerging communities and citizens are optimally subsidized/supported to develop into paying participants. Content creators and communities that curate content are rewarded for their contributions with algorithmic weight and/or revenue. We discuss applications to productivity (e.g. LinkedIn), political (e.g. X), and cultural (e.g. TikTok) platforms.



Figure 3: This diagram illustrates how the X Community Notes algorithm works. Note ratings are decomposed into two factors: political orientation (here, the horizontal axis) and helpfulness independent of politics (vertical axis). Notes that score low in partisanship and high in helpfulness are chosen for display (Wojcik et al. 2022).
AI and the Future of Digital Public Squares
  • Preprint
  • File available

December 2024

·

245 Reads

Two substantial technological advances have reshaped the public square in recent decades: first with the advent of the internet and second with the recent introduction of large language models (LLMs). LLMs offer opportunities for a paradigm shift towards more decentralized, participatory online spaces that can be used to facilitate deliberative dialogues at scale, but also create risks of exacerbating societal schisms. Here, we explore four applications of LLMs to improve digital public squares: collective dialogue systems, bridging systems, community moderation, and proof-of-humanity systems. Building on the input from over 70 civil society experts and technologists, we argue that LLMs both afford promising opportunities to shift the paradigm for conversations at scale and pose distinct risks for digital public squares. We lay out an agenda for future research and investments in AI that will strengthen digital public squares and safeguard against potential misuses of AI.

Download

Fig. 1 | Development of information environments over time. A general trend is observed whereby new technologies increase the speed at which information can be retrieved but decrease transparency with respect to the information source.
How large language models can reshape collective intelligence

September 2024

·

3,540 Reads

·

37 Citations

Nature Human Behaviour

Collective intelligence underpins the success of groups, organizations, markets and societies. Through distributed cognition and coordination, collectives can achieve outcomes that exceed the capabilities of individuals-even experts-resulting in improved accuracy and novel capabilities. Often, collective intelligence is supported by information technology, such as online prediction markets that elicit the 'wisdom of crowds', online forums that structure collective deliberation or digital platforms that crowdsource knowledge from the public. Large language models, however, are transforming how information is aggregated, accessed and transmitted online. Here we focus on the unique opportunities and challenges this transformation poses for collective intelligence. We bring together interdisciplinary perspectives from industry and academia to identify potential benefits, risks, policy-relevant considerations and open research questions, culminating in a call for a closer examination of how large language models affect humans' ability to collectively tackle complex problems.


Figure 1. RCT estimates of political persuasion with large language models. This figure includes all known studies which randomised participants to LLM-generated political messages and measured posttreatment attitudes. For each study, we calculate the simple difference in mean outcomes by condition (with 95% CIs) in order to maximise consistency across studies, but note that this may differ from authors' original analyses. The studies vary in the model used (GPT-3, GPT-4, Claude 3 Opus), treatment format (vignettes, articles, chatbot conversations), reference conditions (experts, laypeople, etc.), as well as in the political issues considered. For descriptive purposes we include a meta-analytic average, but caution against over-interpretation given the substantial heterogeneity.
How will advanced AI systems impact democracy?

August 2024

·

548 Reads

Advanced AI systems capable of generating humanlike text and multimodal content are now widely available. In this paper, we discuss the impacts that generative artificial intelligence may have on democratic processes. We consider the consequences of AI for citizens' ability to make informed choices about political representatives and issues (epistemic impacts). We ask how AI might be used to destabilise or support democratic mechanisms like elections (material impacts). Finally, we discuss whether AI will strengthen or weaken democratic principles (foundational impacts). It is widely acknowledged that new AI systems could pose significant challenges for democracy. However, it has also been argued that generative AI offers new opportunities to educate and learn from citizens, strengthen public discourse, help people find common ground, and to reimagine how democracies might work better.


Figure 1: Personhood credentials rely on two deficits of AI.
Figure 2: Indistinguishability and scalability could drive an increase in AI-powered deception.
Figure 3: Illustration of enrollment and usage of a personhood credential.
Figure 4: Ecosystem design trade-offs-the argument for bounded credentials. Note: Open circles represent individuals, and filled circles represent issuers. Top left: Unlimited credentials per person (red). Less effective against deception but better for privacy and civil liberties due to minimal issuer data storage. Bottom right: One credential per person, from one issuer. Effective against deception but risky for privacy and civil liberties. Top right: One credential per person per issuer, with multiple issuers. This balances privacy protection and countering deception.
Personhood credentials: Artificial intelligence and the value of privacy-preserving tools to distinguish who is real online

August 2024

·

114 Reads

Anonymity is an important principle online. However, malicious actors have long used misleading identities to conduct fraud, spread disinformation, and carry out other deceptive schemes. With the advent of increasingly capable AI, bad actors can amplify the potential scale and effectiveness of their operations, intensifying the challenge of balancing anonymity and trustworthiness online. In this paper, we analyze the value of a new tool to address this challenge: "personhood credentials" (PHCs), digital credentials that empower users to demonstrate that they are real people -- not AIs -- to online services, without disclosing any personal information. Such credentials can be issued by a range of trusted institutions -- governments or otherwise. A PHC system, according to our definition, could be local or global, and does not need to be biometrics-based. Two trends in AI contribute to the urgency of the challenge: AI's increasing indistinguishability (i.e., lifelike content and avatars, agentic activity) from people online, and AI's increasing scalability (i.e., cost-effectiveness, accessibility). Drawing on a long history of research into anonymous credentials and "proof-of-personhood" systems, personhood credentials give people a way to signal their trustworthiness on online platforms, and offer service providers new tools for reducing misuse by bad actors. In contrast, existing countermeasures to automated deception -- such as CAPTCHAs -- are inadequate against sophisticated AI, while stringent identity verification solutions are insufficiently private for many use-cases. After surveying the benefits of personhood credentials, we also examine deployment risks and design challenges. We conclude with actionable next steps for policymakers, technologists, and standards bodies to consider in consultation with the public.


Collective Constitutional AI: Aligning a Language Model with Public Input

June 2024

·

21 Reads

There is growing consensus that language model (LM) developers should not be the sole deciders of LM behavior, creating a need for methods that enable the broader public to collectively shape the behavior of LM systems that affect them. To address this need, we present Collective Constitutional AI (CCAI): a multi-stage process for sourcing and integrating public input into LMs-from identifying a target population to sourcing principles to training and evaluating a model. We demonstrate the real-world practicality of this approach by creating what is, to our knowledge, the first LM fine-tuned with collectively sourced public input and evaluating this model against a baseline model trained with established principles from a LM developer. Our quantitative evaluations demonstrate several benefits of our approach: the CCAI-trained model shows lower bias across nine social dimensions compared to the baseline model, while maintaining equivalent performance on language, math, and helpful-harmless evaluations. Qualitative comparisons of the models suggest that the models differ on the basis of their respective constitutions, e.g., when prompted with contentious topics, the CCAI-trained model tends to generate responses that reframe the matter positively instead of a refusal. These results demonstrate a promising, tractable pathway toward publicly informed development of language models.




Figure 1: Example Frontier AI Lifecycle.
Figure 4: Computation used to train notable AI systems. Note logarithmic y-axis. Source: [50] based on data from [280].
Frontier AI Regulation: Managing Emerging Risks to Public Safety

July 2023

·

439 Reads

·

8 Citations

Advanced AI models hold the promise of tremendous benefits for humanity, but society needs to proactively manage the accompanying risks. In this paper, we focus on what we term "frontier AI" models: highly capable foundation models that could possess dangerous capabilities sufficient to pose severe risks to public safety. Frontier AI models pose a distinct regulatory challenge: dangerous capabilities can arise unexpectedly; it is difficult to robustly prevent a deployed model from being misused; and, it is difficult to stop a model's capabilities from proliferating broadly. To address these challenges, at least three building blocks for the regulation of frontier models are needed: (1) standard-setting processes to identify appropriate requirements for frontier AI developers, (2) registration and reporting requirements to provide regulators with visibility into frontier AI development processes, and (3) mechanisms to ensure compliance with safety standards for the development and deployment of frontier AI models. Industry self-regulation is an important first step. However, wider societal discussions and government intervention will be needed to create standards and to ensure compliance with them. We consider several options to this end, including granting enforcement powers to supervisory authorities and licensure regimes for frontier AI models. Finally, we propose an initial set of safety standards. These include conducting pre-deployment risk assessments; external scrutiny of model behavior; using risk assessments to inform deployment decisions; and monitoring and responding to new information about model capabilities and uses post-deployment. We hope this discussion contributes to the broader conversation on how to balance public safety risks and innovation benefits from advances at the frontier of AI development.


Citations (16)


... Enhancing access to care Large language models have opened a new window into a vast landscape of new possibilities regarding the quality of care that patients can access and how they access it. LLMs can simplify the description of medical conditions, assist in drafting medical documents, create training programs and processes, and streamline research processes, and may potentially transform healthcare by enhancing diagnostics, medical writing, education, and project management 7,8 . ...

Reference:

Patient agency and large language models in worldwide encoding of equity
How large language models can reshape collective intelligence

Nature Human Behaviour

... Of particular importance for the subsequent discussion are the concepts of constitutional AI and AI guardrails. The following paragraphs will explain those in more detail: The main idea behind constitutional AI (Abiri, 2024;Bai et al., 2022;Huang et al., 2024) is to fine-tune instruction-following LLMs so that they align with ethical principles outlined in a constitutionlike framework. This approach builds on reinforcement learning from human feedback and is closely tied to research on value alignment. ...

Collective Constitutional AI: Aligning a Language Model with Public Input
  • Citing Conference Paper
  • June 2024

... Further tools are still needed to evaluate stakeholders' compatibility by a multicriteria approach (Akrivou et al., 2022) and raise awareness of adopting circular economy principles at an industry-wide level (Adams et al., 2017). When it comes to characterizing supply and demand, there are considerable gaps in developing standardized circularity indicators (Dräger et al., 2022), understanding the design, structure, and function of the databases (Łȩ kawska-Andrinopoulou et al., 2021), and allowing scalability and interoperability of digital infrastructures (Bühler et al., 2023). The Circularity Information Platform stands out for its waste-toresource artifact, which simulates matching supply with demand for concrete in the Netherlands (Yu, 2023). ...

Harnessing Digital Federation Platforms and Data Cooperatives to Empower SMEs and Local Communities

SSRN Electronic Journal

... For others, this tension is best described as an interplay of expertise and institutional decisionmaking requiring political legitimacy (Bader, 2014). Overall, the concept of "democratisation of technology" is often used in conflicting ways, from widespread usage, collaborative development, sharing profits or democratic governance (Seger et al. 2023). ...

Democratising AI: Multiple Meanings, Goals, and Methods
  • Citing Conference Paper
  • August 2023

... UNESCO has approved its Recommendation on Open Science, whose implementation makes it essential to incorporate the "open culture" into any international, national, regional, or institutional policy related to information or research. In this regard, the Open University of Catalonia ( Central to the debate is the principle of democratizing access to knowledge, a concept highlighted by the works of Kop, M., et al. (2022), Bühler, M. M., et al. (2023), along with Dutta, M., et al. (2021). These researchers unite in their call for a scientific realm that is equitable, cooperative, and committed to open access principles. ...

Data Cooperatives as Catalysts for Collaboration, Data Sharing, and the (Trans)Formation of the Digital Commons

SSRN Electronic Journal

... That said, in order to reap all the benefits of foundation models for MS-specific applications, several open problems need to be solved. These include sub-par reasoning capabilities (Rae et al., 2021;McKenzie et al., 2023;Arkoudas, 2023) which could be dangerous in high-stakes environments such as healthcare (Richens et al., 2020;Fraser et al., 2018), broader concerns regarding AI safety (Bommasani et al., 2021;Anderljung et al., 2023;Urbina et al., 2022), and predictions that may be unacceptably skewed to the detriment of a particular group of people (Mehrabi et al., 2021). As more solutions to these problems are found, we can expect an increasing focus on large foundation models in the coming years, to help solve some of the most challenging tasks in MS MRI-analysis. ...

Frontier AI Regulation: Managing Emerging Risks to Public Safety

... On the one hand, there are examples of data campaigning like the "Data Save Lives" initiative that emphasizes the potential for data to do good in health care and ensure individuals do not opt out of health data sharing initiatives (26). On the other end of the spectrum is the establishment of data trusts and data cooperatives (27)(28)(29). These projects propose using trusted intermediaries and collective action to create financial and social benefit for the individuals who produce data (27). ...

Unlocking the Power of Digital Commons: Data Cooperatives as a Pathway for Data Sovereign, Innovative and Equitable Digital Communities

Digital

... In mitigating risks through design, collaboration with subject matter experts is crucial when considering knowledge-intensive domains and risks with nuanced definitions [75]. Shevlane et al. [69] discuss how risk evaluation needs to be embedded in governance processes. They advocate embedding risk measurement and governance processes into model training and deployment processes. ...

Model evaluation for extreme risks

... Hence, this chapter intertwingles the state-of-the-art on interrelated concepts, such as smart cities [42], smart villages [22], Living Labs [43], and action research, as well as their impact on the implementation of the SDGs by focusing on (i) the feasibility of technology, (ii) the role of politics and power relations in communities, and (iii) the self-capacity of communities to develop their locally-driven entrepreneurial model based on (data) co-operativism [44,45]. ...

Unlocking the Power of Digital Commons: Data Cooperatives as a Pathway for Data Sovereign, Innovative and Equitable Digital Communities

... In response to increasing criticism, technology companies have explicitly relied on democratic rhetoric as a key source of legitimacy. Frontier AI companies have publicly declared a desire to cede some governance power and orient their technologies toward common interests (Seger et al., 2023). 3 Experiments with stakeholder democracy in the technology sector have taken both aggregative and deliberative forms. ...

Democratising AI: Multiple Meanings, Goals, and Methods