Content uploaded by Manuel Goyanes
Author content
All content in this area was uploaded by Manuel Goyanes on Dec 06, 2023
Content may be subject to copyright.
A Scholarly Denition of Articial Intelligence (AI): Advancing AI
as a Conceptual Framework in Communication Research
Homero Gil de Zúñiga
a,b,c
, Manuel Goyanes
d
, and Timilehin Durotoye
b
a
Democracy Research Unit, Political Science, College of Law & Public Administration, University of Salamanca,
Salamanca;
b
Media Effects Research Lab, Film Production & Media Studies Department, Donald P. Bellisario
College of Communications, Pennsylvania State University, State College;
c
Facultad de Comunicación y Letras,
Universidad Diego Portales, Santiago;
d
Department of Communication, Universidad Carlos III
ABSTRACT
Research on Articial Intelligence (AI) in communication research is
gaining broader interest. This interdisciplinary interest has yet to be
supported by a systematic scholarly denition and by a holistic theo-
retical framework in communication research. First, combining prior
theoretical eorts from diverse disciplines in the social sciences, espe-
cially journalism and communication, this study introduces a wide-
ranging working AI scholarly denition in communication research as
the tangible real-world capability of non-human machines or articial
entities to perform, task solve, communicate, interact, and act logically as
it occurs with biological humans. We also propose its theoretical oper-
ationalization based on two dimensions: level of performance and level
of autonomy, advancing an elementary conceptual framework draw-
ing on AI’s levels of potential actions or performance the AI may
accomplish, including 1) performing tasks, 2) taking decisions, and 3)
making predictions; as well as AI’s level of autonomy, or the agency
results contingent on the degrees of human input, interaction, or
supervision involved.
KEYWORDS
Artificial intelligence; AI;
AI performance level;
AI autonomy level;
social sciences;
communication; journalism
Present-day society is saturated by digital infrastructures, supported by artificial technolo-
gies, which have become essential in navigating the social world (de Lima-Santos & Ceron,
2021; Hepp, 2020; Lobera et al., 2020). Rapid advances in Artificial Intelligence (AI) have
significantly disrupted the world economy and other sectors like engineering, agriculture,
politics, and media, developing technological-driven systems in the process of information
and service creation, dissemination, and preservation (Birtchnell, 2018; Kamble & Shah,
2018; Kieslich et al., 2022; Tubaro et al., 2020). The penetration, use, and reliance on AI have
various implications for different disciplines, including communication.
Some studies in communication and journalism have problematized the potential
impacts of AI in shaping users’ attitudes, perceptions, and behaviors (Brewer et al., 2022;
Chen & Wen, 2021). For instance, AI is redefining present-day journalism through the
introduction of AI-infused workflow systems, the use of data in news stories, and the
reorganization of journalistic labor and recruitment (Deuze & Beckett, 2022; Sun et al.,
2022). AI also has implications in communication, particularly in persuasive communica-
tion, in which scholars focus on the interplay between technology and persuasion to
CONTACT Homero Gil de Zúñiga hgz@usal.es Democracy Research Unit, Political Science, College of Law & Public
Administration, University of Salamanca
POLITICAL COMMUNICATION
https://doi.org/10.1080/10584609.2023.2290497
© 2023 Taylor & Francis Group, LLC
investigate how artificial entities like robotics and algorithms (Dehnert & Mongeau, 2022;
Hermann, 2022; Lee & Liang, 2018; Siegel et al., 2009) may influence persuasion across
diverse contexts, including political and public opinion (Cohen, 2021).
In journalism, this technology has reshaped modern investigative reporting, news pro-
duction, and distribution while targeting users based on their news preferences (Stray, 2019)
by employing cutting-edge AI tools (de Lima-Santos & Ceron, 2021). Thus far, researchers
have identified three phases of computerization in newsrooms: automated information
generation, automated content production, and information distribution and audience
relations (Sánchez-García et al., 2023). Each phase influences news organizations’ utiliza-
tion of machine learning to predict story virality, model audience consumption patterns
(Stray, 2019), and monetize website subscriptions (Goyanes et al., 2023). Additionally,
personalization recommenders, bots, and algorithms have been used by major newsrooms
such as The New York Times, Reuters, China Morning Post, and news aggregators like Google
News and Yahoo News (Das et al., 2007; Sánchez-García et al., 2023).
Considering the recent academic interest in AI within communication and journalism
studies and the implications of its development, it is imperative to reconcile the different
research silos that have been evolving with the consolidation of AI in communication-
related phenomena. However, little theoretical research has provided and operationalized
a definition of AI and a conceptual framework for its empirical application in communica-
tion research. We believe this task is timely and relevant as it helps clarify empirical
discourses on AI’s nature, use, and effect and their social, political, cultural, and ethical
implications in society. This study introduces a wide-ranging working AI scholarly defini-
tion in communication research as the tangible real-world capability of non-human
machines or artificial entities to perform, task solve, communicate, interact, and act logically
as it occurs with biological humans. We also propose its theoretical operationalization based
on two dimensions: level of performance and level of autonomy, advancing an elementary
conceptual framework drawing on AI’s levels of potential actions or performance the AI
may accomplish, including 1) performing tasks, 2) taking decisions, and 3) making predic-
tions; as well as AI’s level of autonomy, or the agency results contingent on the degrees of
human input, interaction, or supervision involved.
Articial Intelligence: Impact in Journalism and Communication
The concept of AI dates back to 1955 when Prof. John McCarthy defined the scientific
procedures of machine learning (de Lima-Santos & Ceron, 2021). Over the years, this
technology has revolutionized how intelligent entities are computed to simplify and per-
form human tasks, make decisions, and solve problems (Broussard et al., 2019). For
decades, academic focus on AI mainly flourished in computer science with greater con-
centration on subfields such as “(i) machine learning (Ng & Leung, 2020); (ii) computer
vision- CV (Russell & Norvig, 2022); (iii) speech recognition (Reddy, 1976); (iv) natural
language processing – NLP (Allen, 2003); (v) planning, scheduling, and optimization
(Wölker & Powell, 2021); (vi) expert systems (Benfer et al., 1991); and (vii) robotics” (de
Lima-Santos & Ceron, 2021, p. 3).
The scope of AI, however, has extended globally. Journalists and communication experts,
for instance, have used computational-based aid to improve their reporting (Broussard,
2015), develop more robust investigative journalism (Stray, 2019), and foster old and new
2H. GIL DE ZÚÑIGA ET AL.
journalistic practices, contributing to better and more informed public opinion (Moran &
Shaikh, 2022). AI automation has proven instrumental to journalism and investigative
reporting (de-Lima-Santos & Ceron, 2021; Borchardt, 2022). Conversely, the era of pseudo-
information (Kim & Gil de Zúñiga, 2021), post-factual ideas (Moon et al., 2022) have
thrived in part due to the contribution of AI tools that have sparked the emergence of fake
news generators, video games, and metaverse (Godulla et al., 2021; Mattke, 2018).
In communication, AI is changing social interactions, developing technologies adapted
to online and offline communication (Guzman & Lewis, 2020). AI aims to intelligently
replicate human communication abilities aided by computational resources (Frankish,
2014). Due to the changing dynamics in the connection between AI technologies and
communication, new inquiries are being raised on how AI fits into the communication
paradigms (Gunkel, 2012).
Generally, the two principal dimensions in communication technology research are
human-computer interaction (HCI) and computer-mediated communication (CMC),
which explain how AI might function as a communicator or a channel for human social
interactions (Sundar & Nass, 2000). These dimensions are based on the usability and the
ability to interpret social cues by human interlocutors. Recently, these subfields have been
classified into human-AI interaction (HAII), a derivative from the Theory of Interactive
Media Effects (TIME) model proposed by Sundar (Sundar & Nass, 2000), and AI-mediated
communication [AIMC] (Hancock et al., 2020). Thus, the conceptual advances have
facilitated the contextualization of AI as a rise in machine agency, conversational agents,
and embodied robots (Beattie et al., 2020; Westerman et al., 2020). These versatilities
explain how AI assists and streamlines communication (Reeves, 2016). Nevertheless, the
field still needs a simple communication research framework that facilitates the study of AI
for such interactions.
With the growing adoption of AI in communication, its functionality has extended from
one – to-one encounter to one – to–many, thereby uncovering new assumptions about the
human communication process, including deeper socio-emotional interactions (Lee &
Sundar, 2009), human-AI socialization and “friendship” (Brandtzaeg et al., 2022), and
autonomy – as visible in algorithmic assistants, work bots, AI co-authorship, and AI
curatorship (Banas et al., 2022; Duan et al., 2022; Hepp, 2020). These facets also highlight
AI’s role as an intelligent-agent communicator in contrast to a simple mediator role and the
degree of agency it affords humans versus machines in journalism, persuasion, and com-
munication (Dehnert & Mongeau, 2022; Sundar & Lee, 2022).
AI is also utilized to resolve issues of hate speech and mis/disinformation online and on
social media (Banas et al., 2022); conduct sentiment analysis, information pattern extraction,
visual analysis, and speech recognition (Ryan, 2020; Vergeer, 2020). Additionally, commu-
nicative bots as artificial companions are becoming more popular among users (Hepp, 2020)
and E-commerce enterprises for sales conversions and learning about customers’ purchasing
behaviors via recommender systems (Sun & Zhang, 2018). However, there are projected
benefits to the future scenarios of AI technology and drawbacks to its malicious uses at
individual and organizational levels (Nah et al., 2020; Araujo et al., 2020; Canavilhas, 2022;
Ciancaglini et al., 2020; Helberger & Diakopoulos, 2022; Natale, 2021).
Despite the advantages of AI, there are ethical, privacy, and authenticity concerns about
its deployment (Hermann, 2022; Kieslich et al., 2022). For instance, there are issues of lack
of transparency and inaccuracy of AI-embedded systems in learning about users’
POLITICAL COMMUNICATION 3
preferences (Jokinen, 2015), or the “cold start problem,” where inferences about users may
not be accurately drawn (Gope & Jain, 2017; Lee, 2020). Some examples are machine
learning algorithms used in recommender systems in social media platforms like Tik Tok,
Instagram, YouTube, or media streaming services like Netflix, Hulu, Apple TV, etc. There
are also uncertainties about increased machine autonomy as AI gains consciousness and
becomes self-aware to accelerate attention processing and emulate human-level intelligence
like ChatGPT (i.e., GPT-4) and DALL-E (Chella & Manzotti, 2013; Ng & Leung, 2020).
Several of these adverse effects of AI inspired the launch of the European AI Act in 2021 by
the European Commission to provide binding ethical guidelines for private and public
integration of AI (Helberger & Diakopoulos, 2022).
Relatedly, Volovici et al. (2022) mention how the abuse of AI in some healthcare systems
can propagate prejudices like racism and sexism. Generative AI, a deep learning machine
model that creates high-dimensional data from simple prompts, like texts, images, or work-
able prototypes in response to users’ commands (i.e., ChatGPT, and Dall-E), has been
identified as a potential threat to journalists (Jamil, 2021; Ciancaglini et al., 2020) and
marketers due to the ability of AI to perform tasks accurately and quickly (Pavlik, 2023).
Although AI tools are expected to reduce the costs of investigative journalism, current meta-
journalistic conversations begin to suggest otherwise due to the uptake of algorithmic
processes that undermine human involvement in news production (Borchardt et al., 2019;
Guzman, 2018; Moran & Shaikh, 2022). Likewise, there are predictions of the death of
journalism values as journalists grapple with over-reliance on robots in the newsrooms,
raising ethical debates on how AI is appropriated in the news (Moran & Shaikh, 2022).
Scholarly Denition of Articial Intelligence in Communication Research
The study of AI in communication has relied on broad descriptions from computer science
and engineering, sociology, and legal studies as “learning algorithm used to approximate
some form of intelligence operating within computing machines” (Ninness & Ninness,
2020, p. 100). In some cases, AI has been subsumed as automation in media use and
production (Hepp, 2020) and categorized as general AI, which explains intelligent systems
designed to conduct limited tasks as opposed to possessing similar or higher human
intelligence (Ryan, 2020). The role of AI in communication and newly proposed conceptual
parameters – human-AI interaction (HAII) and AI-mediated communication, are dis-
cussed by scholars based on factors like source orientation and social exchanges optimized
by machine learning (Guzman & Lewis, 2020; Sundar & Lee, 2022).
Furthermore, there is a lack of operationalization of different constructs to assess the
impact of AI on democracy, news media, social and political attitudes, and perceptions such
as public trust (Chen & Wen, 2021).
We define AI as the tangible real-world capability of non-human machines or artificial
entities to perform, task solve, communicate, interact, and act logically as it occurs with
biological humans. AI may not solely depend on human intelligence to define its existence,
and we argue that its operationalization is based on two parameters: level of performance
and level of autonomy. Accordingly, this study advances an elementary conceptual frame-
work drawing on AI’s levels of performance, or the actions AI may accomplish, which
include 1) performing tasks, 2) making decisions, and 3) making predictions, as well as AI’s
level of autonomy, as the agency results contingent on the degrees of human input,
4H. GIL DE ZÚÑIGA ET AL.
interaction, or supervision involved. Like humans, AI will disclose different levels of
intelligence based on its performance and autonomy capabilities.
Table 1 shows that AI is mostly defined as computational technique capable of actualiz-
ing specific tasks, mimicking human behaviors, cognitive heuristics, and intelligence with-
out considering the relationship between human-human and human-machine
communication patterns. Although none of definitions listed in the table below include
all the different attributes proposed by the current definition offered in this article, we
advance a holistic AI theoretical framework here that can largely be applied regardless of
which academic definition is finally employed.
AI as a Conceptual Framework in Communication Research
Drawing on current communication studies and building on our definition of AI, we argue
that AI embraces a notion that may be categorized into two levels: 1) the depth of analysis
that AI can perform specific tasks, make decisions, communicate, interact, and forecast
results, resulting in unique output, and 2) the degree to which AI requires or does not
require human supervision or interaction. Based on these two features, our framework for
explicating AI in communication research stands as follows:
Level of performance
This level of analysis, which is consistently but dispersedly discussed in the literature as an
inherent feature of AI (Gunkel, 2012; Lee et al., 2023; Park et al., 2021), is implicitly
appraised by extant research but lacks the consistent operationalization we seek to present
next. It entails the potential actions of AI, including 1) performing tasks, 2) making
decisions, and 3) making predictions. These three categories should be understood as
complex actions with different levels of interaction and mutual feedback, as well as an
initial approach to understand and advance a historically primitive framework of AI in
communication (see Figure 1).
Figure 1 illustrates that AI may achieve different performance levels individually (single
level), in a dual interaction (second level), in a triple interactive mode (third level), or as
a series of events occurring one after the other (i.e., perform tasks – make decisions, or make
decisions – make predictions, or perform tasks – make predictions – make decisions).
Performing a task, for instance, can be theoretically assumed to be a simple action for AI,
but as we proceed to deciding or predicting, the level of complexity increases. The level of
complexity rises as the potential connections between these activities are also triggered: for
example, decisions based on predictions (i.e., Gmail and Outlook text predictions,
Grammarly, Google editor) or predictions based on decisions (i.e., Canva and Google
search engine). What we argue here is that the output of AI depends on the intricacies of
an AI interaction between its performative levels, 1) performing tasks, 2) making decisions,
and 3) making predictions to produce results, and the quality of inputs an AI initially
receives or its level of autonomy.
AI’s level of performance can be understood through automated decision-making, which
describes actions conducted using advanced technologies and without human involvement
to make data-driven decisions (Dodge & Kitchin, 2007; European Commission, 2018). In
other words, algorithms are utilized to gather, process, and generate data for making
POLITICAL COMMUNICATION 5
Table 1. Current definitions of AI in communication research.
Author (s) Article Journal/Publisher Definition of AI
Castro and New (2016, p. 2) The Promise of Artificial Intelligence Center for Data Innovation The process of “creating computing machines and systems
that perform operations analogous to human learning
and decision-making.”
Ng and Leung (2020, p. 1) Strong Artificial Intelligence and Consciousness Journal of Artificial Intelligence
and Consciousness
“Artificial Intelligence (AI) is about emulating the human
intelligence process by machines.”
de Lima-Santos and Ceron (2021) Artificial Intelligence in News Media: Current Perceptions
and Future Outlook
Journalism and Media It is a step-by-step process for performing repetitive actions,
designing models, and solving technical problems
without pre-existing concrete solutions.
Nah et al. (2020) Communicating Artificial Intelligence (AI): Theory,
Research, and Practice
Communication Studies It is portrayed as the interactions between virtual and
human agents across social, cultural, political, and ethical
domains to enable machine imitations of human
behavior and thoughts.
European Commission,
Directorate-General for
Communications Networks,
Content, and Technology
(2021)
Proposal For a Regulation Of The European Parliament and
of The Council Laying Down Harmonized Rules on
Artificial Intelligence (Artificial Intelligence Act) and
Amending Certain Union Legislative Acts
European Union AI Act AI systems are defined in an extensive framework, with risk-
based and social-good approaches, which comprises
automation in the form of structured or unstructured
data acquisition, content creation, moderation, and
production, predictive and recommendation systems
employed for decision-making, and targeting users.
Guzman and Lewis (2020) Artificial intelligence and communication: A Human-
Machine Communication research agenda
New Media & Society The concept of AI is focused on understanding human
intelligence and developing technologies to perform
tasks associated with human cognitive and emotional
abilities.
Hancock et al. (2020, p. 90) AI-Mediated Communication: Definition, Research
Agenda, and Ethical Considerations
Journal of Computer-Mediated
Communication
AI “refers broadly to computational systems that involve
algorithms, machine learning methods, natural language
processing, and other techniques that operate on behalf
of an individual to improve communication outcome.”
Russell (2010, p. 4) Artificial Intelligence: A Modern Approach Pearson Education “A computational rational agent that acts given inputs
(percepts) to achieve the best-expected outcome.”
Kurzweil (1990, p. 117) The Age of Intelligent Machines MIT press Cambridge “The art of creating machines that
perform tasks that require intelligence when people carry
them out.”
Jamil (2021) Artificial Intelligence and Journalistic Practice: The
Crossroads on Obstacles and Opportunities for Pakistani
Journalists
Journalism Practice In journalism, AI extends beyond machine storytelling. It is
described as the processing of natural language and the
incorporation of transformed data, algorithms, and
automated modes of news production and dissemination
aided by news bots.
(Continued)
6H. GIL DE ZÚÑIGA ET AL.
Table 1. (Continued).
Author (s) Article Journal/Publisher Definition of AI
Kaplan and Haenlein (2019, p. 17) Siri, Siri, in my hand: Who’s the fairest in the land? On
the interpretations, illustrations, and implications of
artificial intelligence
Business Horizons “a system’s ability to interpret external data correctly, to
learn from such data, and to use those learnings to
achieve specific goals and tasks through flexible
adaptation.”
Sundar (2020) Rise of Machine Agency: A Framework For Studying the
Psychology of Human–AI Interaction (HAII)
Journal of Computer-Mediated
Communication
It entails AI-driven mass personalization of communication
content.
Ryan (2020, p. 3) In AI We Trust: Ethics, Artificial Intelligence, and Reliability Science and Engineering Ethics “It is a field of computer science that focuses on computer
processes that can often function and react in human-like
ways, such as image recognition (vision), speech
recognition (hearing), and natural language generation
(speaking).”
Carlson (2015) The Robotic Reporter: Automated Journalism and the
Redefinition of Labor, Compositional Forms, and
Journalistic Authority
Digital Journalism It is defined as a set of algorithmic processes designed to
generate and distribute media output like texts and
images for general use, with limited or no human
intervention.
Russell and Norvig (2022) Artificial intelligence: a modern approach (4th ed.). London: Pearson Education “Study of agents that receive precepts of the environment
and perform actions.”
The United Nations Information
Economy Report (UNCTAD,
2017, p. 5)
Information Economy Report: Digitalization, Trade, and
Development
United Nations Publication “AI refers to the capability of machines to imitate intelligent
human behavior. This may involve performing various
cognitive tasks, such as sensing, processing oral
language, reasoning, learning, making decisions, and
demonstrating an ability to manipulate objects
accordingly”
POLITICAL COMMUNICATION 7
decisions in some aspects, including political, cultural, economic, and ethical (Elish & Boyd,
2018; Kitchin, 2017). Gunkel (2012) and Natale (2021) emphasized that AI should be
viewed from a technical and material functioning of computing technologies or machine
intelligence, which is pivotal in modeling the relationship between humans and machines.
In digital journalism, social bots are employed to create a scenario where AI tech-
nologies are active producers of communication, including news content (Natale, 2021).
Tong and Walther (2011) mention the use of algorithms and search engines for data
mining news leads and crafting prose to refine news stories. This indicates a practice of
computational journalism where algorithmic tools are employed for tracking topics on
public affairs (Appelgren & Nygren, 2014), analyzing big data sets (Andersen, 2018),
fact-checking falsified information, and restructuring journalistic processes (Lewis et al.,
2020). Some examples include Open Refine (openrefine.org) and Dedupe.io (DataMade,
2016).
Furthermore, AI may perform tasks and make predictions in aggregated forms.
Vergeer (2020) reveals that the gatekeeping function of journalists is disrupted with
the use of machine learning algorithms to build paywalls and recommender systems
that help journalists learn about the audience’s interests (Shaw et al., 2021;
Gruszczynski & Wagner, 2017). It makes decisions using proprietary content man-
agement systems and statistical programs. Automated news is becoming a trend in
digital journalism (Broussard, 2015), with several machine-written news pieces being
created across multiple languages within minutes (Dörr, 2016; Hansen et al., 2017).
Nevertheless, the importance of AI is related to automated news articles, as well as
Figure 1. Venn Diagram Showing Examples of Artificial Intelligence Tools based on their levels of
performance.
8H. GIL DE ZÚÑIGA ET AL.
newsroom planning, post-production, and optimization via chatbots and other AIs
such as Quakebot (Salaverría & de Lima-Santos, 2020), ChatGPT, and Dall-E
(Moran & Shaik, 2022).
Drawing on our AI framework, we argue machine-human interaction and online
communication is also changing. Institutions are increasingly adopting machine agents
or chatbots to carry out a variety of communicative tasks previously designated to
humans, such as customer support (Adam et al., 2021), scheduling appointments
(Srinivas & Ravindran, 2018), and personal banking (Umamaheswari et al., 2023). AI
has also permeated everyday human conversations that can manage people’s mental
health (Adam et al., 2021), convey nonverbal expressions with emojis (Prada et al.,
2018), introduce identity shifts (Gonzales & Hancock, 2008), and foster relationships
(Tong & Walther, 2011). AI models like ChatGPT and DALL-E (Zhang et.al., 2023),
developed by research lab OpenAI, have made headlines for their ability to function as
meaningful AI. Only levels of performance and the level of autonomy will dictate how
this AI will affect the world in the future.
Following the analyses of AI’s functioning, a theoretical evaluation of these AI
activities denotes that they should be understood as multilayered actions that require
the interaction of single, dual, or triple activities at different levels. Moreover, the
level of intricacy increases as mutual feedback and potential connections between
these AI actions are systematically triggered. Digital streaming service Netflix, for
instance, provides customized recommendations based on users’ viewing history and
likes (Sun & Zhang, 2018).
Major digital companies like Google, Alibaba, Amazon, and Apple have intro-
duced conversational AI to revolutionize information systems. These bots are inte-
grated as AI assistants such as Alexa (Amazon), Siri (Apple), Google Assistant, and
Bard (Google’s chat GPT-like AI), which can process instructions, engage in chit-
chat, and allow users to access information (Sun & Zhang, 2018). Figure 1 shows the
majority of present-day AI activities, portraying the three integral actions and their
relationships.
Deuze and Beckett (2022) reference how AI-infused systems create news stories and
gather public issues that resonate most among audiences through various websites or
applications. There are cases of Twitter bot indicators - Botometer and Tweetbotornot -
performing tasks like detecting automated accounts that collate tweets used to pro-
mote political discourses (Martini et al., 2021). With the rapid advancement in AI
technology, sophisticated predictions are made using, for instance, voice recognition
applications to detect health-related issues (Schulz, 2017), or algorithms to discover
hidden trends from big data that might provide insights into relevant news topics
(Stray, 2019).
Hansen et al. (2017) and Marconi and Siegman (2017) mentioned the impact of machine
intelligence in monitoring global feeds, drawing patterns from big data sets and collating the
results in writing a news story that can be shared with the audience. Today, the prevalence
of AI technologies has made such theorizing a reality with several examples of machine
learning used in interpersonal and computer-mediated contexts (Beattie et al., 2020;
Westerman et al., 2020) and by news organizations (Marconi & Siegman, 2017; Schmitt,
2019; Wölker & Powell, 2021). However, these AI’s performance levels are not exclusive to
journalism or communication disciplines.
POLITICAL COMMUNICATION 9
Level of autonomy
This level of analysis provides a nuance on perceptions of the potential of AI-Human
interaction and has been a debate in the literature (Hancock et al., 2020; Milano et al., 2020).
Theoretically assumed to be of high or low levels of autonomy, AI’s autonomy discerns the
agency result of the potential degree of AI or human input, interaction, or supervision
involved. Fortunati (2018) references how AI communicative bots are designed as auton-
omously operating systems to facilitate quasi-communication between humans, and AI
possesses different interfaces of communicative autonomy. The function of communicative
bots depends on datafication and continuous feedback, which determines if a system is
partially or fully automated, as well as the variation of oversight it affords individuals
(Araujo et al., 2020; Esposito, 2017). So, the larger the amount of data processing capabilities
of any given AI, the higher the automation and extent of mediatization (Andersen, 2018)
and, ultimately, autonomy.
Some AI tools are developed to be “functionally automatic, to act when triggered without
any regular human intervention or oversight” (Gillespie et al., 2014, p. 170) This indicates
high visualization of information from databases or the internet. AI companions such as
Apple’s Siri, Amazon’s Alexa, Microsoft’s Cortana, Google’s Assistant, and Shazam are
autonomous systems that meet human socio-communication needs (Diakopoulos, 2016;
Hepp, 2020; Selwyn & Gallo Cordoba, 2022). Nevertheless, these AI tools provide various
degrees of control for users. Shazam, for instance, gives users extended control over their
music consumption because it is designed to identify songs via a smartphone’s microphone
and propose music using search results saved in users’ databases (Hepp, 2020). This
conceptualization of AI-enabled interaction, thus, deals with humans’ control over AI
and not the opposite. Results of a fully supervised AI technology may imply different
societal, economic, or human risks, assuming that if humans fully control AI technologies,
this may influence fewer risking outputs, as opposed to a fully autonomous AI technology
close to the cutting-edge technological singularity.
Discussion
Many news agencies employ AI tools to enhance their users’ experiences (Lee et al., 2023)
and improve automated news production and distribution (Deuze & Beckett, 2022; Moran
& Shaikh, 2022), redefining our society (Barredo-Ibáñez et al., 2021; Dehnert & Mongeau,
2022; Siegel et al., 2009). AI represents a fundamental role in many strands of society, from
industries to scholarship, including communication as a field. Thus, it is crucial to define its
core processes and provide a framework aiming at capturing the broadness of its application
in communication research. Furthermore, as AI grows rapidly, predominant theories on
communication research only provide a partial perspective of AI’s framework. Hence, our
proposed definition of AI broadly conceives the complexity of AI based on its level of
performance and autonomy and the procedures involved. This study provides a holistic
scholarly definition of AI as the tangible real-world capability of non-human machines to
perform, task solve, communicate, and interact as it occurs with humans.
AI’s levels of performance and autonomy situate theoretical perspectives of
human-machine interaction and the variants of the practices conducted by AI.
These classifications will help scholars and users reimagine the different uses of AI
10 H. GIL DE ZÚÑIGA ET AL.
systems and their outcomes for various issues. Scholars, for instance, could investi-
gate the potential and effects of computational journalism to deepen conversations
on AI’s formative role in the newsroom and examine audiences’ reaction to AI-
generated news vs. journalist-written news. Future studies may explore the psycho-
logical effects of AI literacy on individuals’ expectations of how an algorithm’s
performance might help them distinguish between fake and real news, thereby
increasing fake news literacy among audiences who consume news from social
media or alternative media, which are known as hotspots for the spread of mis-
information (Lee et al., 2023; Mason et al., 2018).
Moreso, there is an uncanniness in how AI can accentuate social and digital
disparities. With the growing popularity of generative AI models like Chat GPT,
GPT-3, and GPT-4, there is an opportunity to leverage AI’s performance and
autonomy in designing empirical studies on the personalization and human-like
conversational metrics of AI agents to explore the emotional intelligence in human-
AI interactions and its impact on users’ perceptions. Following the distinctions of
AI, it is clear that media and communication studies will pose different research
questions and attempt to address debates on the communicative relationship
between humans and AI, including ethical contentions.
Numerous AI challenges need to be faced by academics and policymakers. Diverse
AI-like communicative robots, for instance, are based on the premise of the accumula-
tion of individual private data, so there is a need for a clearer understanding of what
data is public or private when AI is embedded in users’ social spaces (Lutz & Tamó-
Larrieux, 2020) as developers make specific trade-off decisions that may threaten the
ethical design of AI tools. Additionally, it is critical to bring about the notion of AI
accountability, that is, who bears the responsibility for AI intrusion in users’ private
spaces (Gunkel, 2018). AI systems may run the risk of undermining transparency for
users concerning the machine learning process (Liao et al., 2020) and disclosure of the
methods of data collection and the extent of individuating information collected (Sundar
et al., 2013); tracking their data, and monitoring their digital footprints such as interests
and media use patterns.
Over time, different institutions like the European Union continually address such issues
through policymaking (Canavilhas, 2022) and the U.S. Congress is also considering AI’s
implications to society and the health of US democracy. The most recent case is the Tik Tok
Congressional Hearing on privacy invasion, data security, political propaganda, and child
safety (Shepardson & Ayyub, 2023; Tolentino, 2023). Despite the risks and vulnerabilities
caused by the advancement of AI, this Forum article aims to present a standard working
definition of AI in communication and provide a conceptual framework for understanding
how AI models are structured, extending research on the examination of their future
influence in society.
Disclosure statement
No potential conflict of interest was reported by the author(s).
POLITICAL COMMUNICATION 11
Funding
The work was supported by the Spanish National Research Agency’s Program for the Generation of
Knowledge and the Scientific and Technological Strengthening Research + Development [PID2020-
115562GB-I00].
Notes on contributors
Homero Gil de Zúñiga: Ph.D. in Politics at Universidad Europea de Madrid and Ph.D. in Mass
Communication at University of Wisconsin – Madison, serves as Distinguished Research Professor at
University of Salamanca, as Distinguished Professor of Media Effects & AI at The Pennsylvania State
University, and as Senior Research Fellow at Universidad Diego Portales. His work aims to shed an
empirical social scientific light over how social media, algorithms, AI, and other technologies affect
society. Relying on survey, experimental, and computational methods his work seeks to clarify the
way we understand some of today’s most pressing challenges for democracies.
Manuel Goyanes: Ph.D. in Journalism at University of Santiago de Compostela, serves as an Assistant
Professor at Carlos III University in Madrid and as a Visiting Fellow at the Democracy Research Unit
– DRU (University of Salamanca). His research addresses the influence of journalism and new
technologies over citizens’ daily lives, as well as the effects of news consumption on citizens’ political
knowledge and participation. He is also interested in global inequalities in academic participation, the
systematic biases towards global South scholars, and publication trends in Communication.
Timilehim Durotoye: Ph.D. student in the Donald P. Bellisario College of Communications at The
Pennsylvania State University. She earned her BA and MA degrees from the Department of
Communication and Language Arts, University of Ibadan, Nigeria. Using quantitative methods,
her research centers on the effects of new media, algorithms, and artificial intelligence (AI) on
human-social interaction, and long-debated/emergent political behavior capable of influencing
democratic processes.
ORCID
Homero Gil de Zúñiga http://orcid.org/0000-0002-4187-3604
References
Adam, M., Wessel, M., & Benlian, A. (2021). AI-based chatbots in customer service and their effects
on user compliance. Electronic Markets, 31, 427–445. https://doi.org/10.1007/s12525-020-00414-7
Allen, J. F. (2003). Natural language processing. In A. Ralston, E. D. Reilly, & D. Hemmendinger
(Eds.), Encyclopedia of computer science (4th ed., pp. 1218–1222). Wiley.
Andersen, J. (2018). Archiving, ordering, and searching: Search engines, algorithms, databases, and
deep mediatization. Media, Culture & Society, 40(8), 1135–1150. https://doi.org/10.1177/
0163443718754652
Appelgren, E., & Nygren, G. (2014). Data journalism in Sweden: Introducing new methods and
genres of journalism into “old” organizations. Digital Journalism, 2(3), 394–405. https://doi.org/10.
1080/21670811.2014.884344
Araujo, T., Helberger, N., Kruikemeier, S., & De Vreese, C. H. (2020). In AI we trust? Perceptions
about automated decision-making by artificial intelligence. AI & Society, 35(3), 611–623. https://
doi.org/10.1007/s00146-019-00931-w
Banas, J. A., Palomares, N. A., Richards, A. S., Keating, D. M., Joyce, N., & Rains, S. A. (2022). When
machine and bandwagon heuristics compete: Understanding users’ response to conflicting AI and
crowdsourced fact-checking. Human Communication Research, 48(3), 430–461. https://doi.org/10.
1093/hcr/hqac010
12 H. GIL DE ZÚÑIGA ET AL.
Barredo-Ibáñez, D., De la-Garza-Montemayor, D. J., Torres-Toukoumidis, Á., & López-López, P. C.
(2021). Artificial intelligence, communication, and democracy in Latin America: A review of the
cases of Colombia, Ecuador, and Mexico. Information Professional (EPI), 30(6). https://doi.org/10.
3145/epi.2021.nov.16
Beattie, A., Edwards, A. P., & Edwards, C. (2020). A bot and a smile: Interpersonal impressions of
chatbots and humans using emoji in computer-mediated communication. Communication Studies,
71(3), 409–427. https://doi.org/10.1080/10510974.2020.1725082
Benfer, R. A., Brent, E. E., & Furbee, L. (1991). Expert systems. Sage Publications.
Birtchnell, T. (2018). Listening without ears: Artificial intelligence in audio mastering. Big Data &
Society, 5(2), 2053951718808553. https://doi.org/10.1177/2053951718808553
Borchardt, A. (2022). Go, robots, Go! The Value and Challenges of Artificial Intelligence for Local
Journalism. Digital Journalism, 10(10), 1919–1924. https://doi.org/10.1080/21670811.2022.
2149584
Borchardt, A., Lück, J., Kieslich, S., Schultz, T., & Simon, F. (2019). Blood components requests at an
orthopedic hospital: A critical survey. Hematology, Transfusion and Cell Therapy, 42(1), 25–32.
https://doi.org/10.1016/j.htct.2019.01.001
Brandtzaeg, P. B., Skjuve, M., & Følstad, A. (2022). My AI friend: How users of a social chatbot
understand their human–AI friendship. Human Communication Research, 48(3), 404–429. https://
doi.org/10.1093/hcr/hqac008
Brewer, P. R., Bingaman, J., Paintsil, A., Wilson, D. C., & Dawson, W. (2022). Media use, inter-
personal communication, and attitudes toward artificial intelligence. Science Communication, 44
(5), 559–592. https://doi.org/10.1177/10755470221130307
Broussard, M. (2015). Artificial intelligence for investigative reporting: Using an expert system to
enhance journalists’ ability to discover original public affairs stories. Digital Journalism, 3(6),
814–831. https://doi.org/10.1080/21670811.2014.985497
Broussard, M., Diakopoulos, N., Guzman, A. L., Abebe, R., Dupagne, M., & Chuan, C. H. (2019).
Artificial intelligence and journalism. Journalism & Mass Communication Quarterly, 96(3),
673–695. https://doi.org/10.1177/1077699019859901
Canavilhas, J. (2022). Inteligencia artificial aplicada al periodismo: estudio de caso del proyecto “A
European Perspective” (UER). Revista Latina de Comunicación Social, 80(80), 1–13. https://doi.
org/10.4185/RLCS-2022-1534
Carlson, M. (2015). The Robotic Reporter. Digital Journalism, 3(3), 416–431. https://doi.org/10.1080/
21670811.2014.976412
Castro, D., & New, J. (2016). The promise of artificial intelligence. Center for Data Innovation, 115
(10), 32–35.
Chella, A., & Manzotti, R. (2013). Artificial consciousness. Andrews UK Limited.
Chen, Y. N. K., & Wen, C. H. R. (2021). Impacts of attitudes toward government and corporations on
public trust in artificial intelligence. Communication Studies, 72(1), 115–131. https://doi.org/10.
1080/10510974.2020.1807380
Ciancaglini, V., Gibson, C., Sancho, D., McCarthy, O., Eira, M., Amann, P., & Klayn, A. (2020).
Malicious uses and abuses of artificial intelligence. Trend Micro Research. United Nations
Interregional Crime and Justice Research Institute (UNICRI). Available onlie here: https://docu
ments.trendmicro.com/assets/white_papers/wp-malicious-uses-and-abuses-of-artificial-intelli
gence.pdf
Cohen, M. D. (2021). Modern political campaigns: How professionalism, technology, and speed have
revolutionized elections. Rowman & Littlefield.
Das, A. S., Datar, M., Garg, A., & Rajaram, S., (2007, May 8–12). Google news personalization:
Scalable online collaborative filtering. In Proceedings of the 16th international conference on World
Wide Web, Banff, Alberta, Canada (pp. 271–280). https://doi.org/10.1145/1242572.1242610
DataMade. (2016). Introducing Dedupe.Io. https://datamade.us/blog/introducing-dedupeio
Dehnert, M., & Mongeau, P. A. (2022). Persuasion in the age of artificial intelligence (AI): Theories
and complications of AI-Based persuasion. Human Communication Research, 48(3), 386–403.
https://doi.org/10.1093/hcr/hqac006
POLITICAL COMMUNICATION 13
de Lima-Santos, M. F., & Ceron, W. (2021). Artificial intelligence in news media: Current perceptions
and future outlook. Journalism and Media, 3(1), 13–26. https://doi.org/10.3390/
journalmedia3010002
Deuze, M., & Beckett, C. (2022). Imagination, algorithms and news: Developing AI literacy for
journalism. Digital Journalism, 10(10), 1913–1918. https://doi.org/10.1080/21670811.2022.
2119152
Diakopoulos, N. (2016). Accountability in algorithmic decision making. Communications of the
ACM, 59(2), 56–62. https://doi.org/10.1145/2844110
Dodge, M., & Kitchin, R. (2007). The automatic management of drivers and driving spaces.
Geoforum, 38(2), 264–275. https://doi.org/10.1016/j.geoforum.2006.08.004
Dörr, K. N. (2016). Mapping the field of algorithmic journalism. Digital Journalism, 4(6), 700–722.
https://doi.org/10.1080/21670811.2015.1096748
Duan, Z., Li, J., Lukito, J., Yang, K. C., Chen, F., Shah, D. V., & Yang, S. (2022). Algorithmic agents in
the hybrid media system: Social bots, selective amplification, and partisan news about COVID-19.
Human Communication Research, 48(3), 516–542. https://doi.org/10.1093/hcr/hqac012
Elish, M. C., & Boyd, D. (2018). Situating methods in the magic of big data and AI. Communication
Monographs, 85(1), 57–80. https://doi.org/10.1080/03637751.2017.1375130
Esposito, E. (2017). Artificial communication? The production of contingency by algorithms.
Zeitschrift für Soziologie, 46(4), 249–265. https://doi.org/10.1515/zfsoz-2017-1014
European Commission. (2018). Data protection working party. Directive 95/46/EC, Articles 29 and 30.
Fortunati, L. (2018). Robotization and the domestic sphere. New Media & Society, 20(8), 2673–2690.
https://doi.org/10.1177/1461444817729366
Franklin, S. (2014). History, motivations, and core themes. In K. Frankish & W. Ramsey (Eds.), The
Cambridge Handbook of Artificial Intelligence (pp. 15–33). Cambridge: Cambridge University
Press. https://doi.org/10.1017/CBO9781139046855.003
Gillespie, T., Boczkowski, P. J., & Foot, K. A. (Eds.). (2014). Media technologies: Essays on commu-
nication, materiality, and society. MIT Press.
Godulla, A., Hoffmann, C. P., & Seibert, D. (2021). Dealing with deepfakes – an interdisciplinary
examination of the state of research and implications for communication studies. SCM Studies in
Communication and Media, 10(1), 72–96. https://doi.org/10.5771/2192-4007-2021-1-72
Gonzales, A. L., & Hancock, J. T. (2008). Identity shift in computer-mediated environments. Media
Psychology, 11(2), 167–185. https://doi.org/10.1080/15213260802023433
Gope, J., & Jain, S. K. (2017, May). A survey on solving cold start problem in recommender systems.
In 2017 International Conference on Computing, Communication and Automation (ICCCA),
Greater Noida, India (pp. 133–138). IEEE. https://doi.org/10.1109/CCAA.2017.8229786
Goyanes, M., Scheffauer, R., & de Zúñiga, H. G. (2023). News distribution and sustainable journalism:
Effects of Social Media news Use and Media skepticism on citizens. Paying Behavior, Mass
Communication and Society, 26(5), 878–901. https://doi.org/10.1080/15205436.2023.2169164
Gruszczynski, M., & Wagner, M. W. (2017). Information flow in the 21st century: The dynamics of
agenda-uptake. Mass Communication and Society, 20(3), 378–402. https://doi.org/10.1080/
15205436.2016.1255757
Gunkel, D. J. (2012). Communication and artificial intelligence: Opportunities and challenges for the
21st century. Communication+, 1(1(1), 1–25.
Gunkel, D. J. (2018). Robot rights. MIT Press.
Guzman, A. L. (Ed.). (2018). What is human-machine communication, anyway. Human-Machine
Communication: Rethinking Communication, Technology, and Ourselves (pp. 1–28). New York,
NY: Peter Lang.
Guzman, A. L., & Lewis, S. C. (2020). Artificial intelligence and communication: A human–machine
communication research agenda. New Media & Society, 22(1), 70–86. https://doi.org/10.1177/
1461444819858691
Hancock, J. T., Naaman, M., & Levy, K. (2020). AI-mediated communication: Definition, research
agenda, and ethical considerations. Journal of Computer-Mediated Communication, 25(1), 89–100.
https://doi.org/10.1093/jcmc/zmz022
14 H. GIL DE ZÚÑIGA ET AL.
Hansen, M., Roca-Sales, M., Keegan, J. M., & King, G. (2017). Artificial intelligence: Practice and
implications for journalism. Columbia University Libraries. Tow Center for Digital Journalism, 1–
21. https://doi.org/10.7916/d8x92prd
Helberger, N., & Diakopoulos, N. (2022). The European AI act and how it matters for research into
AI in media and journalism. Digital Journalism, 10(10), 1–10. https://doi.org/10.1080/21670811.
2022.2152195
Hepp, A. (2020). Artificial companions, social bots and work bots: Communicative robots as research
objects of media and communication studies. Media, Culture & Society, 42(7–8), 1410–1426.
https://doi.org/10.1177/0163443720916412
Hermann, E. (2022). Artificial intelligence and mass personalization of communication content—an
ethical and literacy perspective. New Media & Society, 24(5), 1258–1277. https://doi.org/10.1177/
14614448211022702
Jamil, S. (2021). Artificial intelligence and journalistic practice: The crossroads of obstacles and
opportunities for the Pakistani journalists. Journalism Practice, 15(10), 1400–1422.
Jokinen, J. P. (2015). Emotional user experience: Traits, events, and states. International Journal of
Human-Computer Studies, 76, 67–77. https://doi.org/10.1016/j.ijhcs.2014.12.006
Kamble, R., & Shah, D. (2018). Applications of artificial intelligence in human life. International
Journal of Research, 6(6), 178–188. https://doi.org/10.29121/granthaalayah.v6.i6.2018.1363
Kaplan, A., & Haenlein, M. (2019). Siri, Siri, in my hand: Who’s the fairest in the land? On the
interpretations, illustrations, and implications of artificial intelligence. Business Horizons, 62(1),
15–25. https://doi.org/10.1016/j.bushor.2018.08.004
Kieslich, K., Keller, B., & Starke, C. (2022). Artificial intelligence ethics by design. Evaluating public
perception on the importance of ethical design principles of artificial intelligence. Big Data &
Society, 9(1), 20539517221092956. https://doi.org/10.1177/20539517221092956
Kim, J. N., & Gil de Zúñiga, H. (2021). Pseudo-information, media, publics, and the failing market-
place of ideas: Theory. American Behavioral Scientist, 65(2), 163–179. https://doi.org/10.1177/
0002764220950606
Kitchin, R. (2017). Thinking critically about and researching algorithms. Information,
Communication & Society, 20(1), 14–29. https://doi.org/10.1080/1369118X.2016.1154087
Lee, E. J. (2020). Authenticity model of (mass-oriented) computer-mediated communication:
Conceptual explorations and testable propositions. Journal of Computer-Mediated
Communication, 25(1), 60–73. https://doi.org/10.1093/jcmc/zmz025
Lee, S., Gil de Zúñiga, H., & Munger, K. (2023). Antecedents and consequences of fake news
exposure: A two-panel study on how news use and different indicators of fake news exposure
affect media trust. Human Communication Research, 49(4), 408–420. https://doi.org/10.1093/hcr/
hqad019
Lee, S. A., & Liang, Y. (2018). Theorizing verbally persuasive robots. In A. L. Guzman (Ed.), Human-
machine communication: Rethinking communication, technology, and ourselves (pp. 119–143).
Peter Lang.
Lee, E.-J., & Sundar, S. S. (2009). Human-computer interaction. In C. R. Berger, M. E. Roloff, &
D. R. Roskos-Ewoldsen (Eds.), The handbook of communication science (2nd ed), pp. 507–523).
Sage Publications.
Lewis, N. P., McAdams, M., & Stalph, F. (2020). Data journalism. Journalism & Mass Communication
Educator, 75(1), 16–21. https://doi.org/10.1177/1077695820904971
Liao, Q. V., Gruen, D., & Miller, S. (2020, April 25–30). Questioning the AI: Informing design
practices for explainable AI user experiences. In Proceedings of the 2020 CHI Conference on Human
Factors in Computing Systems, Honolulu, HI, USA (pp. 1–15). CHI.
Lobera, J., Fernández Rodríguez, C. J., & Torres-Albero, C. (2020). Privacy, values and machines:
Predicting opposition to artificial intelligence. Communication Studies, 71(3), 448–465. https://doi.
org/10.1080/10510974.2020.1736114
Lutz, C., & Tamó-Larrieux, A. (2020). The robot privacy paradox: Understanding how privacy
concerns shape intentions to use social robots. Human-Machine Communication, 1, 87–111.
https://doi.org/10.30658/hmc.1.6
POLITICAL COMMUNICATION 15
Marconi, F., & Siegman, A. (2017). The future of augmented journalism: A Guide for newsrooms in the
age of smart machines. Associated Press.
Martini, F., Samula, P., Keller, T. R., & Klinger, U. (2021). Bot, or not? Comparing three methods for
detecting social bots in five political discourses. Big Data & Society, 8(2), 20539517211033566.
https://doi.org/10.1177/20539517211033566
Mason, L. E., Krutka, D., & Stoddard, J. (2018). Media literacy, democracy, and the challenge of fake
news. Journal of Media Literacy Education, 10(2), 1–10. https://doi.org/10.23860/JMLE-2018-10-
2-1
Mattke, S. (2018, June 6). KI gegen KI: Wettbewerb zu F.lschung von Video-Inhalten [AI vs. AI:
Competition on video content counterfeiting]. Heise online. Retrieved from: https://www.heise.de/
newsticker/meldung/DARPA-veranstaltet-Wettbewerb-zu-Faelschung-von-Video-Inhalten
-4074467.html
Milano, S., Taddeo, M., & Floridi, L. (2020). Recommender systems and their ethical challenges. AI &
SOCIETY, 35(4), 957–967. https://doi.org/10.1007/s00146-020-00950-y
Moon, W. K., Chung, M., & Jones-Jang, S. M. (2022). How can we fight partisan biases in the
COVID-19 pandemic? AI source labels on fact-checking messages reduce motivated reasoning.
Mass Communication and Society, 26(4), 1–25. https://doi.org/10.1080/15205436.2022.2097926
Moran, R. E., & Shaikh, J. S. (2022). Robots in the news and newsrooms: Unpacking Meta-Journalistic
Discourse on the use of artificial intelligence in journalism. Digital Journalism, 10(10), 1756–1774.
https://doi.org/10.1080/21670811.2022.2085129
Moran, R. E., & Shaikh, S. J. (2022). Robots in the news and newsrooms: Unpacking meta-journalistic
discourse on the use of artificial intelligence in journalism. Digital Journalism, 10(10), 1–19.
https://doi.org/10.1080/21670811.2022.2085129
Nah, S., McNealy, J., Hyun Kim, J., & Joo, J. (2020). Communicating artificial intelligence (AI):
Theory, Research, and practice. Communication Studies, 71(3), 369–372. https://doi.org/10.1080/
10510974.2020.1788909
Natale, S. (2021). Communicating through or communicating with: Approaching artificial intelli-
gence from a communication and media studies perspective. Communication Theory, 31(4),
905–910. https://doi.org/10.1093/ct/qtaa022
Ng, G. W., & Leung, W. C. (2020). Strong artificial intelligence and consciousness. Journal of Artificial
Intelligence and Consciousness, 7(1), 63–72. https://doi.org/10.1142/S2705078520300042
Ninness, C., & Ninness, S. K. (2020). Emergent virtual analytics: Artificial intelligence and
human-computer interactions. Behavior & Social Issues, 29(1), 100–118. https://doi.org/10.1007/
s42822-020-00031-1
Park, S., Sang, Y., Jung, J., & Stroud, N. J. (2021). News engagement: The roles of technological
affordance, emotion, and social endorsement. Digital Journalism, 9(8), 1007–1017. https://doi.org/
10.1080/21670811.2021.1981768
Pavlik, J. V. (2023). Collaborating with ChatGPT: Considering the Implications of Generative
Artificial Intelligence for Journalism and Media Education. Journalism & Mass Communication
Educator, 78(1), 84–93. https://doi.org/10.7769/58221149577
Prada, M., Rodrigues, D. L., Garrido, M. V., Lopes, D., Cavalheiro, B., & Gaspar, R. (2018). Motives,
frequency and attitudes toward emoji and emoticon use. Telematics and Informatics, 35(7),
1925–1934. https://doi.org/10.1016/j.tele.2018.06.005
Reddy, D. R. (1976). Speech recognition by machine: A review. Proceedings of the IEEE, 64(4),
501–531. https://doi.org/10.1109/PROC.1976.10158
Reeves, J. (2016). Automatic for the people: The automation of communicative labor. Communication
and Critical/Cultural Studies, 13(2), 150–165. https://doi.org/10.1080/14791420.2015.1108450
Russell, S. J. (2010). Artificial intelligence a modern approach. Pearson Education, Inc.
Russell, S., & Norvig, P. (Eds.). (2022). Artificial intelligence: A modern approach (4th ed.). Pearson
Education, Inc.
Ryan, M. (2020). In AI we trust: Ethics, artificial intelligence, and reliability. Science and Engineering
Ethics, 26, 2749–2767. https://doi.org/10.1007/s11948-020-00228-y
Salaverría, R., & de Lima-Santos, M. F. (2020). Towards Ubiquitous Journalism: Impacts of IoT on
News. In J. Vázquez-Herrero, S. Direito-Rebollal, A. Silva-Rodríguez, & X. López-García (Eds.),
16 H. GIL DE ZÚÑIGA ET AL.
Journalistic Metamorphosis. Studies in Big Data, Vol. 70. Springer: Cham. https://doi.org/10.1007/
978-3-030-36315-4_1
Sánchez-García, P., Merayo-Álvarez, N., Calvo-Barbero, C., & Díez-Gracia, A. (2023). Spanish
technological development of artificial intelligence applied to journalism: Companies and tools
for documentation, production, and distribution of information. Profesional de la información, 32
(2). https://doi.org/10.3145/epi.2023.mar.08
Schmitt, B. (2019). From atoms to bits and back: A research curation on digital technology and
agenda for future research. Journal of Consumer Research, 46(4), 825–832. https://doi.org/10.1093/
jcr/ucz038
Schulz, T. (2017). Zukunft. Zuckerbergs Zweifel Der Spiegel, 70(14), 3.
Selwyn, N., & Gallo Cordoba, B. (2022). Australian public understanding of artificial intelligence. AI
& Society, 37(4), 1645–1662. https://doi.org/10.1007/s00146-021-01268-z
Shaw, R., Cook, C., Garcia, E., Gyu Lnazaryan, H., Melano, J., Parusinski, J., & Sabadan, A. (2021).
IMS defending journalism book series. 1–24. Denmark: International Media Support (IMS). https://
www.mediasupport.org/wp-content/uploads/2021/07/AI-ML-in-Emerging-Markets-vFinal4.pdf
Shepardson, D., & Ayyub, R. (2023, March 24). TikTok congressional hearing: CEO Shou Zi Chew
grilled by US lawmakers. Reuters. Retrieved from https://www.reuters.com/technology/tiktok-ceo-
face-tough-questions-support-us-ban-grows-2023-03-23/
Siegel, M., Breazeal, C., & Norton, M. I. (2009). Persuasive robotics: The influence of robot gender on
human behavior. In Proceedings of IEEE/RSJ International Conference on Intelligent Robots and
Systems, St. Louis, MO, USA (pp. 2563–2568). IEEE. https://doi.org/10.1109/IROS.2009.5354116
Srinivas, S., & Ravindran, A. R. (2018). Optimizing outpatient appointment system using machine
learning algorithms and scheduling rules: A prescriptive analytics framework. Expert Systems with
Applications, 102, 245–261. https://doi.org/10.1016/j.eswa.2018.02.022
Stray, J. (2019). Making artificial intelligence work for investigative journalism. Digital Journalism, 7
(8), 1076–1097. https://doi.org/10.1080/21670811.2019.1630289
Sundar, S. S. (2020). Rise of machine agency: A framework for studying the psychology of human–AI
interaction (HAII. Journal of Computer-Mediated Communication, 25(1), 74–88. https://doi.org/
10.1093/jcmc/zmz026 .
Sundar, S. S., Kang, H., Wu, M., Go, E., & Zhang, B. (2013). Unlocking the privacy paradox: Do
cognitive heuristics hold the key?. In CHI’13 extended abstracts on human factors in computing
systems (pp. 811–816). Paris, France: CHI.
Sundar, S. S., & Lee, E. J. (2022). Rethinking communication in the era of artificial intelligence.
Human Communication Research, 48(3), 379–385. https://doi.org/10.1093/hcr/hqac014
Sundar, S. S., & Nass, C. (2000). Source orientation in human-computer interaction: Programmer,
networker, or independent social actor. Communication Research, 27(6), 683–703. https://doi.org/
10.1177/009365000027006001
Sun, M., Hu, W., & Wu, Y. (2022). Public perceptions and attitudes towards the application of
artificial intelligence in journalism: From a China-based survey. Journalism Practice, 1–23. https://
doi.org/10.1080/17512786.2022.2055621
Sun, Y., & Zhang, Y. (2018, July 8–12). Conversational recommender system. In The 41st interna-
tional acm sigir conference on research & development in information retrieval, Ann Arbor, MI,
USA (pp. 235–244). SIGIR.
Tolentino, D. (2023). TikTok CEO gives first public interview since congressional hearing. NBC
News. Retrieved from https://www.nbcnews.com/tech/tiktok-ceo-ted2023-conference-rcna80760
Tong, S., & Walther, J. B. (2011). Relational maintenance and CMC. Computer-Mediated
Communication in Personal Relationships, 53(9), 1689–1699.
Tubaro, P., Casilli, A. A., & Coville, M. (2020). The trainer, the verifier, the imitator: Three ways in
which human platform workers support artificial intelligence. Big Data & Society, 7(1),
2053951720919776. https://doi.org/10.1177/2053951720919776
Umamaheswari, S., Valarmathi, A., & Lackshmi, R.(2023). Role of artificial intelligence in the bank-
ing sector. Journal of Survey in Fisheries Sciences, 10(4S), 2841–2849. https://doi.org/10.17762/sfs.
v10i4S.1722
POLITICAL COMMUNICATION 17
UNCTAD. (2017) The information economy report 2017: Digitalization, Trade and Development,
UNCTAD division on Technology and Logistics, ICT Analysis Section.
Vergeer, M. (2020). Artificial intelligence in the Dutch press: An analysis of topics and trends.
Communication Studies, 71(3), 373–392. https://doi.org/10.1080/10510974.2020.1733038
Volovici, V., Syn, N. L., Ercole, A., Zhao, J. J., & Liu, N. (2022). Steps to avoid overuse and misuse of
machine learning in clinical research. Nature Medicine, 28(10), 1996–1999. https://doi.org/10.
1038/s41591-022-01961-6
Westerman, D., Edwards, A. P., Edwards, C., Luo, Z., & Spence, P. R. (2020). I-It, I-Thou, I-Robot:
The perceived humanness of AI in human-machine communication. Communication Studies, 71
(3), 393–408. https://doi.org/10.1080/10510974.2020.1749683
Wölker, A., & Powell, T. E. (2021). Algorithms in the newsroom? News readers’ perceived credibility
and selection of automated journalism. Journalism, 22(1), 86–103. https://doi.org/10.1177/
1464884918757072
Zhang, C., Zhang, C., Zheng, S., Qiao, Y., Li, C., Zhang, M., & Hong, C. S. (2023). A complete survey
on generative ai (aigc): Is chatgpt from gpt-4 to gpt-5 all you need? arXiv Preprint arXiv:
230311717.
18 H. GIL DE ZÚÑIGA ET AL.