ArticlePDF Available

AI Chatbots in Digital Mental Health

MDPI
Informatics
Authors:
  • Australian Institute for Suicide Research and Prevention

Abstract

Artificial intelligence (AI) chatbots have gained prominence since 2022. Powered by big data, natural language processing (NLP) and machine learning (ML) algorithms, they offer the potential to expand capabilities, improve productivity and provide guidance and support in various domains. Human–Artificial Intelligence (HAI) is proposed to help with the integration of human values, empathy and ethical considerations into AI in order to address the limitations of AI chatbots and enhance their effectiveness. Mental health is a critical global concern, with a substantial impact on individuals, communities and economies. Digital mental health solutions, leveraging AI and ML, have emerged to address the challenges of access, stigma and cost in mental health care. Despite their potential, ethical and legal implications surrounding these technologies remain uncertain. This narrative literature review explores the potential of AI chatbots to revolutionize digital mental health while emphasizing the need for ethical, responsible and trustworthy AI algorithms. The review is guided by three key research questions: the impact of AI chatbots on technology integration, the balance between benefits and harms, and the mitigation of bias and prejudice in AI applications. Methodologically, the review involves extensive database and search engine searches, utilizing keywords related to AI chatbots and digital mental health. Peer-reviewed journal articles and media sources were purposively selected to address the research questions, resulting in a comprehensive analysis of the current state of knowledge on this evolving topic. In conclusion, AI chatbots hold promise in transforming digital mental health but must navigate complex ethical and practical challenges. The integration of HAI principles, responsible regulation and scoping reviews are crucial to maximizing their benefits while minimizing potential risks. Collaborative approaches and modern educational solutions may enhance responsible use and mitigate biases in AI applications, ensuring a more inclusive and effective digital mental health landscape.
Citation: Balcombe, L. AI Chatbots in
Digital Mental Health. Informatics
2023,10, 82. https://doi.org/
10.3390/informatics10040082
Academic Editors: Antony Bryant,
Roberto Montemanni, Min Chen,
Paolo Bellavista, Kenji Suzuki,
Jeanine Treffers-Daller and
Roberto Theron
Received: 3 September 2023
Revised: 17 October 2023
Accepted: 25 October 2023
Published: 27 October 2023
Copyright: © 2023 by the author.
Licensee MDPI, Basel, Switzerland.
This article is an open access article
distributed under the terms and
conditions of the Creative Commons
Attribution (CC BY) license (https://
creativecommons.org/licenses/by/
4.0/).
informatics
Review
AI Chatbots in Digital Mental Health
Luke Balcombe
Australian Institute for Suicide Research and Prevention, School of Applied Psychology, Griffith University,
Messines Ridge Road, Mount Gravatt, QLD 4122, Australia; lukebalcombe@gmail.com
Abstract:
Artificial intelligence (AI) chatbots have gained prominence since 2022. Powered by big
data, natural language processing (NLP) and machine learning (ML) algorithms, they offer the
potential to expand capabilities, improve productivity and provide guidance and support in various
domains. Human–Artificial Intelligence (HAI) is proposed to help with the integration of human
values, empathy and ethical considerations into AI in order to address the limitations of AI chatbots
and enhance their effectiveness. Mental health is a critical global concern, with a substantial impact
on individuals, communities and economies. Digital mental health solutions, leveraging AI and ML,
have emerged to address the challenges of access, stigma and cost in mental health care. Despite
their potential, ethical and legal implications surrounding these technologies remain uncertain. This
narrative literature review explores the potential of AI chatbots to revolutionize digital mental health
while emphasizing the need for ethical, responsible and trustworthy AI algorithms. The review is
guided by three key research questions: the impact of AI chatbots on technology integration, the
balance between benefits and harms, and the mitigation of bias and prejudice in AI applications.
Methodologically, the review involves extensive database and search engine searches, utilizing
keywords related to AI chatbots and digital mental health. Peer-reviewed journal articles and media
sources were purposively selected to address the research questions, resulting in a comprehensive
analysis of the current state of knowledge on this evolving topic. In conclusion, AI chatbots hold
promise in transforming digital mental health but must navigate complex ethical and practical
challenges. The integration of HAI principles, responsible regulation and scoping reviews are crucial
to maximizing their benefits while minimizing potential risks. Collaborative approaches and modern
educational solutions may enhance responsible use and mitigate biases in AI applications, ensuring a
more inclusive and effective digital mental health landscape.
Keywords: human–artificial intelligence; AI chatbots; digital mental health; mental health care
1. Introduction
Artificial intelligence (AI) chatbots are intelligent conversational computer systems
that think, learn and complete tasks in combination with humans or independently, us-
ing big data, natural language processing (NLP) and machine learning (ML) algorithms
to expand their capabilities, improve productivity and provide conversation, guidance
and support [
1
,
2
]. Also indicated as conversational agents or generative AI using large
language models, they are a result of progress in the past 15 years in the fields of robotics,
ML, AI models and NLP. AI chatbots became eminent since the launch of ChatGPT in
November 2022 [3].
There are opportunities where AI chatbots can provide insightful responses beyond
human capacity. However, they may lack a personalized and empathetic touch. It is
proposed that human–artificial intelligence (HAI) may help overcome such limitations,
whereby humans and AI enable each other’s strengths to collaborate on a common task or
goal for efficient, safer, sustainable and enjoyable work and lives. The HAI concept aligns
with the Center of Humane Technology’s work with integrating human values, such as
empathy, compassion and responsibility in AI [4].
Informatics 2023,10, 82. https://doi.org/10.3390/informatics10040082 https://www.mdpi.com/journal/informatics
Informatics 2023,10, 82 2 of 16
Mental health is a critical issue that affects many millions of people worldwide [
5
].
As an example, around 20% of Australian adults have a mental disorder, which increases
to 44% when the experience of mental illness is considered over a lifetime [
6
], costing the
economy billions of dollars [
7
] because of diminished health and reduced life expectancy [
8
].
Unfortunately, many people do not receive the help they need due to various barriers such
as a lack of access to mental health services, stigma and cost [
9
12
]. Digital mental health
solutions target young people with technology for mental health assessment, support,
prevention and treatment [
13
]. For example, AI and ML models are used in the prediction
of mental illness [
14
], and AI chatbots are used for psychological support [
15
]. However,
there is uncertainty around the ethical and legal implications of these tools.
The aim of this narrative literature review is to demonstrate how the potential for
AI chatbots to assist various populations with accessible digital mental health through
relevant, scalable and sustainable data-driven insights is challenged by the need for creating
ethical, responsible and trustworthy AI algorithms.
2. Methods
This narrative literature review is adapted from the four steps outlined by
Demiris et al. [16]:
(1) Conduct a search of numerous databases and search engines; (2) Identify and use perti-
nent keywords from relevant articles; (3) Review the abstracts and text of relevant articles
and include those that address the research aim; and (4) Document results by summarizing
and synthesizing the findings and integrating them into the review.
The heterogeneity of the topic prevented a systematic review. In addition, the topic
of AI chatbots in digital mental health is still evolving, and there were not enough studies
that meet the strict criteria of a systematic review. Additionally, the field of digital mental
health is interdisciplinary, incorporating aspects of psychology, technology and health
care. This leads to a wide range of research approaches, methodologies and study designs,
making it challenging to apply strict systematic review criteria. Instead, a purposively
selected articles approach showed an educational approach to how AI chatbots may impact
digital mental health. By purposively selecting articles that aligned with the research aim
and reviewing them in a comprehensive manner, valuable insights were gathered and are
presented in a coherent narrative literature review. This approach allowed for flexibility in
considering various perspectives and ideas within the topic, contributing to a more holistic
understanding of the subject matter.
The selection of peer-reviewed journal articles, media articles and conference proceed-
ings were retrieved from searches of computerized databases, purposive online searches
and authoritative texts based on an assessment of three of the research questions posed in
the Editorial, “AI Chatbots: Threat or Opportunity?” [
3
]. These research questions were
used as a guide to explore the topic of interest, because it is not yet possible to arrive at a
comprehensive understanding of the state of the science:
1.
The development of AI chatbots has been claimed to herald a new era, offering signif-
icant advances in the incorporation of technology into people’s lives and interactions.
Is this likely to be the case, and if so, where will these impacts be the most pervasive
and effective?
2.
Is it possible to strike a balance regarding the impact of these technologies so that any
potential harms are minimized while potential benefits are maximized and shared?
3.
A growing body of evidence shows that the design and implementation of many
AI applications, i.e., algorithms, incorporate bias and prejudice. How can this be
countered and corrected?
The database searches used were Scopus, ScienceDirect, Sage and the Association for
Computing Machinery (ACM) Digital Library. The search engines used were PubMed,
Google Scholar and IEEE Xplore. The search terms used were “AI chatbots” OR “generative
artificial intelligence” OR “conversational agents” AND “digital mental health” OR “mental
health care”.
The following selection of criteria were used:
Informatics 2023,10, 82 3 of 16
Inclusion criteria:
Studies that have been published in peer-reviewed journals, media articles and
conference proceedings.
Studies that have been published in the English language.
Studies that have been published between 2010 and 2023.
Studies that have investigated the use of AI chatbots, generative artificial intelligence
or conversational agents in digital mental health or mental health care.
Studies that have reported on the effectiveness of AI chatbots, generative artificial
intelligence or conversational agents in digital mental health or mental health care.
Exclusion criteria:
Studies that are not published in peer-reviewed journals, media articles and
conference proceedings.
Studies that are not published in the English language.
Studies that are published before 2010 or after 2023.
Studies that do not investigate the use of AI chatbots, generative artificial intelligence
or conversational agents in digital mental health or mental health care.
Studies that do not report on the effectiveness of AI chatbots, generative artificial
intelligence or conversational agents in digital mental health or mental health care.
Boolean operators such as AND and OR were used to combine search terms and refine
search results. For example, using the Boolean operator OR between “AI chatbots” and
“generative artificial intelligence” retrieved articles that contain either one of these terms.
Similarly, using the Boolean operator AND between “conversational agents” and “digital
mental health” retrieved articles that contain both these terms. Boolean operators helped to
narrow down search results and make them more relevant to the research question.
Relevant articles and their reference lists were explored based on (1) relevance to the
guiding research questions, (2) showing examples of theoretical and empirical research
and development and (3) highlighting issues and possible solutions. These articles were
applied in a best-evidence synthesis for a complete, critical and objective analysis of the
current knowledge on the topic. Overall, the method shows a systematic and transparent
approach that minimizes bias by ensuring a comprehensive search, focusing on relevant
articles and presenting a fair synthesis of findings. However, it is important to note that
bias can still exist in the literature itself; therefore, the studies were critically evaluated, and
any potential limitations or biases within the selected articles were acknowledged.
3. Results
3.1. The Impact of AI Chatbots on Technology Integration
Research Question 1: The development of AI chatbots has been claimed to herald a
new era, offering significant advances in the incorporation of technology into people’s lives
and interactions. Is this likely to be the case, and if so, where will these impacts be the most
pervasive and effective [3]?
The use of AI chatbots has the potential to bring significant advances and impact
various aspects of people’s lives and interactions [
17
], especially where human to human
interaction is not preferred or possible to obtain [
18
]. AI chatbots may provide customer
service and support, health care and mental health support, education and e-learning,
personal productivity and assistance, language translation and communication as well as
social companionship and entertainment [
19
,
20
]. The diversity of uses for AI chatbots and
the large mix of empirical literature means it is reasonable to focus on one area.
Mental health care is a good example, because AI chatbots have been considered a viable
resource in this domain for more than a decade [
21
]. There are promising clinical outcomes
for AI chatbots providing relevant and continually accessible
support [22,23]
for depression
in adults [
24
], anxiety in university students [
25
,
26
] and attention-deficit/hyperactivity
symptoms for adults [27].
Informatics 2023,10, 82 4 of 16
AI chatbots may help address the barriers to the help-seeking process for mental health
issues by offering personalized, accessible, affordable and stigma-free assistance, promoting
early intervention and generating valuable insights for research and policymaking [
28
31
].
AI chatbots may be particularly useful in monitoring, communication, memory assistance,
screening and diagnosis, with the aim of understanding a patient’s emotions and assisting
in the analysis of large datasets. For example, algorithms may identify patterns and trends
that might be missed by human analysts. By analysing a patient’s medical history, genetic
data and other relevant factors, algorithms could generate tailored symptom checks and
treatment recommendations that consider the individual’s unique needs and circumstances.
The opportunities for AI chatbots should also be considered in terms of the challenges
posed, such as a lack of human connection, a reliance on technology, the accuracy and
reliability of information, ethical and privacy considerations as well as misdiagnosis and
limited understanding [2831].
A 2023 overview of mental health chatbots found 10 apps on the market for a variety
of mental health concerns (e.g., anxiety and depression) and users (e.g., rural dwellers, shift
workers, students, veterans and adolescents), for a variety of aims (e.g., to improve social or
job interviewing skills) [
18
]. The overview took interest in AI chatbots for their accessible,
affordable and convenient social and psychological support. However, vulnerable users
may overrate the benefits and encounter risks, especially during a crisis, because AI chatbots
were allegedly incapable of identifying crisis situations. Therefore, poor semantics were
found to undermine AI chatbots, because they were not developed enough to understand
the context of users’ words and failed to respond effectively or at all.
Users may not be aware of the difference between humans and humanlike chatbots.
These limitations are human factors, of which education is the key for effectively collabo-
rating to produce sustainable solutions [
32
]. Users and practitioners need guidance on the
uses of AI chatbots, similar to what is generally required for digital mental health platforms
and interventions [33].
The different fields of psychology, psychiatry, AI and health care, as well as educa-
tors, policymakers, computer scientists and technology developers working on mental
health care, means there are significant challenges to overcome in order to realize overall
benefits [
34
]. Mental health professionals and policymakers hold the key to AI chatbots
being a useful tool in the intelligent system toolbox. However, it appears that graduate
students and research scientists may best drive change through their willingness and ability
to effectively collaborate with computer scientists and technology developers.
AI chatbots offer promise as complementary tools rather than a replacement for human
mental health professionals [
18
,
20
]. A 2021 review of digital mental health interventions
(DMHIs) found AI chatbots to speculatively help mental health professionals meet over-
whelming service demand [
34
]. A 2023 systematic review and meta-analysis of randomized
controlled trials (RCTs) found AI chatbots to be acceptable for a wide range of mental health
problems [
35
]. For example, an RCT found a fully automated conversational agent, Woebot,
to be a feasible, engaging and effective way to deliver cognitive behavioural therapy (CBT)
for anxiety and depression in young adults [
25
]. There is promise for Woebot [
36
] and
Wysa [37] in establishing a therapeutic bond with users.
Although AI chatbots are feasible as an engaging and acceptable way to deliver therapy,
more studies are required for what may facilitate a digital therapeutic alliance [
36
,
37
]
and to reduce misunderstandings [
38
]. Mental health chatbot attrition rates are lower in
comparison to other digital interventions [
24
,
39
]. However, dropout rates require attention,
as does clarity around what disorders they are useful for [
40
]. Some reviews found a high
potential for AI chatbots in identifying patients at risk of suicide [
41
43
], and triage and
treatment development through NLP integrated to social media in real-time [4446].
Developments in Generative Pre-Trained Transformer (GPT) programs like ChatGPT
4 means AI chatbots may be used in suicide prevention [
47
]. However, there is a need for
better understanding AI chatbot limitations such as negative sentiment, constrictive think-
ing, idioms, hallucinations and logical fallacies. A study of messages related to people’s
Informatics 2023,10, 82 5 of 16
suicidal thoughts sought insights from the arrangement of their words, the sentiment and
rationale [
48
]. While AI chatbot hallucinations and fallacies require human intervention, it
is possible to detect idioms, negative sentiment and constrictive language with off-the-shelf
algorithms and publicly available data. However, safety concerns were publicized after a
chatbot, Eliza, was blamed by a Belgian man’s widow for her husband’s suicide [49].
There is a need for qualitative studies to help reduce poor semantics and errors as well
as increase trust in AI chatbots. For example, thematic analysis from retrospective data is
required to identify common themes of messages sent to mental health chatbots in order
to increase the effectiveness of AI chatbots as a source of support. AI chatbots may help
improve problem areas through NLP for sentiment analysis, which is fast and effective
qualitative data analysis, to assist in understanding multidimensional online feedback.
Recommendations for identifying and evaluating the impact of AI chatbots are
as follows:
Conduct qualitative studies using AI chatbots to demonstrate how they assist with
accessibility, engagement and effectiveness through (1) identifying user needs,
(2) understanding
barriers to its use, (3) evaluating user experience and AI chatbot
impact and (4) integrating human–AI approaches to overcome problem areas.
Contribute to empirical evidence with longitudinal studies and RCTs to see which
mental health conditions and populations AI chatbots may be recommended for.
Determine a practical attrition prediction possibility to identify individuals at a high
risk of dropping out through applying advanced machine learning models (e.g.,
deep neural networks) to the leveraging analyses of feature sets (e.g., baseline user
characteristics, self-reported user context and AI chatbot feedback, passively detected
user behaviour and the clinical functioning of users).
3.2. The Balance between the Benefits and Harms of AI Chatbots
This is difficult to answer on a global scale because of a lack of widely collaborative
international standards in addition to the diversity of applications for AI chatbots. However,
current investment in AI research, education, societal adaptation, innovation, employment
opportunities and jobs creation appear to be insufficient upon considering the scale of the
impending changes.
The novelty and complexity of AI in mental health means it is timely to focus on
cutting-edge education such as specialized university courses in digital mental health
and informatics that use peer-reviewed and routinely updated textbooks and modules.
The intent should be to stimulate discerning skills and critical thought from a mix of
subjects that will assist in pioneering benefits to mental health care and AI technology
industries while also mitigating the increasing costs from mental illness. While AI chatbots
are eminent, they have yet to reach their potential in assisting with mental health problems
in digital users, who are predominantly young people [18].
Quality, effective and usable chatbots such as Woebot and Wysa are available to assist
with mental health support [
36
,
37
]. However, various studies are needed to show evidence
for a broader array of mental health disorders and symptoms. Furthermore, development
is mostly being driven by the technology side with an interest in mental health rather
than by mental health professionals who are technologically savvy. The differences in
communication styles and methodologies between technology and mental health care
researchers (i.e., pattern-based versus hypothesis-derived) has limited the combination of
these approaches. Another hindrance is the limited opportunities for high-level researchers
who are capable of understanding and implementing hybrid methods.
However, there is good potential for mental health care to serve as an example where
AI chatbots may assist in providing (cost-)effective solutions for a range of users and
aims [
21
24
,
36
38
]. Mental health care professionals may need to embrace AI chatbots
for their use to become more productive [
50
]. There also needs to be conscious efforts to
broaden the way in which productivity is measured if significant advances integrating
technology into people’s lives and interactions are to be realized. For example, how can AI
Informatics 2023,10, 82 6 of 16
chatbots’ contribution to the economy and the health of people be accurately measured?
While there may be a gain to the gross domestic product (GDP) of developed countries, there
may also be some job losses because of AI disruption. Although productivity, affordability
and accessibility are important levers, so are policies that consider mental health and
human capital.
The impact of AI chatbots on productivity needs to be considered in terms of national
and international economics, standards and regulations. It is apparent that not all gov-
ernments are aligned in principle. Also, the digital divide is questionable in terms of not
further marginalizing the underserved and the unserved [
34
]. Therefore, productivity and
humanity need to be considered in terms of global risks such as war and the costs that the
physical effects of climate change will bring [
50
]. While some governments heavily invest
in defence, the decarbonisation of heavy industries and transition among energy systems,
there will be competing demands for investment into AI technologies. Meanwhile, the
example of the emergence of ChatGPT shows the difficulties of stakeholders grappling
with the pace of technological development.
It is unclear how the Productivity Commission’s forecasts of AI boosting the Australian
economy have been calculated to arrive at a predicted boon between a 66.67% and 266.67%
higher GDP in the next decade [
7
]. In 2023, the Australian Government projected an outlook
for the next 40 years in terms of intergenerational equity, with forecasts of higher financial
burdens on younger generations [
51
]. This leads to the question of how such countries
manage and maximize the major shifts that are underway in their national economy while
also effectively integrating the impact of AI technologies.
In the example of mental health care in Australia, it is necessary to explore the existing
structure for safety and quality to see AI’s consistency with it before examining its economic
potential. Australia’s National Standards in mental health services provide a framework
for safety and quality in hospitals and community services and are primarily intended to
regulate the practice of mental health professionals [
52
]. Yet, with overwhelming demand
and limited supply in mental health care exacerbated by further strain during the COVID-19
pandemic [
53
,
54
], digital mental health filled a service gap, culminating in the development
of the National Safety and Quality Digital Mental Health Standards in 2020, which aimed
to improve the safety and quality of digital mental health service provision [
55
]. However,
mental health professionals and policymakers are currently confronting the opportunities
and challenges of AI [
56
]. For example, prompt engineering is used with ChatGPT to
bypass content filters in social media. This could lead to harm and problems such as
exploiting vulnerability.
The Australian Government adopted a voluntary ethics framework in 2018 for “re-
sponsible” AI, in order to guide businesses and governments to responsibly design, develop
and implement AI [
57
]. However, mainstream AI chatbots are mostly being developed in
the US. Australia and the US are among the various countries seeking input or planning on
AI chatbot regulation [
58
]. The EU implemented the Digital Services Act and the Digital
Market Act, aiming to create a safer digital space where the fundamental rights of users
are protected and to establish a level playing field for businesses [
59
]. There is a need to
ensure that AI algorithms are developed and trained using diverse and representative
datasets and that any insights generated by AI are rigorously validated and verified by
human experts. ChatGPT’s owners, OpenAI, suggested proactively managing the risks of
these “frontier AI” models [
60
]. OpenAI initially proposed conducting pre-deployment
risk assessments, external scrutiny of model behaviour, using risk assessments to inform
deployment decisions and monitoring and responding to new information about model
capabilities and uses post deployment.
It is essentially up to users to be transparent about their use of AI, take steps to
protect privacy and confidentiality and take care to use it responsibly for optimising its
performance [
61
]. For example, recruits may use AI chatbots to accomplish their duties,
which detracts from the intent of seeking human input and raises critical questions about
the value of cooperative work if trust cannot be established and maintained [
62
]. A
Informatics 2023,10, 82 7 of 16
main problem with AI chatbots is that they are a new technology with the potential of
becoming fundamentally pervasive in terms of cybersecurity risk because of their ability to
create all sorts of malicious codes and algorithms that can cause infrastructure or financial
system chaos [63].
The use of AI in mental health research has been well established as potentially
yielding important insights and improving outcomes for individuals with mental health
disorders [
64
,
65
]. However, it is important to carefully classify and regulate “high” risks
and prioritise ethical considerations at every step. The increasing use of AI chatbots for
mental health and crisis support means that stakeholders need to increase their attention
and education in order to effectively leverage these tools [
18
,
66
]. For example, fair aware
AI has been called for in digital mental health to promote diversity and inclusion [
67
],
and explainable AI has been suggested as a tool for demonstrating transparency and trust
between users and practitioners [68].
It is proposed that HAI may complement these concepts in an evolving AI system
where multiple AI models work together with human input to generate recommendations
and predictions rather than relying on a single algorithm. A next step is establishing
optimal combinations of humans and AI chatbots for various tasks in research, practice
and policy [
69
]. However, it is necessary to consider AI technologies overall in terms of
plans for broad-ranging regulation.
According to Australia’s AI Ethics Principles [
70
], regulation can help achieve safer,
more reliable and fairer outcomes for all Australians; reduce the risk of negative impacts
on those affected by AI applications; and encourage businesses and governments to prac-
tice the highest ethical standards when designing, developing and implementing AI. A
subsequent position statement on generative AI suggested that regulation can help ad-
dress concerns about potential harms such as algorithmic bias and errors, the spread of
misinformation, inappropriate content and the creation of deepfakes [71].
By implementing measures such as transparency, accountability and risk mitigation
strategies, regulation can help ensure that AI is used responsibly and ethically [
72
]. Fur-
thermore, regulation can help raise public trust in AI technologies by ensuring that they are
developed and used in a manner that is consistent with societal values and expectations [
73
].
This can help facilitate the adoption of AI technologies and enable society to fully realize
their potential benefits.
The regulation of AI should include defining what constitutes “unsafe” AI and de-
termining what aspects of AI should be subject to regulation [
74
]. This requires a clear
understanding of the anticipated risks and benefits of AI technologies on a global scale as
well as insights into the public’s trust and the acceptance of AI systems. While people in
Western countries are more cautious of AI and less confident that the benefits outweigh the
risks, those in the emerging economies (i.e., Brazil, India, China and South Africa) are more
trusting and embracing of AI, in addition to young, university-educated people as well as
those in managerial roles [75].
Overly stringent regulations could stifle innovation and hinder the development of AI
technologies [
76
]. As such, regulation requires international cooperation to be truly effec-
tive. Without a global consensus, companies might simply migrate their AI development
activities to less regulated jurisdictions, leading to a regulatory race to the bottom. There
is a need to secure AI models and their associated systems by using industry-standard
security protocols. AI models and systems should be regularly updated and patched to
address any discovered vulnerabilities.
Recommendations for regulating and/or promoting the responsible use of AI applica-
tions are as follows:
Invest in research to evaluate the efficacy and potential harms of AI applications and
develop systems to monitor and audit AI systems for unusual or suspicious activity.
Implement rigorous safety measures, robust regulations and collaborative standards
to ensure the responsible use of AI technologies.
Informatics 2023,10, 82 8 of 16
Validate a HAI model combining AI chatbots with human experts in research, practice
and policy to optimise mental health care assistance.
3.3. The Mitigation of Bias and Prejudice in AI Applications
The World Health Organisation offered a warning, stating that the use of generative AI
for health care must be approached with caution [
77
]. AI algorithms are only as good as the
data they are trained on, and biases in the data can lead to biased results [
78
]. Additionally,
the use of AI in mental health care raises important risks and ethical considerations [
79
] as
well as security, bias and privacy concerns, particularly when it comes to the storage and
use of sensitive medical and personal data [80].
More generally, there are “high-risk” uses of AI and automated decision-making,
which warrant the warning of potential harms, including the creation of deepfakes and
algorithmic bias [
81
]. There is also concern about AI perpetuating or amplifying biases or
narrow viewpoints [
82
,
83
] as well as automating jobs and therefore replacing humans in
some capacities [
84
]. However, AI can be used to counter disinformation and to improve
the accuracy and reliability of reporting [
85
,
86
]. The challenge lies in defining and deciding
what constitutes “unsafe” AI. Various Australian science experts have called for rigorous
safety measures, robust regulations and standards to be implemented for these “unsafe”
AI [
76
]. It is apparent that mitigating measures for high-risk AI should be quickly and
proactively sought to avoid hampering progress in AI.
Generative AI is being used in the media to create more personalised and targeted
advertising, to automate content creation and curation and to analyse audience behaviour
and preferences [87,88]. Misinformation or disinformation may stem from tools like Chat-
GPT [
87
] in combination with social media, which produce the mass de-prioritization of
legitimate news outlets in favour of spam as well as false or manipulative user-uploaded
content [
87
]. Bias and errors in generative AI [
67
,
87
] highlight the questionability of existing
information assessment guidelines because of evidence credibility, source transparency
and limitation acknowledgment. Generative AI has shown the need for new guidelines to
promote ethics, fairness, privacy and transparency [
76
] as well as recognize the intellectual
property rights of human creators and organizations [
89
]. This may be exacerbated by po-
tentially anticompetitive practices used by dominant technology platforms such as Google
and Meta [88].
There is a need to counter and correct the AI applications that help perpetuate bias,
harassment and marginalization as well as the loss of critical thinking and independent
thought. AI chatbots may be a part of innovative solutions to answer calls for the de-
tection and moderation of fake news [
90
] and the transparent regulation of social media
platforms [
91
93
]. As an example, a review of the impact of YouTube on loneliness and
mental health found its recommendation algorithms may inadvertently reinforce existing
beliefs and biases, spread misinformation and disinformation as well as enable unhelpful
or harmful content [
46
]. However, the review also found that YouTube can have positive
effects on loneliness, anxiety and depression if users actively engage with the platform and
use it as a tool for education, social connection and emotional support.
There are opportunities for biased and prejudiced AI applications to be countered and
corrected through education and research with the assistance of AI chatbots [
94
]. However,
human researchers/experts who understand the history and context of research problems
may need to help prompt and supervise AI chatbots for solutions. For example, YouTube’s
recommendation algorithm is designed to suggest videos based on users’ viewing history,
search queries and other data points [95]. Since YouTube’s launch in 2005 up until 2011, it
was designed to recommend videos that attracted the most views or clicks. Then, in 2012,
it was designed to directly respond to metrics such as shares, likes and, to a lesser extent,
dislikes. From 2016 onwards, it was designed to increase safety, with efforts made to remove
dangerous content and demonetise those who did not follow regulations. However, the
development of AI chatbots means that continuous adaptation is critical through legislation
and setting ethical values [94] in addition to improving current AI systems [46].
Informatics 2023,10, 82 9 of 16
YouTube has initiated mental health policies, algorithm changes, content moderation,
content creator and user psychoeducation, mental health and crisis resource panels, self-
harm and suicide content warnings and parental controls and settings [
96
]. YouTube
reported largely reduced borderline content as it immediately removes offensive content
where detected [
97
]. However, the algorithm can also create filter bubbles and echo
chambers where users are exposed to content that reinforces their existing beliefs and
biases [
98
]. This can lead to polarisation and misinformation, which can have negative
consequences for mental health. Improved algorithms are called for to detect bias and
errors as well as moderate how videos appear in a watch list in order to steer users to
safe, well-informed and inclusive content as well as refer them to mental health and crisis
resource panels with suitable information and resources, in conjunction with assistance
from AI chatbots [46].
However, problematic social media use in young people affects one in three indi-
viduals in the Australian youth, and it is not only limited to YouTube [
99
]. For example,
cyberbullying is also an issue in other social media (e.g., Facebook, Twitter, Snapchat,
Instagram, TikTok) [
100
]. Various studies have found a clear link between heavy social
media use and an increased risk for depression, anxiety, loneliness, self-harm and suicidal
thoughts [
101
103
]. Although there is a lack of psychological studies on TikTok [
104
], a
causal study across American colleges found that access to Facebook led to an increase in
severe depression by 7% and anxiety disorder by 20% [
102
]. This significant link between
the presence of Facebook and a decline in mental health in young people is concerning
when considering the fact that the arrival of Facebook in 2004 was followed by an increase
of 57% in deaths by suicide among Americans aged 10–24 between 2007 and 2017 [105].
Major data breaches and the use of “psychological warfare tools” on Facebook were
reported in 2018 with the Cambridge Analytica files [
106
]. After calls were sounded for
data to be used following principles of ethics, privacy and security, Australia took the
international lead in social media regulation with the Online Safety Act 2021 after public
hearings revealed that Facebook’s algorithms were potentially harmful and unsafe [
107
].
However, in 2022, the Australian government and the technology industry realized that
an outdated classification system hindered the creation of new codes for regulating online
content [
108
]. In 2023, Australia expressed interest in pursuing risk-based classification
systems for AI chatbots as were being drafted in Canada and the EU [109].
Advances in AI chatbots and other tools such as predictive models and virtual assis-
tants means that multiple models may be combined with human expert input to address
mental health challenges and suicide prevention, improve access to care and reduce the
barriers to seeking help. These tools use NLP and ML to mine mental health data, to
understand and respond to individuals’ needs and to provide personalised support. A
theoretical framework proposed an adaptive Social Media Virtual Companion (SMVC) for
educating and supporting adolescent students in interactions in social media environments
in order to achieve a measure of collective well-being [
110
]. This SMVC framework is an
example of how to design social media systems and embedded educational interventions
through HAI, because automatic processing powered by a recommendation algorithm is
combined with educator/expert intervention and guidance.
HAI mental health strategies are proposed to be useful for the design and devel-
opment of a multi-model responsible social media system in education settings. For
example, an adaptive SMVC may assist in promoting the obtaining of more balanced and
diverse content as well as reducing the impact of bias and errors in algorithmic recom-
mendation systems such as filter bubbles and echo chambers. By integrating off-the-shelf
solutions like Viable for sentiment analysis and DataMinr for monitoring and analysing
social media, the SMVC system can learn from HAI feedback and recent data to adjust
recommendations accordingly.
However, AI-generated sentiment affects the emotional language used in human
conversation, therefore potentially affecting social relationships. Randomized experiments
found that algorithmic recommendation systems change how people interact with and
Informatics 2023,10, 82 10 of 16
perceive one another socially; people are evaluated more negatively if they are suspected of
using an algorithmic response from AI chatbots like ChatGPT [
111
]. Therefore, educators
should proactively and transparently encourage the use of AI chatbots to avoid negative
perceptions. Users may need to be taught how to be discerning and critical of the infor-
mation AI chatbots provide and learn how to effectively leverage these tools to help solve
complex problems in their studies as well as cautiously use them in self-care for mental
health—obtaining assistance where required.
Recommendations for countering and correcting the flaws of AI applications are
as follows:
Vulnerable people need more informed guidance on how to self-manage their mental
health when assisted by AI chatbots in order to connect with resources and treatments.
Social media mental health and crisis resource panels may be enhanced by linking to
AI chatbots that provide vetted digital mental health and crisis services or referrals
as necessary.
HAI mental health strategies with SVMC may be explored for cautiously navi-
gating a safer, more responsible social media with humane, fair and explainable
system recommendations.
4. Conclusions
This narrative literature review has explored the multifaceted impacts of AI chatbots
on various aspects of society, particularly focusing on their potential in the field of mental
health care. This review is useful for providing an overview of the topic, identifying gaps
in the literature and generating new research questions. By synthesizing both theoretical
and empirical research, this study provided a comprehensive overview of the current state
of AI chatbots in mental health care. The evidence presented indicates that AI chatbots
hold promise for revolutionizing mental health support, offering accessibility, engagement
and effectiveness in assisting individuals and populations with a wide range of mental
health concerns and aims. However, it is crucial to approach their implementation and
regulation with caution and responsibility. The novelty of AI chatbots in mental health
means that this narrative literature review shows examples of theoretical and empirical
research that future studies can apply.
The development of AI chatbots brings opportunities for serving underserved and
unserved populations, as well as blending care for the well-served, especially in treating
common disorders such as anxiety and depression. However, there are challenges in know-
ing which AI chatbots are of good quality and which are useful and effective. Therefore, it
is important for future research to clarify these areas as well as the level of care required for
crisis support. The human factors of human–computer interaction require more attention
through empirical research. AI chatbots offer accessible and convenient support, helping to
address the barriers in the help-seeking process for mental health issues and have shown
promise in various clinical trials. Nevertheless, limitations such as poor semantics, biases
and the need for qualitative studies to improve user experience must be acknowledged and
addressed. AI chatbots should be seen as complementary tools rather than replacements for
human mental health professionals. Notwithstanding, there is a need for more empirical
evidence and advocacy for users and practitioners to distinguish the quality, usability and
effectiveness of AI as well as its uses and the populations that would benefit from them. If
AI chatbots evolve to provide appropriate answers to these areas for clarification, then a
more autonomous use of these tools will become progressively possible.
Furthermore, there is a need for regulation and responsible use of AI applications,
given the potential for biases, privacy concerns and the amplification of misinformation.
This emphasizes the importance of international collaboration in establishing standards
and regulations to ensure the ethical and transparent use of AI technologies. The balance
between innovation and regulation must be carefully struck in order to avoid stifling
progress while safeguarding against potential harm.
Informatics 2023,10, 82 11 of 16
Additionally, the review highlights the role of collaborative AI in countering biases
and errors in AI applications, especially in the context of social media and mental health
support. By integrating human expertise and sentiment analysis into AI models, it becomes
possible to provide more balanced and diverse content while reducing the impact of
algorithmic biases.
Overall, the review shows promise for the use of AI chatbots in mental health care,
but it also highlights the need for further research, such as scoping reviews, to evaluate
their effectiveness and address the risk of bias and ethical concerns. It emphasizes the
need for careful consideration, research and collaboration in harnessing the potential of
AI chatbots. While they offer transformative possibilities in various domains, responsible
development, regulation and ongoing evaluation are essential to maximize their benefits
while minimizing risks. Collaborative efforts between technology developers, mental
health professionals, policymakers, researchers and educators can help ensure that AI
chatbots contribute positively to society’s well-being and mental health support.
Funding: This research received no external funding.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Not applicable.
Conflicts of Interest: The author declares no conflict of interest.
Abbreviations
ACM Association for Computing Machinery
AI artificial intelligence
CBT cognitive behavioural therapy
DMHIs digital mental health interventions
EU European Union
GDP gross domestic product
GPT Generative Pre-Trained Transformer
HAI Human–artificial intelligence
HCI human–computer interaction
IEEE the Institute of Electrical and Electronics Engineers
ML machine learning
NLP natural language processing
RCT randomized controlled trial
UK United Kingdom
US United States
WHO World Health Organization
References
1.
Team Capacity. The Complete Guide to AI Chatbots: The Future of AI and Automation. 2023. Available online: https:
//capacity.com/learn/ai- chatbots/ (accessed on 19 August 2023).
2. Caldarini, G.; Jaf, S.; McGarry, K. A Literature Survey of Recent Advances in Chatbots. Information 2022,13, 41. [CrossRef]
3. Bryant, A. AI Chatbots: Threat or Opportunity? Informatics 2023,10, 49. [CrossRef]
4.
The Center for Humane Technology. Align Technology with Humanity’s Best Interests. 2023. Available online: https://www.
humanetech.com/ (accessed on 19 August 2023).
5.
World Health Organization. Mental Health. 2023. Available online: https://www.who.int/health-topics/mental-health#tab=
tab_1 (accessed on 19 August 2023).
6.
Australian Bureau of Statistics. National Study of Mental Health and Wellbeing. 2021. Available online: https://www.abs.gov.
au/statistics/health/mental-health/national-study-mental-health-and-wellbeing/latest-release (accessed on 19 August 2023).
7.
Australian Productivity Commission. Mental Health. 2020. Available online: https://www.pc.gov.au/inquiries/completed/
mental-health#report (accessed on 19 August 2023).
8.
Queensland Brain Institute. Life Expectancy Mapped for People with Mental Disorders. 2019. Available online: https://qbi.uq.
edu.au/article/2019/10/life-expectancy-mapped-people- mental-disorders (accessed on 19 August 2023).
Informatics 2023,10, 82 12 of 16
9.
Clement, S.; Schauman, O.; Graham, T.; Maggioni, F.; Evans-Lacko, S.; Bezborodovs, N.; Morgan, C.; Rüsch, N.; Brown, J.S.L.;
Thornicroft, G. What is the impact of mental health-related stigma on help-seeking? A systematic review of quantitative and
qualitative studies. Psychol. Med. 2015,45, 11–27. [CrossRef]
10.
Oexle, N.; Müller, M.; Kawohl, W.; Xu, Z.; Viering, S.; Wyss, C.; Vetter, S.; Rüsch, N. Self-stigma as a barrier to recovery: A
longitudinal study. Eur. Arch. Psychiatry Clin. Neurosci. 2017,268, 209–212. [CrossRef]
11.
Australian Institute of Health and Welfare. Mental Health: Prevalence and Impact. 2022. Available online: https://www.aihw.
gov.au/reports/mental-health-services/mental- health (accessed on 19 August 2023).
12.
U.S. Department of Health and Human Services, Substance Abuse and Mental Health Services Administration. Key Substance Use
and Mental Health Indicators in the United States: Results from the 2018 National Survey on Drug Use and Health. 2018. Available
online: https://www.samhsa.gov/data/sites/default/files/cbhsq-reports/NSDUHDetailedTabs2018R2/NSDUHDetTabsSect8
pe2018.htm#tab8-28a (accessed on 19 August 2023).
13.
Wies, B.; Landers, C.; Ienca, M. Digital Mental Health for Young People: A Scoping Review of Ethical Promises and Challenges.
Front. Digit. Health 2021,3, 697072. [CrossRef]
14.
Iyortsuun, N.K.; Kim, S.-H.; Jhon, M.; Yang, H.-J.; Pant, S. A Review of Machine Learning and Deep Learning Approaches on
Mental Health Diagnosis. Healthcare 2023,11, 285. [CrossRef]
15.
Andreou, A. Generative AI Could Help Solve the U.S. Mental Health Crisis. Psychology Today. Available online: https://www.
psychologytoday.com/au/blog/the- doctor-of- the-future/202303/generative-ai-could-help- solve-the-us- mental-health-crisis
(accessed on 19 August 2023).
16.
Demiris, G.; Oliver, D.P.; Washington, K.T. The Foundations of Behavioral Intervention Research in Hospice and Palliative Care.
In Behavioral Intervention Research in Hospice and Palliative Care; Academic Press: Cambridge, MA, USA, 2019; pp. 17–25. [CrossRef]
17.
Adamopoulou, E.; Moussiades, L. Chatbots: History, technology, and applications. Mach. Learn. Appl.
2020
,2, 100006. [CrossRef]
18.
Haque, M.D.R.; Rubya, S. An Overview of Chatbot-Based Mobile Mental Health Apps: Insights from App Description and User
Reviews. JMIR mHealth uHealth 2023,11, e44838. [CrossRef] [PubMed]
19.
Denecke, K.; Abd-Alrazaq, A.; Househ, M. Artificial Intelligence for Chatbots in Mental Health: Opportunities and Challenges. In
Multiple Perspectives on Artificial Intelligence in Healthcare: Opportunities and Challenges; Lecture Notes in Bioengineering; Springer:
Berlin/Heidelberg, Germany, 2021; pp. 115–128. [CrossRef]
20.
Rizvi, M. AI Chatbots Revolutionize Depression Management and Mental Health Support—DATAVERSITY. 2023. Available
online: https://www.dataversity.net/ai-chatbots-revolutionize-depression-management- and-mental-health- support/ (accessed
on 21 August 2023).
21.
Vaidyam, A.N.; Wisniewski, H.; Halamka, J.D.; Kashavan, M.S.; Torous, J.B. Chatbots and Conversational Agents in Mental
Health: A Review of the Psychiatric Landscape. Can. J. Psychiatry 2019,64, 456–464. [CrossRef] [PubMed]
22.
Daley, K.; Hungerbuehler, I.; Cavanagh, K.; Claro, H.G.; Swinton, P.A.; Kapps, M. Preliminary Evaluation of the Engagement and
Effectiveness of a Mental Health Chatbot. Front. Digit. Health 2020,2, 576361. [CrossRef] [PubMed]
23.
Inkster, B.; Sarda, S.; Subramanian, V. An Empathy-Driven, Conversational Artificial Intelligence Agent (Wysa) for Digital Mental
Well-Being: Real-World Data Evaluation Mixed-Methods Study. JMIR mHealth uHealth 2018,6, e12106. [CrossRef]
24.
Lim, S.M.; Shiau, C.W.C.; Cheng, L.J.; Lau, Y. Chatbot-Delivered Psychotherapy for Adults with Depressive and Anxiety
Symptoms: A Systematic Review and Meta-Regression. Behav. Ther. 2022,53, 334–347. [CrossRef]
25.
Fitzpatrick, K.K.; Darcy, A.; Vierhile, M. Delivering Cognitive Behavior Therapy to Young Adults with Symptoms of Depression
and Anxiety Using a Fully Automated Conversational Agent (Woebot): A Randomized Controlled Trial. JMIR Ment. Health
2017
,
4, e19. [CrossRef]
26.
Klos, M.C.; Escoredo, M.; Joerin, A.; Lemos, V.N.; Rauws, M.; Bunge, E.L. Artificial Intelligence–Based Chatbot for Anxiety and
Depression in University Students: Pilot Randomized Controlled Trial. JMIR Form. Res. 2021,5, e20678. [CrossRef]
27.
Jang, S.; Kim, J.-J.; Kim, S.-J.; Hong, J.; Kim, S.; Kim, E. Mobile app-based chatbot to deliver cognitive behavioral therapy and
psychoeducation for adults with attention deficit: A development and feasibility/usability study. Int. J. Med. Inform.
2021
,
150, 104440. [CrossRef]
28.
Viduani, A.; Cosenza, V.; Araújo, R.M.; Kieling, C. Chatbots in the Field of Mental Health: Challenges and Opportunities. In
Digital Mental Health; Springer: Berlin/Heidelberg, Germany, 2023; pp. 133–148. [CrossRef]
29.
Helmy, B.S.; Helmy, A.S. Role of Artificial Intelligence in Mental Wellbeing: Opportunities and Challenges. J. Artif. Intell.
2022
,
15, 1–8. [CrossRef]
30.
Singh, O. Artificial intelligence in the era of ChatGPT—Opportunities and challenges in mental health care. Indian J. Psychiatry
2023,65, 297–298. [CrossRef]
31.
Boucher, E.M.; Harake, N.R.; Ward, H.E.; Stoeckl, S.E.; Vargas, J.; Minkel, J.; Parks, A.C.; Zilca, R. Artificially intelligent chatbots
in digital mental health interventions: A review. Expert Rev. Med. Devices 2021,18, 37–49. [CrossRef] [PubMed]
32. Balcombe, L.; De Leo, D. Human-Computer Interaction in Digital Mental Health. Informatics 2022,9, 14. [CrossRef]
33.
Balcombe, L.; De Leo, D. Evaluation of the Use of Digital Platforms and Digital Mental Health Interventions: Scoping Review. Int.
J. Environ. Res. Public Health 2022,20, 362. [CrossRef] [PubMed]
34.
Balcombe, L.; De Leo, D. Digital Mental Health Challenges and the Horizon Ahead for Solutions. JMIR Ment. Health
2021
,
8, e26811. [CrossRef]
Informatics 2023,10, 82 13 of 16
35.
He, Y.; Yang, L.; Qian, C.; Li, T.; Su, Z.; Zhang, Q.; Hou, X. Conversational Agent Interventions for Mental Health Problems:
Systematic Review and Meta-analysis of Randomized Controlled Trials. J. Med. Internet Res. 2023,25, e43862. [CrossRef]
36.
Darcy, A.; Daniels, J.; Salinger, D.; Wicks, P.; Robinson, A. Evidence of Human-Level Bonds Established with a Digital Conversa-
tional Agent: Cross-sectional, Retrospective Observational Study. JMIR Form. Res. 2021,5, e27868. [CrossRef] [PubMed]
37.
Beatty, C.; Malik, T.; Meheli, S.; Sinha, C. Evaluating the Therapeutic Alliance with a Free-Text CBT Conversational Agent (Wysa):
A Mixed-Methods Study. Front. Digit. Health 2022,4, 847991. [CrossRef] [PubMed]
38.
Dosovitsky, G.; Bunge, E.L. Bonding with Bot: User Feedback on a Chatbot for Social Isolation. Front. Digit. Health
2021
,3, 735053.
[CrossRef] [PubMed]
39.
Sinha, C.; Cheng, A.L.; Kadaba, M. Adherence and Engagement with a Cognitive Behavioral Therapy–Based Conversational
Agent (Wysa for Chronic Pain) among Adults with Chronic Pain: Survival Analysis. JMIR Form. Res.
2022
,6, e37302. [CrossRef]
40.
Prochaska, J.J.; Vogel, E.A.; Chieng, A.; Kendra, M.; Baiocchi, M.; Pajarito, S.; Robinson, A. A Therapeutic Relational Agent
for Reducing Problematic Substance Use (Woebot): Development and Usability Study. J. Med. Internet Res.
2021
,23, e24850.
[CrossRef]
41.
Martínez-Miranda, J. Embodied Conversational Agents for the Detection and Prevention of Suicidal Behaviour: Current
Applications and Open Challenges. J. Med. Syst. 2017,41, 135. [CrossRef]
42.
Laranjo, L.; Dunn, A.G.; Tong, H.L.; Kocaballi, A.B.; Chen, J.; Bashir, R.; Surian, D.; Gallego, B.; Magrabi, F.; Lau, A.Y.S.; et al.
Conversational agents in healthcare: A systematic review. J. Am. Med. Inform. Assoc. 2018,25, 1248–1258. [CrossRef]
43.
Lejeune, A.; Le Glaz, A.; Perron, P.-A.; Sebti, J.; Baca-Garcia, E.; Walter, M.; Lemey, C.; Berrouiguet, S. Artificial intelligence and
suicide prevention: A systematic review. Eur. Psychiatry 2022,65, 1–22. [CrossRef]
44.
Robinson, J.; Cox, G.; Bailey, E.; Hetrick, S.; Rodrigues, M.; Fisher, S.; Herrman, H. Social media and suicide prevention: A
systematic review. Early Interv. Psychiatry 2015,10, 103–121. [CrossRef]
45.
Bernert, R.A.; Hilberg, A.M.; Melia, R.; Kim, J.P.; Shah, N.H.; Abnousi, F. Artificial Intelligence and Suicide Prevention: A
Systematic Review of Machine Learning Investigations. Int. J. Environ. Res. Public Health 2020,17, 5929. [CrossRef] [PubMed]
46. Balcombe, L.; De Leo, D. The Impact of YouTube on Loneliness and Mental Health. Informatics 2023,10, 39. [CrossRef]
47.
Korngiebel, D.M.; Mooney, S.D. Considering the possibilities and pitfalls of Generative Pre-trained Transformer 3 (GPT-3) in
healthcare delivery. NPJ Digit. Med. 2021,4, 93. [CrossRef] [PubMed]
48.
Ireland, D.; Bradford, D.K. Pandora’s Bot: Insights from the Syntax and Semantics of Suicide Notes. Stud. Health Technol. Inform.
2021,276, 26. [CrossRef]
49.
Tamim, B. Belgian Woman Blames ChatGPT-Like Chatbot ELIZA for Her Husband’s Suicide. 2023. Available online: https:
//interestingengineering.com/culture/belgian-woman-blames- chatgpt-like-chatbot-eliza-for-her-husbands- suicide (accessed
on 23 August 2023).
50.
Sweeney, C.; Potts, C.; Ennis, E.; Bond, R.; Mulvenna, M.D.; O’neill, S.; Malcolm, M.; Kuosmanen, L.; Kostenius, C.; Vakaloudis,
A.; et al. Can Chatbots Help Support a Person’s Mental Health? Perceptions and Views from Mental Healthcare Professionals and
Experts. ACM Trans. Comput. Healthc. 2021,2, 1–15. [CrossRef]
51.
Evans, J. Intergenerational Report Spells Out Australia’s Future by 2063, with Warnings for Work, Climate and the Budget. 2023.
Available online: https://www.abc.net.au/news/2023-08-24/intergenerational-report-work-ageing-economy-climate-in-2063
/102769156 (accessed on 23 August 2023).
52.
Australian Commission on Safety and Quality in Health Care. National Standards in Mental Health Services. 2017. Available
online: https://www.safetyandquality.gov.au/our-work/mental-health/national-standards-in-mental-health (accessed on 24
August 2023).
53.
Balcombe, L.; De Leo, D. An Integrated Blueprint for Digital Mental Health Services Amidst COVID-19. JMIR Ment. Health
2020
,
7, e21718. [CrossRef]
54. Balcombe, L.; De Leo, D. Digital Mental Health Amid COVID-19. Encyclopedia 2021,1, 1047–1057. [CrossRef]
55.
Australian Commission on Safety and Quality in Health Care. National Safety and Quality Digital Mental Health Standards.
2020. Available online: https://www.safetyandquality.gov.au/standards/national-safety-and-quality-digital-mental- health-
standards#about-the-standards (accessed on 23 August 2023).
56.
Abrams, Z. AI Is Changing Every Aspect of Psychology. Here’s What to Watch for. 2023. Available online: https://www.apa.org/
monitor/2023/07/psychology-embracing-ai (accessed on 23 August 2023).
57.
Australian Government. Australia’s Artificial Intelligence Ethics Framework. 2019. Available online: https://www.industry.gov.
au/publications/australias-artificial-intelligence-ethics- framework (accessed on 23 August 2023).
58.
Reuters. Which Countries Are Trying to Regulate Artificial Intelligence? 2023. Available online: https://www.euronews.com/
next/2023/05/03/which-countries-are-trying-to-regulate-artificial-intelligence (accessed on 23 August 2023).
59.
European Commission. The Digital Services Act Package. 2023. Available online: https://digital- strategy.ec.europa.eu/en/
policies/digital-services-act-package (accessed on 9 October 2023).
60.
OpenAI. Frontier AI Regulation: Managing Emerging Risks to Public Safety. 2023. Available online: https://openai.com/
research/frontier-ai-regulation (accessed on 23 August 2023).
61.
Veselovsky, V.; Horta Ribeiro, M.; West, R. Artificial Artificial Artificial Intelligence: Crowd Workers Widely Use Large Language
Models for Text Production Tasks. arXiv 2023. [CrossRef]
Informatics 2023,10, 82 14 of 16
62.
Ahmed, A. Revealing the Influence of Artificial Intelligence Data on Human Contributions: Insights from Research. 2023.
Available online: https://www.digitalinformationworld.com/2023/06/revealing-influence-of-artificial.html (accessed on 23
August 2023).
63.
Muhammad, Z. AI Chatbots Might Be a Security Risk for Business Operations, Here’s Why. 2023. Available online:
https://www.digitalinformationworld.com/2023/09/ai-chatbots-might-be-security-risk-for.html#:~:text=Malicious%20
actors%20can%20wreak%20havoc,the%20conducting%20of%20illegal%20transactions (accessed on 23 August 2023).
64.
Graham, S.; Depp, C.; Lee, E.E.; Nebeker, C.; Tu, X.; Kim, H.-C.; Jeste, D.V. Artificial Intelligence for Mental Health and Mental
Illnesses: An Overview. Curr. Psychiatry Rep. 2019,21, 116. [CrossRef]
65. D’alfonso, S. AI in mental health. Curr. Opin. Psychol. 2020,36, 112–117. [CrossRef] [PubMed]
66.
van der Schyff, E.L.; Ridout, B.; Amon, K.L.; Forsyth, R.; Campbell, A.J. Providing Self-Led Mental Health Support through an
Artificial Intelligence–Powered Chat Bot (Leora) to Meet the Demand of Mental Health Care. J. Med. Internet Res.
2023
,25, e46448.
[CrossRef] [PubMed]
67.
Timmons, A.C.; Duong, J.B.; Fiallo, N.S.; Lee, T.; Vo, H.P.Q.; Ahle, M.W.; Comer, J.S.; Brewer, L.C.; Frazier, S.L.; Chaspari, T. A Call
to Action on Assessing and Mitigating Bias in Artificial Intelligence Applications for Mental Health. Perspect. Psychol. Sci.
2022
,
18, 1062–1096. [CrossRef]
68.
Joyce, D.W.; Kormilitzin, A.; Smith, K.A.; Cipriani, A. Explainable artificial intelligence for mental health through transparency
and interpretability for understandability. NPJ Digit. Med. 2023,6, 6. [CrossRef] [PubMed]
69.
Dwivedi, Y.K.; Kshetri, N.; Hughes, L.; Slade, E.L.; Jeyaraj, A.; Kar, A.K.; Baabdullah, A.M.; Koohang, A.; Raghavan, V.; Ahuja,
M.; et al. Opinion Paper: “So what if ChatGPT wrote it?” Multidisciplinary perspectives on opportunities, challenges and
implications of generative conversational AI for research, practice and policy. Int. J. Inf. Manag. 2023,71, 102642. [CrossRef]
70.
Australian Government. Australia’s AI Ethics Principles. 2023. Available online: https://www.industry.gov.au/publications/
australias-artificial-intelligence-ethics- framework/australias-ai-ethics- principles (accessed on 23 August 2023).
71.
Australian Government. Tech Trends Position Statement—Generative AI. 2023. Available online: https://www.esafety.gov.au/
industry/tech-trends-and-challenges (accessed on 23 August 2023).
72.
Trend, A. Responsible AI Is the Business for Australia. 2023. Available online: https://www.csiro.au/en/news/all/articles/2023
/july/business-potential-responsible-ai (accessed on 23 August 2023).
73.
Bello y Villarino, J.-M.; Hua, D.; Wang, B.; Trezise, M. Standardisation, Trust and Democratic Principles: The Global Race
to Regulate Artificial Intelligence. 2023. Available online: https://www.ussc.edu.au/standardisation-trust-and-democratic-
principles-the-global-race- to-regulate-artificial-intelligence (accessed on 23 August 2023).
74.
Curtis, C.; Gillespie, N.; Lockey, S. AI-deploying organizations are key to addressing ‘perfect storm’ of AI risks. AI Ethic
2022
,
3, 145–153. [CrossRef]
75.
Gillespie, N.; Lockey, S.; Curtis, C.; Pool, J.; Akbari, A. Trust in Artificial Intelligence: A Global Study; The University of Queensland
& KPMG Australia: Brisbane, Australia, 2023. [CrossRef]
76.
Australian Science Media Centre. EXPERT REACTION: Australia Considers Ban on ‘High-Risk’ Uses of AI. 2023. Available
online: https://www.scimex.org/newsfeed/expert-reaction-australia-considers-ban-on-high-risk- uses-of-ai (accessed on 29
August 2023).
77.
Morrison, R. WHO Urges Caution over Use of Generative AI in Healthcare. 2023. Available online: https://techmonitor.ai/
technology/ai-and-automation/ai-in- healthcare-who (accessed on 23 August 2023).
78.
Hasnain, A. Understanding the Biases Embedded Within Artificial Intelligence. 2023. Available online: https://www.
digitalinformationworld.com/2023/06/understanding-biases-embedded-within.html (accessed on 23 August 2023).
79.
Lee, M.; Kruger, L. Risks and Ethical Considerations of Generative AI. 2023. Available online: https://ukfinancialservicesinsights.
deloitte.com/post/102i7s2/risks-and-ethical-considerations- of-generative-ai (accessed on 23 August 2023).
80.
Pearl, R. ChatGPT’s Use in Medicine Raises Questions of Security, Privacy, Bias. 2023. Available online: https://www.forbes.com/
sites/robertpearl/2023/04/24/chatgpts-use-in-medicine-raises-questions-of-security-privacy-bias/?sh=5cc178415373 (accessed
on 23 August 2023).
81.
Karp, P. Australia Considers Ban on ‘High-Risk’ Uses of AI Such as Deepfakes and Algorithmic Bias. 2023. Available on-
line: https://www.theguardian.com/technology/2023/jun/01/australian-government-considers-ban-on-high- risk-uses-of-
ai-such-as-deepfakes- and-algorithmic-bias (accessed on 29 August 2023).
82.
Wang, Y.; Kosinski, M. Deep neural networks are more accurate than humans at detecting sexual orientation from facial images. J.
Pers. Soc. Psychol. 2018,114, 246–257. [CrossRef]
83.
Bidon, H.G. Mitigating Bias in AI. 2023. Available online: https://www.reinventedmagazine.com/post/mitigating-bias-in-ai
(accessed on 31 August 2023).
84.
Semuels, A. Millions of Americans Have Lost Jobs in the Pandemic—And Robots and AI Are Replacing Them Faster Than Ever.
2023. Available online: https://time.com/5876604/machines-jobs-coronavirus/ (accessed on 31 August 2023).
85.
Akhtar, P.; Ghouri, A.M.; Khan, H.U.R.; Haq, M.A.U.; Awan, U.; Zahoor, N.; Khan, Z.; Ashraf, A. Detecting fake news and
disinformation using artificial intelligence and machine learning to avoid supply chain disruptions. Ann. Oper. Res.
2022
,
327, 633–657. [CrossRef]
Informatics 2023,10, 82 15 of 16
86.
Santos, F.C.C. Artificial Intelligence in Automated Detection of Disinformation: A Thematic Analysis. J. Media
2023
,4, 679–687.
[CrossRef]
87.
Zhou, J.; Zhang, Y.; Luo, Q.; Parker, A.G.; De Choudhury, M. Synthetic Lies: Understanding AI-Generated Misinformation and
Evaluating Algorithmic and Human Solutions. In Proceedings of the 2023 CHI Conference on Human Factors in Computing
Systems, Hamburg, Germany, 23–29 April 2023. [CrossRef]
88.
Chin, C. Navigating the Risks of Artificial Intelligence on the Digital News Landscape. 2023. Available online: https://www.csis.
org/analysis/navigating-risks-artificial-intelligence-digital-news-landscape (accessed on 31 August 2023).
89.
Appel, G.; Neelbauer, J.; Schweidel, D.A. Generative AI Has an Intellectual Property Problem. 2023. Available online: https:
//hbr.org/2023/04/generative-ai-has-an- intellectual-property-problem (accessed on 31 August 2023).
90.
Machete, P.; Turpin, M. The Use of Critical Thinking to Identify Fake News: A Systematic Literature Review. In Proceedings of
the Responsible Design, Implementation and Use of Information and Communication Technology: 19th IFIP WG 6.11 Conference
on e-Business, e-Services, and e-Society, I3E 2020, Skukuza, South Africa, 6–8 April 2020; pp. 235–246. [CrossRef]
91.
Cusumano, M.A.; Gawer, A.; Yoffie, D.B. Social Media Companies Should Self-Regulate. Now. 2021. Available online: https:
//hbr.org/2021/01/social-media-companies-should- self-regulate-now (accessed on 31 August 2023).
92.
Susskind, J. We Can Regulate Social Media without Censorship. Here’s How. 2022. Available online: https://time.com/6199565/
regulate-social-media-platform-reduce-risks/ (accessed on 31 August 2023).
93.
McCarthy, M. Transparency Is Essential for Effective Social Media Regulation. 2022. Available online: https://www.brookings.
edu/articles/transparency-is-essential-for-effective-social-media-regulation/ (accessed on 31 August 2023).
94.
Kooli, C. Chatbots in Education and Research: A Critical Examination of Ethical Implications and Solutions. Sustainability
2023
,
15, 5614. [CrossRef]
95.
McLachlan, S.; Cooper, P. How the YouTube Algorithm Works in 2023: The Complete Guide. 2023. Available online: https:
//blog.hootsuite.com/how-the-youtube-algorithm- works/ (accessed on 31 August 2023).
96.
Balcombe, L. Collaborative AI to Shine a Light on YouTube Mental Health Rabbit Holes. 2023. Available online: https:
//www.digitalinformationworld.com/2023/06/collaborative-ai-to-shine- light-on.html (accessed on 31 August 2023).
97.
Gillespie, T. Do Not Recommend? Reduction as a Form of Content Moderation. Soc. Media + Soc.
2022
,8, 20563051221117552.
[CrossRef]
98.
Ross Arguedas, A.; Robertson, C.T.; Fletcher, R.; Neilsen, R.K. Echo Chambers, Filter Bubbles, and Polarisation: A Literature Re-
view. 2022. Available online: https://reutersinstitute.politics.ox.ac.uk/echo-chambers-filter-bubbles-and-polarisation-literature-
review (accessed on 31 August 2023).
99.
Headspace. Young People Want to Disconnect from Social Media—But FOMO Won’t Let Them. 2023. Available online:
https://headspace.org.au/our-organisation/media-releases/young-people-want-to-disconnect-from-social-media-but-fomo-
wont-let-them/ (accessed on 31 August 2023).
100.
Polanin, J.R.; Espelage, D.L.; Grotpeter, J.K.; Ingram, K.; Michaelson, L.; Spinney, E.; Valido, A.; El Sheikh, A.; Torgal, C.; Robinson,
L. A Systematic Review and Meta-analysis of Interventions to Decrease Cyberbullying Perpetration and Victimization. Prev. Sci.
2021,23, 439–454. [CrossRef]
101. Pantic, I. Online Social Networking and Mental Health. Cyberpsychol. Behav. Soc. Netw. 2014,17, 652–657. [CrossRef]
102. Braghieri, L.; Levy, R.; Makarin, A. Social Media and Mental Health. Am. Econ. Rev. 2022,112, 3660–3693. [CrossRef]
103. Zsila, Á.; Reyes, M.E.S. Pros & cons: Impacts of social media on mental health. BMC Psychol. 2023,11, 201. [CrossRef]
104.
Montag, C.; Yang, H.; Elhai, J.D. On the Psychology of TikTok Use: A First Glimpse from Empirical Findings. Front. Public Health
2021,9, 641673. [CrossRef] [PubMed]
105.
Curtin, S.; Garnett, M. Suicide and Homicide Death Rates among Youth and Young Adults Aged 10–24: United States, 2001–2021; CDC:
Atlanta, GA, USA, 2023. [CrossRef]
106.
Cadwalladr, C.; Graham-Harrison, E. Revealed: 50 Million Facebook Profiles Harvested for Cambridge Analytica in Major Data
Breach. 2018. Available online: https://www.theguardian.com/news/2018/mar/17/cambridge-analytica-facebook-influence-
us-election (accessed on 31 August 2023).
107.
Australian Government. Inquiry into Social Media and Online Safety. 2022. Available online: https://www.oaic.gov.au/
newsroom/inquiry-into-social-media-and-online-safety (accessed on 31 August 2023).
108.
Bogle, A. Australia’s Changing How It Regulates the Internet—And No-One’s Paying Attention. 2022. Available online: https:
//www.abc.net.au/news/science/2022-09-21/internet-online-safety-act-industry-codes/101456902 (accessed on 31 August
2023).
109.
Evans, J. Artificial Intelligence Technologies Could Be Classified by Risk, as Government Consults on AI Regulation. 2023.
Available online: https://www.abc.net.au/news/2023-06-01/ai-government-regulation-risk-classification-plan/102417294
(accessed on 31 August 2023).
Informatics 2023,10, 82 16 of 16
110.
Ognibene, D.; Wilkens, R.; Taibi, D.; Hernández-Leo, D.; Kruschwitz, U.; Donabauer, G.; Theophilou, E.; Lomonaco, F.; Bursic,
S.; Lobo, R.A.; et al. Challenging social media threats using collective well-being-aware recommendation algorithms and an
educational virtual companion. Front. Artif. Intell. 2023,5, 654930. [CrossRef]
111.
Hohenstein, J.; Kizilcec, R.F.; DiFranzo, D.; Aghajari, Z.; Mieczkowski, H.; Levy, K.; Naaman, M.; Hancock, J.; Jung, M.F. Artificial
intelligence in communication impacts language and social relationships. Sci. Rep. 2023,13, 5487. [CrossRef]
Disclaimer/Publisher’s Note:
The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.
... L. King et al., 2022). Certain crisis response teams have also adopted AIdriven chatbot interventions on platforms such as WhatsApp and Facebook (Balcombe, 2023;Cheng & Jiang, 2020). Through digital literacy workshops, social media fact-checking initiatives, and community-led digital advocacy, practitioners play a crucial role in connecting vulnerable populations with trustworthy health information. ...
Chapter
Full-text available
Public health crises, such as pandemics, natural disasters, and environmental emergencies, disproportionately affect vulnerable populations, exacerbating existing health disparities. Effective crisis communication requires an interdisciplinary approach that integrates public health strategies, social work interventions, and marketing-based behavioral insights to ensure accessibility, cultural competence, and engagement. This chapter explores the intersection of social work and social marketing in crisis communication, highlighting the role of social workers in advocating for equity, combating misinformation, and fostering trust within communities. Social marketing principles—such as audience segmentation, behavioral insights, and tailored messaging—are examined as tools to enhance public health outreach and promote behavior change. The study underscores the necessity of a holistic framework that combines social determinants of health, digital engagement strategies, and interdisciplinary collaboration to optimize crisis communication outcomes. The findings contribute to the development of inclusive, evidence-based public health interventions aligned with Sustainable Development Goals (SDGs), particularly in reducing inequalities and strengthening partnerships for global health resilience.
... Woebot, an automated conversational agent, has been created to provide Cognitive Behavioural Therapy (CBT). It has been deemed trustworthy by multiple researchers, particularly in its ability to alleviate depressive symptoms (Balcombe, 2023;Fitzpatrick et al., 2017;Ly et al., 2017;Nicol et al., 2022;Pham et al., 2022). Additionally, it has been proven effective for treating Substance Use Disorders (Prochaska et al., 2021a;Prochaska et al., 2021b). ...
Chapter
Full-text available
AI's integration into mental healthcare has opened new frontiers in the diagnosis, treatment, and management of mental disorders. This leap brings with it a host of ethical considerations. This chapter provides an analysis of the ethical dimensions of AI interventions in mental healthcare, examining the implications for patient autonomy, privacy, and the integrity of the therapeutic relationship. It discusses the challenges of ensuring informed consent in the context of complex AI systems, the risks of algorithmic bias, its impact on equitable care, and the importance of maintaining confidentiality in the face of expansive data requirements. The chapter 194 also explores the need for transparent AI algorithms that can be scrutinized and understood by both practitioners and patients. Furthermore, it addresses the ethical responsibility of AI developers and healthcare providers to prevent harm and ensure the beneficence of AI applications. By highlighting these ethical issues, the chapter aims to foster a responsible approach to the development and deployment of AI interventions.
... Hal ini terjadi secara sadar oleh para penggunanya. Berbagai data pribadi, seperti nama lengkap, jenis kelamin, tempat dan tanggal lahir, tingkat pendidikan, pekerjaan, hobi, status hubungan, hingga foto-foto dari masa kecil, remaja, dewasa, hingga lanjut usia, dengan mudah dibagikan dan diakses oleh publik mampu diketahui oleh semua orang (Balcombe, 2023) . Dengan berkembangya teknologi muncul isu baru yang berkaitan dengan privasi dan keamanan data, memerlukan perhatian lebih besar menjaga manfaat teknologi tanpa mengorbankan aspek penting kehidupan pribadi (L. ...
Article
Teknologi telah menjadi bagian dari kehidupan masyarakat modern, membawa dampak positif maupun negatif terhadap kesehatan mental. Seminar ini bertujuan untuk mengeksplorasi peran teknologi dan dampak media sosial terhadap kesehatan mental, serta bagaimana seminar dan pengecekan kesehatan dapat menjadi solusi untuk generasi z. Dengan menggunakan pendekatan partisipatif, seminar ini melibatkan 100 peserta yang mengikuti seminar edukasi tentang dampak teknologi khususnya penggunaan media sosial terhadap kesehatan mental. Data diperoleh melalui pendekatan sosial pra dan pasca kegiatan, serta wawancara mendalam. Hasil menunjukkan bahwa seminar meningkatkan kesadaran peserta tentang pola penggunaan media sosial yang sehat dan pentingnya deteksi dini gangguan kesehatan mental. Pengecekan kesehatan mental secara berkala juga terbukti efektif dalam mengidentifikasi risiko dan memberikan intervensi dini. Seminar ini menyimpulkan bahwa komunikasi yang efektif melalui seminar dan pengecekan kesehatan dapat memitigasi dampak negatif media sosial, mendukung kesehatan mental, dan meningkatkan kualitas hidup kalangan muda yang tidak bisa lepas dari penggunaan sosial media dalam kehidupan sehari-hari.
... USA with United Kingdom, United Kingdom with Singapore, and United Kingdom with Australia were the 3 closest partnerships in this research area. There is a pressing need for the establishment of standardized guidelines and responsible deployment of chatbots in the mental health sector on a global scale, owing to the potential risks associated with biases, data privacy concerns, and the dissemination of misinformation (34). This underscores the imperative of international cooperation in developing and implementing universally accepted standards and regulatory frameworks that prioritize the ethical and transparent use of mental health chatbots. ...
Article
Full-text available
Background The proliferation of chatbots in the digital mental health sector is gaining momentum, offering a promising solution to address the pressing shortage of mental health professionals. By providing accessible and convenient mental health services and support, chatbots are poised to become a primary technological intervention in bridging the gap between mental health needs and available resources. Objective This study undertakes a thorough bibliometric analysis and discourse on the applications of chatbots in mental health, with the objective of elucidating the underlying scientific patterns that emerge at the intersection of chatbot technology and mental health care on a global scale. Methods The bibliometric software Biblioshiny and VOSviewer were used to conduct a comprehensive analysis of 261 scientific articles published in the Web of Science Core Collection between 2015 and 2024. Publications distribution are analyzed to measure productivity on countries, institutions, and sources. Scientific collaboration networks are generated to analyze the influence as well as communications between countries and institutions. Research topics and trends are formulated by using a keyword co-occurrence network. Results Over the last decade, researches on utilization of chatbots in mental health has appeared to be increasing steadily at an annual rate of 46.19%. The United States have made significant contributions to the development and expansion of publications, accounting for 27.97% of the total research output with 2452 citation counts. England came second to the US in terms of publications and citations, and followed by Australia, China, and France. National Center for Scientific Research in France ranked first among all institutions, followed by Imperial College London and University of Zurich. The number of articles published in Journal of Medical Internet Research was exceptionally high, accounting for 12.26% of the total number of articles, and JMIR Mental Health is the most influential publication sources in terms of average citations per article. Collaboration among universities in the USA, United Kingdom, Switzerland, and Singapore demonstrated a high level. The keyword co-occurrence network highlights the prominent techniques in this multidisciplinary area and reveals 5 research topics, showing a significant overlap between clusters. High-frequency terms such as “ChatGPT”, “machine learning”, and “large language models” underscore the current state of research, highlighting the cutting-edge advancements and frontiers in this field. Conclusions This study provides an in-depth analysis of the most prominent countries, institutions, publications, collaboration status, and research topics associated with utilization of chatbots in mental health over the last decade. It offers insights to mental health professionals without an AI background and individuals interested in the development of mental health chatbots. The findings suggest that chatbots hold a significant role in promoting mental health well-being and exhibit considerable potential in demonstrating empathy, curiosity, understanding, and collaborative capabilities with users.
... These virtual interviews allow for earlier detection of potential health issues, enabling timely medical intervention (Carmona et al., 2022). Advanced chatbots also integrate with electronic health records (EHRs) to tailor interactions based on patients' medical histories, enhancing both relevance and accuracy of the information provided (Balcombe, 2023). ...
Conference Paper
Chronic illness precipitates numerous challenges and often leading to a multifaceted grieving process. Traditional grief management approaches, such as counseling, support groups, and medication, often fall short of providing immediate and personalized support. The integration of artificial intelligence (AI) in healthcare, particularly AI-driven chatbots, offers promising enhancements for grief support. However, ethical considerations must be addressed. This study examines the application of AI-driven chatbots in managing grief among individuals with chronic illness using the Value Sensitive Design (VSD) framework and two kernel theories. By exploring and integrating human values and technical features, this research proposes guidelines for designing AI-driven grief chatbots.
... We have also highlighted how ethical frameworks play a vital role in the development of these technologies responsibly, ensuring alignment with a host of values such as beneficence, nonmaleficence and respect for autonomy (Fournier-Tombs & McHardy 2023, Benosman 2024. Algorithmic bias within artificial systems has the potential to also affect the treatment of marginalized populations (Balcombe 2023, Eslami 2023. Poor use of AI chatbots in the mental health field without strict rules and regulations leads to fears of quality assistance and harm (Haque & Rubya 2023, Omarov et al. 2023. ...
Article
Full-text available
Artificial Intelligence (AI) chatbot 'ChatGPT' in the education sector has changed the learning patterns among students, staff, and lecturers. The rise of AI-driven tools is bringing forth significant ethical concerns, which are considered essential to discuss. The present study was conducted in three public sector universities in Malaysia, where we opted for the quantitative research design and approached the research participants through personal invitation and snowball sampling procedures. A total of 406 respondents were involved in this study, and the data was gathered through a survey, using structured questionnaire. The purpose of this method is to collect diverse data from a variety of participants, including students, staff, and lecturers. The findings show that it is imperative to highlight the need for robust ethical guidelines and a higher education institution framework to ensure that implementation is indispensable. Furthermore, addressing these ethical challenges can harness the potential of chatbots in academia along with ensuring ethical concerns being addressed. The study concluded that academicians had clear understanding about the ethical issues of using ChatGPT or AIs. The study also suggests recommendations regarding the ethical usage by the government of Malaysia. Abstrak Chatbot Kecerdasan Buatan (AI) 'ChatGPT' di sektor pendidikan telah mengubah pola pembelajaran di kalangan mahasiswa, staf, dan dosen. Meningkatnya alat yang digerakkan oleh AI memunculkan berbagai masalah etika yang signifikan, yang dianggap penting untuk dibahas. Studi saat ini dilakukan di tiga universitas sektor publik di Malaysia, di mana kami memilih desain penelitian kuantitatif dan mendekati partisipan penelitian melalui undangan pribadi dan prosedur pengambilan sampel bola salju. Sebanyak 406 responden terlibat dalam penelitian ini, dan data dikumpulkan melalui survei, menggunakan kuesioner terstruktur. Tujuan dari metode ini adalah untuk mengumpulkan beragam data dari berbagai partisipan, termasuk mahasiswa, staf, dan dosen. Temuan menunjukkan bahwa sangat penting untuk menyoroti perlunya pedoman etika yang kuat dan kerangka kerja lembaga pendidikan tinggi untuk memastikan bahwa implementasi sangat diperlukan. Lebih jauh, mengatasi tantangan etika ini dapat memanfaatkan potensi chatbot di dunia akademis sekaligus memastikan masalah etika ditangani. Studi ini menyimpulkan bahwa akademisi memiliki pemahaman yang jelas tentang masalah etika penggunaan ChatGPT atau AI. Studi ini juga menyarankan rekomendasi mengenai penggunaan etika oleh pemerintah Malaysia. Kata kunci: kecerdasan buatan dalam pendidikan; ChatGPT; masalah etika; lembaga pendidikan tinggi 433
Article
Background/Objectives: With advancements in Large Language Models (LLMs), counseling chatbots are becoming essential tools for delivering scalable and accessible mental health support. Traditional evaluation scales, however, fail to adequately capture the sophisticated capabilities of these systems, such as personalized interactions, empathetic responses, and memory retention. This study aims to design a robust and comprehensive evaluation scale, the Comprehensive Evaluation Scale for LLM-Powered Counseling Chatbots (CES-LCC), using the eDelphi method to address this gap. Methods: A panel of 16 experts in psychology, artificial intelligence, human-computer interaction, and digital therapeutics participated in two iterative eDelphi rounds. The process focused on refining dimensions and items based on qualitative and quantitative feedback. Initial validation, conducted after assembling the final version of the scale, involved 49 participants using the CES-LCC to evaluate an LLM-powered chatbot delivering Self-Help Plus (SH+), an Acceptance and Commitment Therapy-based intervention for stress management. Results: The final version of the CES-LCC features 27 items grouped into nine dimensions: Understanding Requests, Providing Helpful Information, Clarity and Relevance of Responses, Language Quality, Trust, Emotional Support, Guidance and Direction, Memory, and Overall Satisfaction. Initial real-world validation revealed high internal consistency (Cronbach’s alpha = 0.94), although minor adjustments are required for specific dimensions, such as Clarity and Relevance of Responses. Conclusions: The CES-LCC fills a critical gap in the evaluation of LLM-powered counseling chatbots, offering a standardized tool for assessing their multifaceted capabilities. While preliminary results are promising, further research is needed to validate the scale across diverse populations and settings.
Chapter
The integration of chatbots in mental health services necessitates a close examination of data security and privacy issues. This chapter explores these concerns in psychology and psychiatry, highlighting the unique challenges and opportunities posed by digital tools in protecting sensitive patient information. The authors analyze current vulnerabilities and threats, drawing on case studies of data breaches to emphasize the need for robust protection measures. Legal and ethical considerations, including HIPAA and GDPR, are reviewed to underscore the importance of compliance. The chapter proposes strategies for enhancing data security, such as secure communication protocols and regular audits. The goal is to provide mental health professionals, cybersecurity experts, policymakers, and researchers with practical guidelines to ensure the ethical and secure use of chatbots in mental health services, balancing innovation with patient confidentiality.
Article
Background Health professionals face significant psychological burdens including burnout, anxiety, and depression. These can negatively impact their well-being and patient care. Traditional psychological health interventions often encounter limitations such as a lack of accessibility and privacy. Artificial intelligence (AI) chatbots are being explored as potential solutions to these challenges, offering available and immediate support. Therefore, it is necessary to systematically evaluate the characteristics and effectiveness of AI chatbots designed specifically for health professionals. Objective This scoping review aims to evaluate the existing literature on the use of AI chatbots for psychological health support among health professionals. Methods Following Arksey and O’Malley’s framework, a comprehensive literature search was conducted across eight databases, covering studies published before 2024, including backward and forward citation tracking and manual searching from the included studies. Studies were screened for relevance based on inclusion and exclusion criteria, among 2465 studies retrieved, 10 studies met the criteria for review. Results Among the 10 studies, six chatbots were delivered via mobile platforms, and four via web-based platforms, all enabling one-on-one interactions. Natural language processing algorithms were used in six studies and cognitive behavioral therapy techniques were applied to psychological health in four studies. Usability was evaluated in six studies through participant feedback and engagement metrics. Improvements in anxiety, depression, and burnout were observed in four studies, although one reported an increase in depressive symptoms. Conclusions AI chatbots show potential tools to support the psychological health of health professionals by offering personalized and accessible interventions. Nonetheless, further research is required to establish standardized protocols and validate the effectiveness of these interventions. Future studies should focus on refining chatbot designs and assessing their impact on diverse health professionals.
Cover Page
Full-text available
The use of social media significantly impacts mental health. It can enhance connection, increase self-esteem, and improve a sense of belonging. But it can also lead to tremendous stress, pressure to compare oneself to others, and increased sadness and isolation. Mindful use is essential to social media consumption.
Article
Full-text available
Deaths due to suicide and homicide, often referred to collectively as violent deaths, have been a leading cause of premature death to people aged 10-24 in the United States (1-3). A previous version of this report with data through 2017 showed that suicide and homicide rates for people aged 10-24 were trending upward (4). This report updates the previous report using the most recent data from the National Vital Statistics System and presents trends from 2001 through 2021 in suicide and homicide rates for people aged 10-24 and for age groups 10-14, 15-19, and 20-24.
Preprint
Full-text available
Large language models (LLMs) are remarkable data annotators. They can be used to generate high-fidelity supervised training data, as well as survey and experimental data. With the widespread adoption of LLMs, human gold--standard annotations are key to understanding the capabilities of LLMs and the validity of their results. However, crowdsourcing, an important, inexpensive way to obtain human annotations, may itself be impacted by LLMs, as crowd workers have financial incentives to use LLMs to increase their productivity and income. To investigate this concern, we conducted a case study on the prevalence of LLM usage by crowd workers. We reran an abstract summarization task from the literature on Amazon Mechanical Turk and, through a combination of keystroke detection and synthetic text classification, estimate that 33-46% of crowd workers used LLMs when completing the task. Although generalization to other, less LLM-friendly tasks is unclear, our results call for platforms, researchers, and crowd workers to find new ways to ensure that human data remain human, perhaps using the methodology proposed here as a stepping stone. Code/data: https://github.com/epfl-dlab/GPTurk
Article
Full-text available
In November 2022, OpenAI launched ChatGPT, an AI chatbot that gained over 100 million users by February 2023 [...]
Article
Full-text available
The increasing prevalence of disinformation has led to a growing interest in leveraging artificial intelligence (AI) for detecting and combating this phenomenon. This article presents a thematic analysis of the potential benefits of automated disinformation detection from the perspective of information sciences. The analysis covers a range of approaches, including fact checking, linguistic analysis, sentiment analysis, and the utilization of human-in-the-loop systems. Furthermore, the article explores how the combination of blockchain and AI technologies can be used to automate the process of disinformation detection. Ultimately, the article aims to consider the integration of AI into journalism and emphasizes the importance of ongoing collaboration between these fields to effectively combat the spread of disinformation. The article also addresses ethical considerations related to the use of AI in journalism, including concerns about privacy, transparency, and accountability.
Article
Full-text available
Digital mental health services are becoming increasingly valuable for addressing the global public health burden of mental ill-health. There is significant demand for scalable and effective web-based mental health services. Artificial intelligence (AI) has the potential to improve mental health through the deployment of chatbots. These chatbots can provide round-the-clock support and triage individuals who are reluctant to access traditional health care due to stigma. The aim of this viewpoint paper is to consider the feasibility of AI-powered platforms to support mental well-being. The Leora model is considered a model with the potential to provide mental health support. Leora is a conversational agent that uses AI to engage in conversations with users about their mental health and provide support for minimal-to-mild symptoms of anxiety and depression. The tool is designed to be accessible, personalized, and discreet, offering strategies for promoting well-being and acting as a web-based self-care coach. Across all AI-powered mental health services, there are several challenges in the ethical development and deployment of AI in mental health treatment, including trust and transparency, bias and health inequity, and the potential for negative consequences. To ensure the effective and ethical use of AI in mental health care, researchers must carefully consider these challenges and engage with key stakeholders to provide high-quality mental health support. Validation of the Leora platform through rigorous user testing will be the next step in ensuring the model is effective.
Technical Report
Full-text available
This research examines the public’s trust and attitudes towards AI use, and expectations of AI management and governance, across 17 countries. The report provides timely, comprehensive global insights into the public’s trust and acceptance of AI systems, including: who is trusted to develop, use, and govern AI, the perceived benefits and risks of AI use, community expectations of the development, regulation, and governance of AI, and how organisations can support trust in their AI use. It also sheds light on how people feel about the use of AI at work, public understanding and awareness of AI, the key drivers of trust in AI systems, and how trust and attitudes to AI have changed over time.
Article
Full-text available
Background: Chatbots are an emerging technology that show potential for mental health care apps to enable effective and practical evidence-based therapies. As this technology is still relatively new, little is known about recently developed apps and their characteristics and effectiveness. Objective: In this study, we aimed to provide an overview of the commercially available popular mental health chatbots and how they are perceived by users. Methods: We conducted an exploratory observation of 10 apps that offer support and treatment for a variety of mental health concerns with a built-in chatbot feature and qualitatively analyzed 3621 consumer reviews from the Google Play Store and 2624 consumer reviews from the Apple App Store. Results: We found that although chatbots' personalized, humanlike interactions were positively received by users, improper responses and assumptions about the personalities of users led to a loss of interest. As chatbots are always accessible and convenient, users can become overly attached to them and prefer them over interacting with friends and family. Furthermore, a chatbot may offer crisis care whenever the user needs it because of its 24/7 availability, but even recently developed chatbots lack the understanding of properly identifying a crisis. Chatbots considered in this study fostered a judgment-free environment and helped users feel more comfortable sharing sensitive information. Conclusions: Our findings suggest that chatbots have great potential to offer social and psychological support in situations where real-world human interaction, such as connecting to friends or family members or seeking professional support, is not preferred or possible to achieve. However, there are several restrictions and limitations that these chatbots must establish according to the level of service they offer. Too much reliance on technology can pose risks, such as isolation and insufficient assistance during times of crisis. Recommendations for customization and balanced persuasion to inform the design of effective chatbots for mental health support have been outlined based on the insights of our findings.
Article
Full-text available
There are positives and negatives of using YouTube in terms of loneliness and mental health. YouTube’s streaming content is an amazing resource, however, there may be bias or errors in its recommendation algorithms. Parasocial relationships can also complicate the impact of YouTube use. Intervention may be necessary when problematic and risky content is associated with unhealthy behaviors and negative impacts on mental health. Children and adolescents are particularly vulnerable. Although YouTube might assist in connecting with peers, there are privacy, safety, and quality issues to consider. This paper is an integrative review of the positive and negative impacts of YouTube with the aim to inform the design and development of a technology-based intervention to improve mental health. The impact of YouTube use on loneliness and mental health was explored by synthesizing a purposive selection (n = 32) of the empirical and theoretical literature. Next, we explored human–computer interaction issues and proposed a concept whereby an independent-of-YouTube algorithmic recommendation system steers users toward verified positive mental health content or promotions.
Article
Full-text available
Artificial intelligence (AI) is already widely used in daily communication, but despite concerns about AI’s negative effects on society the social consequences of using it to communicate remain largely unexplored. We investigate the social consequences of one of the most pervasive AI applications, algorithmic response suggestions (“smart replies”), which are used to send billions of messages each day. Two randomized experiments provide evidence that these types of algorithmic recommender systems change how people interact with and perceive one another in both pro-social and anti-social ways. We find that using algorithmic responses changes language and social relationships. More specifically, it increases communication speed, use of positive emotional language, and conversation partners evaluate each other as closer and more cooperative. However, consistent with common assumptions about the adverse effects of AI, people are evaluated more negatively if they are suspected to be using algorithmic responses. Thus, even though AI can increase the speed of communication and improve interpersonal perceptions, the prevailing anti-social connotations of AI undermine these potential benefits if used overtly.