ArticlePDF AvailableLiterature Review

Automatically extracting social determinants of health for suicide: a narrative literature review

Authors:

Abstract and Figures

Suicide is a complex phenomenon that is often not preceded by a diagnosed mental health condition, therefore making it difficult to study and mitigate. Artificial Intelligence has increasingly been used to better understand Social Determinants of Health factors that influence suicide outcomes. In this review we find that many studies use limited SDoH information and minority groups are often underrepresented, thereby omitting important factors that could influence risk of suicide.
Content may be subject to copyright.
npj | mentalhealth research Review
https://doi.org/10.1038/s44184-024-00087-6
Automatically extracting social
determinants of health for suicide: a
narrative literature review
Check for updates
Annika M. Schoene1, Suzanne Garverich2,ImanIbrahim
2,SiaShah
2, Benjamin Irving1&
Clifford C. Dacso3,4,5
Suicide is a complex phenomenon that is often not preceded by a diagnosed mental health condition,
therefore making it difcult to study and mitigate. Articial Intelligence has increasingly been used to
better understand Social Determinants of Health factors that inuence suicide outcomes. In this review
we nd that many studies use limited SDoH information and minority groups are often
underrepresented, thereby omitting important factors that could inuence risk of suicide.
There has been an increased use of Natural Language Processing (NLP) and
Machine Learning (ML), sub-disciplines in AI, in understanding and
identifying suicidal ideation, behavior, risk and attempts in both public and
clinical contexts. Suicide is one of the leading causes of death for people aged
15-29; over 700,000 people die every year as of 20191. For every person who
has died by suicide, there are many more who attempt suicide, which
increases their risk of dying by suicide in the future2.
The onset and progression of mental health issues have been linked to a
persons social, economic, political, and physical circumstances3.These
circumstances have been summarized in a framework of social factors
referred to as Social Determinants of Health (SDoH)4where each factor is
considered a driving force behind adverse health outcomes and inequalities
(e.g.: hospitalization, increased mortality and lack of access to treatment)5.
SDoH also affects people living with mental health conditions6,where
addressing these inequalities across all stages of a persons life at an indi-
vidual, local and national level are vital to reduce the number of suicide
attempts and deaths by suicide overall7.Recentwork
6highlights that there is
a set of factors and circumstances that are unique to individuals living with
mental health concerns, calling for an expansion of traditional SDoH
categories to a new subset called Social Determinants of Mental Health
(SDoMH). The use of NLP and ML to extract SDoH in the contex t of suicide
is sparsely investigated. Work in this space has predominantly focused on
reviewing (i) AI for mental health dete ction and understanding8,9,(ii)theuse
of NLP and ML to predict suicide, suicidal ideation or attempts without
considering social factors1025, (iii) the use of NLP to extract SDoH without
focusing on specic mental health incidents26,27, where suicide is one of
many possible health outcomes or (iv) focus on suicidality in context of
specicSDoMHswithouttheuseofAI
13,2830. It is also well understood that
there are some groups that are underrepresented, which leads to greater
health disparities31, which include but are not limited to groups of racial and
ethnic minorities, underserved rural communities, sexual and gender
minorities and people with disabilities. This underrepresentation can have
great impacts on NLP and ML methods (e.g.: issues of bias and fairness).
In this literature review we analyze 94 studies at the intersection of
suicide and SDoH and aim to answer the following research questions:
What NLP and ML methods are used to extract SDoH from textual
data and what are the most common data sources?
What are the most commonly identied SDoH factors for suicide?
What socio-demographic groups are these algorithms developed from
and for?
What are the most frequently used health factors and behaviors in NLP
and ML algorithms?
Methodology
Data extraction and categories for review
For each paper in this review, we extracted metadata to identify overall
trends in the collection. Data extraction is divided into four main categories,
aiming to capture (i) general information about the NLP or ML method and
data used, (ii) SDoH and SDoMH variables, (iii) socio-demographic factors,
and (iv) physical, mental, and behavioral health factors.
General information. The year of publication, the data source for the
presented study, and what type of method has been used (e.g.: NLP or ML)
are captured.
Social determinants of (Mental) health. In this category we draw on
previous and include general SDoH categories32 and categories that dis-
proportionately affect people living with mental health disorders
(SDoMH)5,6. For this, we grouped factors into broader categories called
Social, Psychosocial and Economic, where within each category we
1Northeastern University, Institute for Experiential AI, Boston, USA. 2Northeastern University, Institute for Health Equity and Social Justice Research, Boston, USA.
3Medicine Baylor College of Medicine, Houston, USA. 4Electrical and Computer Engineering Rice University, Houston, USA. 5Knox Clinic, Rockland, Maine, USA.
e-mail: a.schoene@northeastern.edu
npj Mental Health Research | (2024) 3:51 1
1234567890():,;
1234567890():,;
capture more granular factors as outlined in Table 1. In this work, we use
SDoH as the overarching term to describe such factors, where SDoMH
factors are included and our categories were developed empirically
through the wording and descriptions used in the works we reviewed.
However, we would like to note that as we attempted to categorize each
SDoH in the context of mental health we were not able to nd clear
distinctions between some terminology/categories in the literature. For
example, many works use references to adverse life events and trauma as
mutually exclusive categories and others use it as the same category. It is
beyond the scope of this review, to make such a distinction but should be
examined by further research.
Sociodemographic factors. We capture basic demographic informa-
tion about research participants, such as age, sex, and marital status.In
addition to this, we also include factors for groups that are at greater risk
of health disparities, such as gender, sexuality, disability, race and
ethnicity31.
Health factors. In addition to the aforementioned categories, we also
label each paper for a set of health conditions, treatments, behaviors and
outcomes that were frequently mentioned as referenced in the original
paper. This includes references to Physical, neurocognitive and mental
health conditions, the use of psychiatric medications,intreatments (e.g.:
stay in hospital, outpatient, ED visits, admissions), previous attempts,
history of self-harm, aggressive/antisocial behavior, substance abuse, level
of physical activity/overall health (e.g.: BMI, weight changes) and risky
behavior. Prior research3335 has shown that such factors can signicantly
increases a personsrisk of suicide.
Search strategy and query. We retrieved 1,661 bibliographic records in
December 2023 from two scientic databases (PubMed and the
Anthology of the Association for Computational Linguistics). We use the
following search queries on two popular scientic database (PubMed and
ACL Anthology) and retrieved 1,585 and 76 papers respectively:
(Natural Language Processing OR Information Extraction OR Infor-
mation Retrieval OR Text Mining OR Machine Learning OR Deep Learning)
AND (Suicide OR Suicidality OR Suicidal) AND (Social Determinants of
Health OR behavioral determinants of health)
Due to the different layout of the ACL Anthology (https://aclanthology.
org/) we use a combination of terms to identify related literature:
Table 1 | Description of Social Determinants of Health categories
Category Sub category Examples
Social Relationships and social support Being in a romantic relationship
Level of support in relations (incl. Romantic, platonic, family and at work)
Breakdown of relationships
References to family life
Social Exclusion
Stigmatization
Religion / Culture
Discrimination based on protected characteristics
Cultural or family attitudes towards suicide and mental health
The inuence of a persons belief system on mental health (e.g.: religious views on suicide)
Level of socialization Lack of social connections
No or low social support
Feeling lonely
Social isolation / loneliness (e.g.: Becks scale of Belonging67)
Psychosocial Adverse life experiences Abortion
Bullying
Responsible of enemy death
Abuse / Trauma [adult] Sexual, physical harassment and/ or abuse
Military
Death of family or loved one / witnessing suicide
Full lista
Legal issues [adult] Incarceration
Ongoing criminal investigations
Abuse / Trauma [childhood] Parentsdivorce or losing a caregiver
Foster care
Bullying
Adverse childhood experiences Unspecied (e.g.: the reviewed literature does not give detailed examples)
Imprisonment/criminal behavior [teens] Trouble with the law during childhood/ teens
Time spent in jail
Economic Housing Insecurity Living with others
Having housing insecurity
Homelessness (in the past or present)
Build environment / Neighborhood Urban vs Rural
Access to healthcare / insurance Type of payer
Level of coverage
Education Level of education
Occupation Type of job held now or in the past (e.g.: doctor, soldier, legal professional)
Employment status Unemployment, lack of security in employment
Level of employment (e.g.: military: active duty, deployed, etc.)
Socioeconomic Income level
Financial difculties or pressures
Food Insecurity Lack of access to food, changes in food intake
ahttps://www.ptsd.va.gov/professional/assessment/documents/LEC5_Standard_Self-report.PDF.
https://doi.org/10.1038/s44184-024-00087-6 Review
npj Mental Health Research | (2024) 3:51 2
Suicide +keyword or Suicide +related keyword
Our queries were constructed to capture as many relevant SDoH fac-
tors by name and also include related terms (see Table 2for a full overview of
keywords). Similarly to26,werst surveyed existing literature for SDOH
related keywords, where we identied a total of 17 relevant keywords for our
search.
Filtering and review strategy. First, we removed duplicates and inclu-
ded all papers that have been peer-reviewed, published as a full text, in
English between 2013 and 2023. Next, we screened both title and abstract
using RobotAnalyst36 to reduce human workload. RobotAnalyst is a web-
based software system that uses both text mining and machine learning
methods to prioritize papers for their relevance based on human feedback
(Free access to RobotAnalyst can be requested to reproduce this work
here: http://www.nactem.ac.uk/robotanalyst/). For this, an iterative
classication approach is used and RobotAnalyst was retrained six times
during our screening process. We developed a set of inclusion and
exclusion criteria to screen 452 papers that were predicted to be relevant
by RobotAnalyst. Both title and abstract were screened for each paper
based on the following exclusion and inclusion criteria:
Inclusion criteria:
I. Papers that focus on the use of NLP or ML to extract SDOHs related to
suicide
II. Published work and studies that take place in English-speaking
countries, including USA, United Kingdom, Australia, and Ireland
Exclusion criteria:
I. Papers that only focus on suicide and SDOH without NLP or ML
methods
II. All retracted papers
III. Workshop proceedings, commentaries, proposals, previous literature
reviews
IV. Papers focusing on unintentional death or homicides, methods of
suicide
V. Research that proposes new ideas, policies or intervention for suicide
with or without the use of Articial Intelligence
Based on our inclusion and exclusion criteria, we retrieved 94 papers
for a full review. In Fig. 1we show the full workow of our search and lter
strategy.
Findings
General Information
We generated a Sankey diagram37 to give a comprehensive overview of how
many articles in our collection were assigned to each category. Figure 2
shows each rectangle node as a metadata category, where the node height
represents the value and each line is proportional to the value. For example,
we can see that both age and ethnicity are often considered as important
factor alongside the economic SDoH, and mental health as a health factor.
Figure 3illustratestheoverallnumberofpublicationsandtypeof
methodsusedovertime,reecting trends over the last 10 years. Here, we can
see (i) that research is increasingly focusing on SDOH factors to better
understand suicide and (ii) that Machine Learning (ML) based methodol-
ogy are used more over time. This highlights not just the rising popularity of
NLP and ML methods, but also their usefulness to understand and extract
additional information at scale . At the same time, it is important to note that
research in Information Extraction has also grown over recent with the
advancements of new AI and NLP methods.
Data types and methods. In our collection of papers, two main
approaches are typically used to predict a persons risk of suicide or
suicidal behavior. Feature-based approaches utilize information, such as a
persons demographic characteristics, frequency of treatments or other
related health behaviors as input into an algorithm, whereas NLP
approaches take advantage of language (e.g.: written online posts or
EHRs) to gain insight into how suicidal ideation is expressed to predict
risk. Within these approaches, we further distinguish between (i) tradi-
tional machine learning methods, such as linear/logistic regression,
Decision Trees, or Support Vector Machines (SVM), (ii) deep learning
methods (e.g.: articial neural networks (ANN), CNNs (Convolutional
Neural Networks) or Transformers) and (iii) unsupervised learning
methods, such as topic modeling, to discover patterns from unlabeled
data. Finally, some studies have utilized existing out-of-the box tools and
software to conduct experiments and analysis and two papers did not
disclose their full approach. In Table 3, we categorize each paper in our
collection according to the methodology used.
Table 2 | Overview of keywords used in our literature search
Keywords Related keywords
smoking Cigarette, cigarette use, tobacco, tobacco use, nicotine
Alcohol Alcohol abuse, alcohol withdrawal, alcohol consume,
alcoholism, alcohol disorder
Drug Drug use, drug abuse
Substance abuse Substance use, Substance misuse
Diet Nutrition, food
Exercise Activity, physical activity
Sexual activity Sex, sexual
Child abuse Child, children
Abuse Abusive, trauma
Healthcare Quality care, access to care, access to primary care, access
health care, insurance
Education Educational stress, studying
Income Financial condition, nancial constraint, economic condition
Work Occupation, working condition, employment,
unemployment
Social support Social isolation, social connections, social cohesion
Relationships Family, friends, friendship, relationship, mother, father,
sister, brother, sibling
Food security Food insecurity
Housing Homeless, homelessness, living condition
Fig. 1 | Overview of article selection process.
https://doi.org/10.1038/s44184-024-00087-6 Review
npj Mental Health Research | (2024) 3:51 3
Furthermore, we nd that the most commonly used data source to
extract data from are Electronic Health Records (EHR) among the reviewed
papers, closely followed by surveys and interviews (see Fig. 4). Multiple data
sources refer to work that utilizes more than one type of data in their work,
such as text data (e.g.: insurance claims) in combination with traditional
clinical data38. Other data sources include audio recordings39, newspaper
articles40, and mobile data collected via an app or smart device41.
Social Determinants of (Mental) Health and socio-demographic
factors
For each paper in our collection, we extracted SDoH and socio-
demographic information for the population studied, where in Fig. 5
shows a heatmap that indicates the number of papers with combinations of
SDoH and socio-demographic information. The diagonal indicates the total
number of papers focusing on a single variable and there are a total of 18
SDoHs and 8 socio-demographic variables shown.
For SDoH variables the majority of papers focus on different types of
abuse or trauma that has been experienced in adulthood (57.44%), followed
by socioeconomic issues (36.17%), issues in relationships, and type of
occupation held (31.91%). Very few works investigate the importance of
legal issues experienced either in childhood (3.19%)oradulthood(7.44%),
or the impact of discrimination (2.12%) and food insecurity (2.21%). For
socio-demographic information we have found that the vast majority of
papers focus on age (22.9%), race (19.3%)andethnicity(19.3%). Only very
few papers focus on other variables, such as sex (12.7%), gender (11.3%),
marital status (10.5%), sexual orientation (2.5%) and disability (1.5%).
However, the vast majority of papers focus on the intersection of multiple
socio-demographic factors. This is particularly important as previous
research has shown42 how multiple elements of a personsidentity(e.g.,
gender, race and age) can lead to compounded discrimination. Here it is
noticeable that (i) most papers only control for age, race/ethnicity, sex/
gender factors and (ii) very little research is investigating the impact of
sexual orientation and disability in relation to any of the other frequently
investigated factors. It is important to note that not all papers disclosed
socio-demographic information in their study or did not have this
Fig. 2 | Sankey diagram of all reviewed articles. A Sankey diagram37, showing the methods, socio-demographic information, SDoH, and other health factors for our selected
articles.
Fig. 3 | Number of publications per year. Number of papers published between
2014 - 2024 and categorized by methodology.
Table 3 | All reviewed papers categorized by ML and NLP
method used
Approach Method Papers
Feature based Traditional ML 41,68109
Neural Network 100,109114
Tools 115
NLP Traditional ML 39,51,116126
Neural Network 27,39,48,51,127133
Topic modeling 116,134137
Tools 40,138147
https://doi.org/10.1038/s44184-024-00087-6 Review
npj Mental Health Research | (2024) 3:51 4
information available to them. We nd similar patterns when considering
how many studies look at a combination of socio-demographic factors and
SDoH information, where there are considerable gaps in research working
on the intersection. These gaps may be due to a lack of data, where studies
collecting such data points from participants or databases are leaving out
this type of information by design. For example, for people with disabilities
there are no considerations of education level, experiencing legal issues in
adulthood, discrimination, belonging, having recorded negative lifetime
events or trauma and abuse in childhood in our set of papers. There is also the
question about how generalizable the ndings are in relation to 1) who is
represented in our healthcare system; 2) who these ndings apply to; and 3)
that by design some minority groups (e.g.: transgender people) are not
represented in these studies. In Supplementary Table 1 we categorized each
paper according to their focus on SDoH categories and in section 4 we
provide a full discussion detailing the implications of our ndings.
Social Determinants of Health and physical, mental, and beha-
vioral health factors
Based on our ndings, we can see that the most commonly researched
health factors are existing mental health conditions (84.04%), substance use
(60.63%), physical health (54.25%) and previous attempts (47.87%). How-
ever, few works consider levels of physical activity (10.63%) or aggression
(8.51%). Similar to section 3.2 we also compared SDoH information to
physical, mental and behavioral health factors, where in Fig. 6we show a
heatmap of the most frequently co-occurring factors in every paper. We nd
that existing mental health conditions and substance use are most often
considered as a variable for SDoHs related to psychosocial (e.g.: abuse/
trauma in adulthood (30.85% and 30.85%), negative lifetime experiences
(22.34% and 21.27%) and economic factors (e.g.: socioeconomic status
(34.04% and 24.46%), occupation (25.53% and22.34%)andbuildenvir-
onment/neighborhood (28.72% and 23.40%). Very few studies look at social
SDoHs, such as discrimination (2.12%) or sexual orientation (7.44%). These
results also illustrate that there is a gap in the research related to disability
status.
Discussion
The increased use of ML and NLP methods to successfully extract SDoH
related to suicide brings new challenges that require multidisciplinary
solutions. In the following section we highlight three key areas, based on the
information from the review of the literature, that need to be further
explored in future research as the algorithms and subsequent tools become
more sophisticated.
Fig. 4 | Type of data source used in each ML/NLP experiment . Type of data sources
in % .
Fig. 5 | Heatmap showing the number of papers for each socio-demographic factor.
https://doi.org/10.1038/s44184-024-00087-6 Review
npj Mental Health Research | (2024) 3:51 5
Data sources
The majority of papers in this review utilize data that is typically not
accessible to the wider research community to protect patient privacy
and adhere to HIPPA (Health Insurance Portability and Account-
ability Act) legal requirements, which means that any kind of experi-
mental results are hard to reproduce. In addition to this, many datasets
are small in comparison to the treatment population and therefore
ndings may not be generalizable to larger or more diverse
populations.
Similar concerns apply to data that is more readily available to
researchers and is often sourced from social media. In such cases, (i)
researchers usually do not have any ground truth information about
the real mental health status of a user, (ii) data is sourced from plat-
forms that have limited demographics (e.g.: Redditsuserpopulationis
70% male43) and (iii) often use a single post to assess risk of suicide24.
Previous work has called for the use of whole user timelines from social
media to predict risk44,45. However, this also raises concerns around
clinical validity of public social media data, which can be taken out of
context and the risks to a users privacy may not warrant such an
intrusive approach. In order to address concerns around ground truth
and clinical validity of such approaches, researchers often ask human
annotators to label data for levels of perceived suicide risk, however in
few cases annotators have any kind of medical, psychological, or health
science training that would enable them to make a more informed
judgment4648. However, this also increases the risk of annotation bias,
which can develop when throughout the data annotation process
systematic errors are introduced that impact the tools performance
and fairness.
Developing tools to predict suicide from either clinical or social media
can lead to a number of ethical questions, including but not limited to:
Who is responsible for the tools decision making? And what do we do
when a user receives a high suicide risk score? Therefore, we recom-
mend that future research takes an interdisciplinary approach that
incorporates perspectives from bioethics, law and policy, computer
science and health science to outline how such technologies can be
developed and deployed responsibly.
Bias and Fairness concerns
Existing disparities in health research49 also lead to inequalities in who
is represented in the data that is used to develop methods and tools.
Therefore, many vulnerable populations (e.g.: transgender people,
people who are hard of hearing to name just two) are left out of this type
of research by design, leading to an increased risk of biased and unfair
automated diagnosis, treatment and possible health outcomes for
different groups of people. Therefore, extracting SDoH information
from public and clinical records can lead to further amplication of
health disparities, biases and fairness concerns.
Recent research has proposed a variety of bias and fairness metrics to
measure and mitigate biases, however; biases are not mutually exclusive
and each method comes with its own benets and disadvantages50.This
ultimately means the person or group designing the tool decides what is
fair. Future research in this eld should carefully consider all aspects of
the machine learning lifecycle (e.g.: data collection and preparation,
algorithm choices) to reduce harm. For example, examining the dataset
prior to training a new model, rebalancing training samples, and
involving a diverse group of people in the development process to get
feedback should be essential for any new algorithm or tool
development.
Computational tasks on suicide
There have been a number of different tasks on detecting suicidal
ideation5153 or attempts14,54 often with the goal to predict risk24,55,56,
using additional information that is not related to SDoH such as
emotions or emojis44,57.Typically,suchtasksareformulatedusing
categorical labels that are meant to reect levels of risk and whilst, ML
models have been able to produce accurate predictions of suicide
ideation, attempts and death58, they are rarely grounded in clinically
Fig. 6 | A heatmap showing both SDoH and physical, mental, and behavioral health factors.
https://doi.org/10.1038/s44184-024-00087-6 Review
npj Mental Health Research | (2024) 3:51 6
theories of suicide as they are hard to implement due to their
complexity59.
Furthermore, a diagnosed mental health condition is not a necessary
precursor to dying by suicide60 nor is expressing suicidal ideation61,62,
which can be used as a form of self-regulation63. Therefore, using AI/
ML based tools to predict risk, especially from public sources, can lead
to an increased risk of discrimination and stigmatization for those
affected. This is particularly concerning given that some research
proposes to use of such algorithms to predict individual risk of suicide
and therefore making people with stigmatized conditions publicly
identiable.
At the same time, healthcare systems are under more pressure than
ever to provide adequate mental health care64 and AI/ML based tools
have often been sold as promising solutions, but often overlook real-
world challenges of such systems (e.g.: clinical applicability, general-
izability, methodological issues)20,65. Similar to66, we caution against
overenthusiasm for the use of such technologies in the real world as it is
yet to be determined whether these tools are competing, comple-
mentary, or merely duplicative21. Especially, given the current use of
and task set up for suicide related tasks using AI/ML it is vital that we
have multi-stakeholder conversations (clinicians, patient advocate
groups, developers) to establish guidelines and regulations that ensures
safeguarding of those most affected by such technologies. Subse-
quently, leading to the development of more promising tools and
technologies that could aid in preventing suicide.
Conclusion
In this work, we have reviewed and manually categorized 94 papers that
use ML or NLP to extract SDoH information on suicide, including but
not limited to suicide risk and attempt prediction and ideation
detection. We nd that ML and NLP methods are increasingly used to
extract SDoH information and the majority of current research focuses
on (i) clinical records as a data source, (ii) traditional ML approaches
(e.g.: SVMs, Regression), and (iii) a limited number of SDoH factors
(e.g.: trauma and abuse as experienced in adulthood, socio-economic
factors etc.) as they relate to demographic information (e.g.: binary
gender, age and race) and other health factors (e.g.: existing mental
health diagnosis and current or historic substance abuse) compared to
the general population and those most in need of care. In our discus-
sion we have highlighted challenges and necessary next steps for future
research, which include not using AI to predict suicide, but rather uses
it as a tool of many to aid in suicide prevention. Finally, this work is
limited in that it focuses on only English-speaking countries and
western nations, where SDoHs identied in this research may not be
applicable in different contexts. Furthermore, we have chosen to only
report broad categories of methods and dataset in order to identify
general trends and patterns.
In spite of these limitations, this review highlights the need for future
research to focus on not only on the responsible development of technol-
ogies in suicide prevention, but also more modern machine learning
approaches that incorporate existing social science and psychology research.
Data availability
No datasets were generated or analysed during the current study.
Received: 25 June 2024; Accepted: 9 September 2024;
References
1. World Health Organization. Suicide in the World: Global Health
Estimates (No. WHO/MSD/MER/19.3) (World Health Organization,
2019).
2. CDC. Suicide Data and Statistics. [online] Suicide Prevention (2024).
Available at https://www.cdc.gov/suicide/facts/data.html?CDC_
AAref_Val=https://www.cdc.gov/suicide/suicide-data-statistics.html.
3. World Health Organization. Social Determinants of Mental Health
(World Health Organization, 2014).
4. Link, B. G. & Phelan, J. Social conditions as fundamental causes of
disease. J. Health Social Behavior 8094 (1995).
5. Alegría, M., NeMoyer, A., Falgàs Bagué, I., Wang, Y. & Alvarez, K..
Social determinants of mental health: where we are and where we
need to go. Curr. Psychiatry Rep. 20, 95.
6. Jeste, D. V. & Pender, V. B. Social determinants of mental health:
recommendations for research, training, practice, and policy. JAMA
Psychiatry 79, 283284 (2022).
7. World Health Organization. Comprehensive Mental Health Action
Plan 20132030 (World Health Organization, 2021).
8. Li, H., Zhang, R., Lee, Y. C., Kraut, R. E. & Mohr, D. C. Systematic
review and meta-analysis of AI-based conversational agents for
promoting mental health and well-being. NPJ Digital Med. 6, 236
(2023).
9. Su, C., Xu, Z., Pathak, J. & Wang, F. Deep learning in mental health
outcome research: a scoping review. Transl. Psychiatry 10, 116
(2020).
10. Boggs, J. M. & Kafka, J. M. A critical review of text mining
applications for suicide research. Curr. Epidemiol. Rep. 9, 126134
(2022).
11. Yeskuatov, E., Chua, S. L. & Foo, L. K. Leveraging reddit for suicidal
ideation detection: a review of machine learning and natural
language processing techniques. Int. J. Environ. Res. Public Health
19, 10347 (2022).
12. Nordin, N., Zainol, Z., Noor, M. H. M. & Chan,L. F.. Suicidal behaviour
prediction models using machine learning techniques: a systematic
review. Artif. Intell. Med. 132, 102395 (2022).
13. Castillo-Sánchez, G. et al. Suicide risk assessment using machine
learning and social networks: a scoping review. J. Med. Syst. 44, 205
(2020).
14. Arowosegbe, A. & Oyelade, T. Application of Natural Language
Processing (NLP) in detecting and preventing suicide ideation: a
systematic review. Int. J. Environ. Res. Public Health 20, 1514 (2023).
15. Kusuma, K. et al. The performance of machine learning models in
predicting suicidal ideation, attempts, and deaths: a meta-analysis
and systematic review. J. Psychiatric Res. 155, 579588 (2022).
16. Cheng, Q. & Lui, C. S. Applying text mining methods to suicide
research. Suicide LifeThreatening Behav. 51, 137147 (2021).
17. Wulz, A. R., Law, R., Wang, J. & Wolkin, A. F. Leveraging data
science to enhance suicide prevention research: a literature review.
Inj. Prev. 28,7480 (2021).
18. Bernert, R. A. et al. Articial intelligence and suicide prevention: a
systematic review of machine learning investigations. Int. J. Environ.
Res. Public Health 17, 5929 (2020).
19. LopezCastroman, J. et al. Mining social networks to improve
suicide prevention: a scoping review. J. Neurosci. Res. 98, 616625
(2020).
20. Whiting, D. & Fazel, S. How accurate are suicide risk prediction
models? Asking the right questions for clinical practice. Evid.-based
Ment. Health 22, 125 (2019).
21. Simon, G. E. et al. Reconciling statistical and clinicianspredictions
of suicide risk. Psychiatr. Serv. 72, 555562 (2021).
22. Corke, M., Mullin, K., Angel-Scott, H., Xia, S. & Large, M. Meta-
analysis of the strength of exploratory suicide prediction models;
from clinicians to computers. BJPsych Open 7, e26 (2021).
23. Vahabzadeh, A., Sahin, N. & Kalali, A. Digital suicide prevention: can
technology become a game-changer? Innov. Clin. Neurosci. 13,16
(2016).
24. Coppersmith, G., Leary, R., Crutchley, P. & Fine, A. Natural language
processing of social media as screening for suicide risk. Biomed.
Inform. Insights 10, 1178222618792860 (2018).
25. Hopkins, D., Rickwood, D. J., Hallford, D. J. & Watsford, C.
Structured data vs. unstructured data in machine learning prediction
https://doi.org/10.1038/s44184-024-00087-6 Review
npj Mental Health Research | (2024) 3:51 7
models for suicidal behaviors: a systematic review and meta-
analysis. Front. Digital Health 4, 945006 (2022).
26. Patra, B. G. et al. Extracting social determinants of health from
electronic health records using natural language processing: a
systematic review. J. Am. Med. Inform. Assoc. 28, 27162727
(2021).
27. Mitra, A. et al. Associations between natural language
processingenriched social determinants of health and suicide
death among US veterans. JAMA Netw. Open 6, e233079e233079
(2023).
28. Nouri, E., Moradi, Y. & Moradi, G. The global prevalence of suicidal
ideation and suicide attempts among men who have sex with men: a
systematic review and meta-analysis. Eur. J. Med. Res. 28, 361
(2023).
29. Mills, P. D., Watts, B. V., Huh, T. J., Boar, S. & Kemp, J. Helping
elderly patients to avoid suicide: a review of case reports from a
National Veterans Affairs database. J. Nerv. Ment. Dis. 201,1216
(2013).
30. Wood, D. S., Wood, B. M., Watson, A., Shefeld, D. & Hauter, H.
Veteran suicide risk factors: a national sample of nonveteran and
veteran men who died by suicide. Health Soc. work 45,2330(2020).
31. National Institutes of Health. Minority health and health disparities:
denitions and parameters (2024).
32. Wilkinson, R. G. & Marmot, M. (eds). Social Determinants of Health:
the Solid Facts (World Health Organization, 2003).
33. McMahon, E. M. et al. Psychosocial and psychiatric factors
preceding death by suicide: a casecontrol psychological autopsy
study involving multiple data sources. Suicide LifeThreatening
Behav. 52, 10371047 (2022).
34. Cha, C. B. et al. Annual Research Review: Suicide among
youthepidemiology, (potential) etiology, and treatment. J. Child
Psychol. Psychiatry 59, 460482 (2018).
35. Renaud, J. et al. Suicidal ideation and behavior in youth in low-and
middle-income countries: A brief review of risk factors and
implications for prevention. Front. Psychiatry 13, 1044354 (2022).
36. Przybyła, P. et al. Prioritising references for systematic reviews with
RobotAnalyst: a user study. Res. Synth. Methods 9, 470488 (2018).
37. Riehmann, P., Haner, M. & Froehlich, B., October. Interactive
sankey diagrams. In: IEEE Symposium on Information Visualization,
2005. INFOVIS 2005 233240 (IEEE, 2005).
38. Green, C. A. et al. Identifying and classifying opioidrelated
overdoses: a validation study. Pharmacoepidemiology drug Saf. 28,
11271137 (2019).
39. Belouali, A. et al. Acoustic and language analysis of speech for
suicidal ideation among US veterans. BioData Min. 14,117 (2021).
40. Moreno, M. A., Gower, A. D., Brittain, H. & Vaillancourt, T. Applying
natural language processing to evaluate news media coverage of
bullying and cyberbullying. Prev. Sci. 20, 12741283 (2019).
41. Haines-Delmont, A. et al. Testing suicide risk prediction algorithms
using phone measurements with patients in acute mental health
settings: feasibility study. JMIR mHealth uHealth 8, e15901 (2020).
42. Crenshaw, K. W. Mapping the margins: Intersectionality, identity
politics, and violence against women of color. In The public nature of
private violence (pp. 93-118). Routledge, 2013.
43. Anderson, S. Understanding Reddit demographics in 2024 (2024).
https://www.socialchamp.io/blog/reddit-demographics/#:~:text=
Reddit%20Demographics%20Gender,men%20and%20about%
2035.1%25%20women.
44. Sawhney, R., Joshi, H., Gandhi, S. & Shah, R. A time-aware
transformer based model for suicide ideation detection on social
media. In: Proc. 2020 Conference on Empirical Methods in Natural
Language Processing (EMNLP) (76857697) (2020).
45. Sawhney, R., Joshi, H., Flek, L. & Shah, R. Phase: Learning
emotional phase-aware representations for suicide ideation
detection on social media. In: Proc. 16th Conference of the European
Chapter of the Association for Computational Linguistics: Main
Volume 24152428 (2021).
46. Rawat, B. P. S., Kovaly, S., Pigeon, W. R. & Yu, H.. Scan: suicide
attempt and ideation events dataset. In: Proc. Conference.
Association for Computational Linguistics. North American Chapter.
Meeting Vol. 2022, 1029 (NIH Public Access, 2022).
47. Liu, D. et al. Suicidal ideation cause extraction from social texts. IEEE
Access 8, 169333169351 (2020).
48. Cusick, M. et al. Using weak supervision and deep learning to
classify clinical notes for identication of current suicidal ideation. J.
Psychiatr. Res. 136,95102 (2021).
49. Fiscella, K., Franks, P., Gold, M. R. & Clancy, C. M. Inequality in
quality: addressing socioeconomic, racial, and ethnic disparities in
health care. JAMA 283, 25792584 (2000).
50. Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K. & Galstyan, A. A
survey on bias and fairness in machine learning. ACM Comput. Surv.
(CSUR) 54,135 (2021).
51. Roy, A. et al. A machine learning approach predicts future risk to
suicidal ideation from social media data. NPJ Digital Med. 3,112
(2020).
52. Tadesse, M. M., Lin, H., Xu, B. & Yang, L. Detection of suicide
ideation in social media forums using deep learning. Algorithms 13,7
(2019).
53. De Choudhury, M., Kiciman, E., Dredze, M., Coppersmith, G. &
Kumar, M.. Discovering shifts to suicidal ideation from mental health
content in social media. In: Proc. 2016 CHI Conference on Human
Factors in Computing Systems 20982110 (2016).
54. Coppersmith,G.,Ngo,K.,Leary,R.&Wood,A.Exploratory
analysis of social media prior to a suicide attempt. In Proc. 3rd
wOrkshop on Computational Linguistics and Clinical Psychology
106117 (2016).
55. Mbarek, A., Jamoussi, S., Char, A. & Hamadou, A. B. Suicidal
proles detection in Twitter. In: WEBIST 289296 (2019).
56. Zirikly, A., Resnik, P., Uzuner, O. & Hollingshead, K. CLPsych
2019 shared task: predicting the degree of suicide risk in Reddit
posts. In: Proc. 6th Workshop on Computational Linguistics and
Clinical Psychology 2433 (2019).
57. Zhang, T., Yang, K., Ji, S. & Ananiadou, S. Emotion fusion for mental
illness detection from social media: A survey. Inf. Fusion 92, 231246
(2023).
58. Schafer, K. M., Kennedy, G., Gallyer, A. & Resnik, P. A direct
comparison of theory-driven and machine learning prediction of
suicide: a meta-analysis. PLoS ONE 16, e0249833 (2021).
59. Parsapoor, M., Koudys, J. W. & Ruocco, A. C. Suicide risk detection
using articial intelligence: the promise of creating a benchmark
dataset for research on the detection of suicide risk. Front.
Psychiatry 14 (2023).
60. World Health Organization. Suicide [online] (World Health
Organization) (2023). Available at: https://www.who.int/news-room/
fact-sheets/detail/suicide.
61. Khazem, L. R. & Anestis, M. D. Thinking or doing? An examination of
well-established suicide correlates within the ideation-to-action
framework. Psychiatry Res. 245, 321326 (2016).
62. Baca-Garcia, E. et al. Estimating risk for suicide attempt: Are we
asking the right questions?: Passive suicidal ideation as a marker for
suicidal behavior. J. Affect. Disord. 134, 327332 (2011).
63. Coppersmith, D. D. et al. Suicidal thinking as affect regulation. J.
Psychopathol. Clin. Sci. 132, 385 (2023).
64. Ford, S. Mental Health Services Now under Unsustainable
Pressure. [online] Nursing Times (2022). Available at https://www.
nursingtimes.net/news/mental-health/mental-health-services-
now-under-unsustainable-pressure-02-12-2022/.
65. Franklin, J. C. et al. Risk factors for suicidal thoughts and behaviors:
a meta-analysis of 50 years of research. Psychol. Bull. 143, 187
(2017).
https://doi.org/10.1038/s44184-024-00087-6 Review
npj Mental Health Research | (2024) 3:51 8
66. Smith, W. R. et al. The ethics of risk prediction for psychosis and
suicide attempt in youth mental health. J. Pediatrics 263, 113583
(2023).
67. Beck, M. & Malley, J. A pedagogy of belonging. Reclaiming Child.
Youth 7, 133137 (1998).
68. Lynch, K. E. et al. Evaluation of suicide mortality among sexual
minority US veterans from 2000 to 2017. JAMA Netw. Open 3,
e2031357e2031357 (2020).
69. Kuroki, Y. Risk factors for suicidal behaviors among Filipino
Americans: a data mining approach. Am. J. Orthopsychiatry 85,34
(2015).
70. Gradus, J. L., King, M. W., GalatzerLevy, I. & Street, A. E. Gender
differences in machine learning models of trauma and suicidal
ideation in veterans of the Iraq and Afghanistan Wars. J. Trauma.
Stress 30, 362371 (2017).
71. Kessler, R. C. et al. Developing a practical suicide risk prediction
model for targeting highrisk patients in the Veterans health
Administration. Int. J. Methods Psychiatr. Res. 26, e1575 (2017).
72. Burke, T. A. et al. Identifying the relative importance of non-suicidal
self-injury features in classifying suicidal ideation, plans, and
behavior using exploratory data mining. Psychiatry Res. 262,
175183 (2018).
73. Walsh, C. G., Ribeiro, J. D. & Franklin, J. C. Predicting suicide
attempts in adolescents with longitudinal clinical data and machine
learning. J. Child Psychol. Psychiatry 59, 12611270 (2018).
74. Van Schaik, P., Peng, Y., Ojelabi, A. & Ling, J. Explainable statistical
learning in public health for policy development: the case of real-
world suicide data. BMC Med. Res. Methodol. 19,114 (2019).
75. Simon, G. E. et al. What health records data are required for accurate
prediction of suicidal behavior? J. Am. Med. Inform. Assoc. 26,
14581465 (2019).
76. Allan, N. P., Gros, D. F., Lancaster, C. L., Saulnier, K. G. & Stecker, T.
Heterogeneity in shortterm suicidal ideation trajectories: Predictors
of and projections to suicidal behavior. Suicide LifeThreatening
Behav. 49, 826837 (2019).
77. Tasmim, S. et al. Early-life stressful events and suicide attempt in
schizophrenia: machine learning models. Schizophrenia Res. 218,
329331 (2020).
78. Hill, R. M., Oosterhoff, B. & Do, C. Using machine learning to identify
suicide risk: a classication tree approach to prospectively identify
adolescent suicide attempters. Arch. Suicide Res. 24, 218235
(2020).
79. Haroz, E. E. et al. Reaching those at highest risk for suicide:
development of a model using machine learning methods for use
with Native American communities. Suicide LifeThreatening Behav.
50, 422436 (2020).
80. Su, C. et al. Machine learning for suicide risk prediction in children
and adolescents with electronic health records. Transl. Psychiatry
10, 413 (2020).
81. Burke, T. A., Jacobucci, R., Ammerman, B. A., Alloy, L. B. &
Diamond, G. Using machine learning to classify suicide attempt
history among youth in medical care settings. J. Affect. Disord. 268,
206214 (2020).
82. Oppenheimer, C. W. et al. Informing the study of suicidal thoughts
and behaviors in distressed young adults: The use of a machine
learning approach to identify neuroimaging, psychiatric, behavioral,
and demographic correlates. Psychiatry Res.: Neuroimaging 317,
111386 (2021).
83. Weller, O. et al. Predicting suicidal thoughts and behavior among
adolescents using the risk and protective factor framework: a large-
scale machine learning approach. Plos One 16, e0258535
(2021).
84. De La Garza, Á. G., Blanco, C., Olfson, M. & Wall, M. M. Identication
of suicide attempt risk factors in a national US survey using machine
learning. JAMA Psychiatry 78, 398406 (2021).
85. Cho, S. E., Geem, Z. W. & Na, K. S. Development of a suicide
prediction model for the elderly using health screening data. Int. J.
Environ. Res. Public Health 18, 10150 (2021).
86. Coley, R. Y., Walker, R. L., Cruz, M., Simon, G. E. & Shortreed, S. M.
Clinical risk prediction models and informative cluster size:
assessing the performance of a suicide risk prediction algorithm.
Biometrical J. 63, 13751388 (2021).
87. Lekkas, D., Klein, R. J. & Jacobson, N. C. Predicting acute suicidal
ideation on Instagram using ensemble machine learning models.
Internet Interventions 25, 100424 (2021).
88. Harman, G. et al. Prediction of suicidal ideation and attempt in 9 and
10 year-old children using transdiagnostic risk features. PLoS ONE
16, e0252114 (2021).
89. Parghi, N. et al. Assessing the predictive ability of the Suicide Crisis
Inventory for nearterm suicidal behavior using machine learning
approaches. Int. J. Methods Psychiatr. Res. 30, e1863 (2021).
90. Kim, S., Lee, H. K. & Lee, K. Detecting suicidal risk using MMPI-2
based on machine learning algorithm. Sci. Rep. 11, 15310 (2021).
91. Edgcomb, J. B., Thiruvalluru, R., Pathak, J. & Brooks, J. O. III.
Machine learning to differentiate risk of suicide attempt and self-
harm after general medical hospitalization of women with mental
illness. Med. Care 59, S58S64 (2021).
92. Dolsen, E. A. et al. Identifying correlates of suicide ideation during the
COVID-19 pandemic: a cross-sectional analysis of
148 sociodemographic and pandemic-specic factors. J. Psychiatr.
Res. 156, 186193 (2022).
93. Van Velzen, L. S. et al. Classication of suicidal thoughts and
behaviour in children: results from penalised logistic regression
analyses in the Adolescent Brain Cognitive Development study. Br.
J. Psychiatry 220, 210218 (2022).
94. Cruz, M. et al. Machine learning prediction of suicide risk does not
identify patients without traditional risk factors. J. Clin. Psychiatry 83,
42525 (2022).
95. Nemesure, M. D. et al. Predictive modeling of suicidal ideation in
patients with epilepsy. Epilepsia 63, 22692278 (2022).
96. Wilimitis, D. et al. Integration of face-to-face screening with real-time
machine learning to predict risk of suicide among adults. JAMA
Netw. Open 5, e2212095e2212095 (2022).
97. Yarborough, B. J. H. et al. Opioid-related variables did not improve
suicide risk prediction models in samples with mental health
diagnoses. J. Affect. Disord. Rep. 8, 100346 (2022).
98. Stanley, I. H. et al. Predicting suicide attempts among US Army
soldiers after leaving active duty using information available before
leaving active duty: results from the Study to Assess Risk and
Resilience in Servicemembers-Longitudinal Study (STARRS-LS).
Mol. Psychiatry 27, 16311639 (2022).
99. Horwitz, A. G. et al. Using machine learning with intensive
longitudinal data to predict depression and suicidal ideation among
medical interns over time. Psychol. Med. 53, 57785785 (2022).
100. Cheng,M. et al. Polyphenic risk score shows robust predictive ability
for long-term future suicidality. Discov. Ment. Health 2, 13 (2022).
101. Czyz, E. K. et al. Ecological momentary assessments and passive
sensing in the prediction of short-term suicidal ideation in young
adults. JAMA Netw. Open 6, e2328005e2328005 (2023).
102. Czyz, E. K., Koo, H. J., Al-Dajani, N., King, C. A. & Nahum-Shani, I.
Predicting short-term suicidal thoughts in adolescents using
machine learning: Developing decision tools to identify daily level
risk after hospitalization. Psychol. Med. 53, 29822991 (2023).
103. Edwards,A. C., Gentry, A. E., Peterson, R. E., Webb, B. T. & Mościcki,
E. K. Multifaceted risk for non-suicidal self-injury only versus suicide
attempt in a population-based cohort of adults. J. Affect. Disord.
333, 474481 (2023).
104. Wang, J. et al. Prediction of suicidal behaviors in the middle-aged
population: machine learning analyses of UK Biobank. JMIR Public
Health Surveill. 9, e43419 (2023).
https://doi.org/10.1038/s44184-024-00087-6 Review
npj Mental Health Research | (2024) 3:51 9
105. Sheu,Y. H. et al. An efcient landmark model for prediction of suicide
attempts in multiple clinical settings. Psychiatry Res. 323, 115175
(2023).
106. Coley, R. Y., Liao, Q., Simon, N. & Shortreed, S. M. Empirical
evaluation of internal validation methods for prediction in large-scale
clinical data with rare-event outcomes: a case study in suicide risk
prediction. BMC Med. Res. Methodol. 23, 33 (2023).
107. Jankowsky, K., Steger, D. & Schroeders, U.. Predicting lifetime
suicide attempts in a community sample of adolescents using
machine learning algorithms. Assessment 31, 557573 (2023).
108. Kirlic, N. et al. A machine learning analysis of risk and protective
factors of suicidal thoughts and behaviors in college students. J. Am.
Coll. Health 71, 18631872 (2023).
109. Shortreed, S. M. et al. Complex modeling with detailed temporal
predictors does not improve health records-based suicide risk
prediction. NPJ Digital Med. 6, 47 (2023).
110. DelPozo-Banos, M. et al. Using neural networks with routine health
records to identify suicide risk: feasibility study. JMIR Ment. Health 5,
e10144 (2018).
111. Sanderson, M., Bulloch, A. G., Wang, J., Williamson, T. & Patten, S.
B. Predicting death by suicide using administrative health care
system data: can feedforward neural network models improve upon
logistic regression models? J. Affect. Disord. 257, 741747 (2019).
112. Gong, J., Simon, G. E. & Liu, S. Machine learning discovery of
longitudinal patterns of depression and suicidal ideation. PLoS ONE
14, e0222665 (2019).
113. Zheng,L. et al. Development of an early-warning system for high-risk
patients for suicide attempt using deep learning and electronic
health records. Transl. Psychiatry 10, 72 (2020).
114. Choi, D. et al. Development of a machine learning model using
multiple, heterogeneous data sources to estimate weekly US suicide
fatalities. JAMA Netw. Open 3, e2030932e2030932 (2020).
115. Rozek, D. C. et al. Using machine learning to predict suicide
attempts in military personnel. Psychiatry Res. 294, 113515 (2020).
116. Homan, C. et al. Toward macro-insights for suicide prevention:
analyzing ne-grained distress at scale. In: Proc. Workshop on
Computational Linguistics and Clinical Psychology: From Linguistic
Signal to Clinical Reality (107117 (2014).
117. Zhang, Y. et al. Psychiatric stressor recognition from clinical notes to
revealassociation with suicide. Health Inform. J. 25, 18461862 (2019).
118. Carson, N. J. et al. Identication of suicidal behavior among
psychiatrically hospitalized adolescents using natural language
processing and machine learning of electronic health records. PLoS
ONE 14, e0211116 (2019).
119. Zhong, Q. Y. et al. Use of natural language processing in electronic
medical records to identify pregnant women with suicidal behavior:
towards a solution to the complex classication problem. Eur. J.
Epidemiol. 34, 153162 (2019).
120. Buckland, R. S., Hogan, J. W. and Chen, E. S.. Selection of clinical
text features for classifying suicide attempts. In: AMIA Annual
Symposium Proceedings Vol 2020, 273 (American Medical
Informatics Association, 2020).
121. Levis, M., Westgate, C. L., Gui, J., Watts, B. V. & Shiner, B. Natural
language processing of clinical mental health notes may add
predictive value to existing suicide risk models. Psychol. Med. 51,
13821391 (2021).
122. Tsui, F. R. et al. Natural language processing and machine learning
of electronic health records for prediction of rst-time suicide
attempts. JAMIA Open 4, ooab011 (2021).
123. Levis, M. et al. Leveraging unstructured electronic medical record
notes to derive population-specic suicide risk models. Psychiatry
Res. 315, 114703 (2022).
124. Rahman, N. et al. Using natural language processing to improve
suicide classication requires consideration of race. Suicide Life
Threatening Behav. 52, 782791 (2022).
125. Goldstein, E. V. et al. Characterizing female rearm suicide
circumstances: a natural language processing and machine learning
approach. Am. J. Prev. Med. 65, 278285 (2023).
126. Goldstein, E. V., Bailey, E. V. & Wilson, F. A. Poverty and suicidal
ideation among Hispanic mental health care patients leading up to
the COVID-19 Pandemic. Hispanic Health Care Int.
15404153231181110 (2023).
127. Ophir, Y., Tikochinski, R., Asterhan, C. S., Sisso, I. & Reichart, R.
Deep neural networks detect suicide risk from textual facebook
posts. Sci. Rep. 10, 16685 (2020).
128. Yao, H. et al. Detection of suicidality among opioid users on reddit:
machine learningbased approach. J. Med. Internet Res. 22, e15293
(2020).
129. Wang, S. et al. An NLP approach to identify SDoH-related
circumstance and suicide crisis from death investigation narratives.
J. Am. Med. Inform. Assoc. 30, 14081417 (2023).
130. Dobbs,M. F. et al. Linguistic correlates of suicidal ideation in youth at
clinical high-risk for psychosis. Schizophr. Res. 259,2027 (2023).
131. Lu, H. et al. Predicting suicidal and self-injurious events in a
correctional setting using AI algorithms on unstructured medical
notes and structured data. J. Psychiatr. Res. 160,1927 (2023).
132. Workman, T. E. et al. Identifying suicide documentation in clinical
notes through zeroshot learning. Health Sci. Rep. 6,e1526
(2023).
133. Purushothaman, V., Li, J. & Mackey, T. K. Detecting suicide and
self-harm discussions among opioid substance users on
instagram using machine learning. Front. Psychiatry 12, 551296
(2021).
134. Grant, R. N. et al. Automatic extraction of informal topics from online
suicidal ideation. BMC Bioinform. 19,5766 (2018).
135. Falcone, T. et al. Digital conversations about suicide among
teenagers and adults with epilepsy: A bigdata, machine learning
analysis. Epilepsia 61, 951958 (2020).
136. Kim, K. et al. Thematic analysis and natural language processing of
jobrelated problems prior to physician suicide in 20032018.
Suicide LifeThreatening Behav. 52, 10021011 (2022).
137. Levis, M., Levy, J., Dufort, V., Russ, C. J. & Shiner, B. Dynamic
suicide topic modelling: Deriving populationspecic, psychosocial
and timesensitive suicide risk variables from Electronic Health
Record psychotherapy notes. Clin. Psychol. Psychother. 30,
795810 (2023).
138. Zhong, Q. Y. et al. Screening pregnant women for suicidal behavior
in electronic medical records: diagnostic codes vs. clinical notes
processed by natural language processing. BMC Med. Inform.
Decis. Mak. 18,111 (2018).
139. Fernandes, A. C. et al. Identifying suicide ideation and suicidal
attempts in a psychiatric clinical research database using natural
language processing. Sci. Rep. 8, 7426 (2018).
140. Bittar, A., Velupillai, S., Roberts, A. & Dutta, R. Text classication to
inform suicide risk assessment in electronic health records. Stud.
Health Technol. Inform. 264,4044 (2019).
141. McCoy, T. H. Jr, Pellegrini, A. M. & Perlis, R. H. Research Domain
Criteria scores estimated through natural language processing are
associated with risk for suicide and accidental death. Depression
Anxiety 36, 392399 (2019).
142. Holden, R.et al. Investigating bullying as a predictor of suicidality in a
clinical sample of adolescents with autism spectrum disorder.
Autism Res. 13, 988997 (2020).
143. Cliffe, C. et al. Using natural language processing to extract self-
harm and suicidality data from a clinical sample of patients with
eating disorders: a retrospective cohort study. BMJ Open 11,
e053808 (2021).
144. Morrow,D. et al. A case for developing domain-specic vocabularies
for extracting suicide factors from healthcare notes. J. Psychiatr.
Res. 151, 328338 (2022).
https://doi.org/10.1038/s44184-024-00087-6 Review
npj Mental Health Research | (2024) 3:51 10
145. Xie, F., Grant, D. S. L., Chang, J., Amundsen, B. I. & Hechter, R. C.
Identifying suicidal ideation and attempt from clinical notes within a
large integrated health care system. Perm. J. 26, 85 (2022).
146. Boggs, J. M., Quintana, L. M., Powers, J. D., Hochberg, S. & Beck,A.
Frequency of cliniciansassessments for access to lethal means in
persons at risk for suicide. Arch. Suicide Res. 26, 127136 (2022).
147. Cliffe, C. et al. A multisite comparison using electronic health records
and natural language processing to identify the association between
suicidality and hospital readmission amongst patients with eating
disorders. Int. J. Eating Disorders 56, 15811592 (2023).
Acknowledgements
The study was supported by a grant from the Institute for Collaboration in
Health.
Author contributions
A.M.S., S.S., and I.I. searched the literature, and categorized each paper
according to predened categories. B.I. generated visualizations. A.M.S.
wrote theinitial draft andA.M.S., S.G., and C.D.revised the paper.All authors
reviewed the paper.
Competing interests
The authors declare no competing interests.
Additional information
Supplementary information The online version contains
supplementary material available at
https://doi.org/10.1038/s44184-024-00087-6.
Correspondence and requests for materials should be addressed to
Annika M. Schoene.
Reprints and permissions information is available at
http://www.nature.com/reprints
Publishers note Springer Nature remains neutral with regard to jurisdictional
claims in published maps and institutional afliations.
Open Access This article is licensed under a Creative Commons
Attribution-NonCommercial-NoDerivatives 4.0 International License,
which permits any non-commercial use, sharing, distribution and
reproduction in any medium or format, as long as you give appropriate
credit to the original author(s) and the source, provide a link to the Creative
Commons licence, and indicate if you modied the licensed material. You
do not have permission under this licence to share adapted material
derived from this article or parts of it. The images or other third party
material in this article are included in the articles Creative Commons
licence, unless indicated otherwise in a credit line to the material.If material
is not included in the articles Creative Commons licence and your intended
use is not permitted by statutory regulation or exceeds the permitted use,
you will need to obtain permission directly from the copyright holder. To
view a copy of this licence, visit http://creativecommons.org/licenses/by-
nc-nd/4.0/.
© The Author(s) 2024
https://doi.org/10.1038/s44184-024-00087-6 Review
npj Mental Health Research | (2024) 3:51 11
... [5] Systematic reviews conducted over the past five years demonstrated that an increasing number of studies are beginning to integrate social determinants of health (SDOH) into EHR predictive risk models. [6][7][8][9][10][11] Note that SDOH is a nomenclature used to either refer to social and community factors such as poverty level and crime rates, or to individual factors extracted from EHR such as LE. In this paper we use it for the latter. ...
Preprint
Full-text available
Background/Aims: Predictive models of suicide risk have focused on predictors extracted from structured data found in electronic health records (EHR), with limited consideration of predisposing life events (LE) expressed in unstructured clinical text such as housing instability and marital troubles. Additionally, there has been limited work in large-scale analysis of natural language processing (NLP) derived predictors for suicide risk and integration of extracted LE into longitudinal models of suicide risk. This study aims to expand upon previous research, demonstrating how high-performance computing (HPC) and machine learning technologies such as language models (LM) can be used to annotate and integrate 8 LE across all Veterans Health Administration (VHA) unstructured clinical text data with enriched performance metrics. Materials/Methods: VHA-wide clinical text from January 2000 to January 2022 were pre-processed and analyzed using HPC. Data-driven lexicon curation was performed for each LE by scaling a nearest-neighbor search over a precomputed index with LM embeddings. Data parallelism was applied to a rule-based annotator to extract LE, followed by random forest for improved positive predictive value (PPV). NLP results were analyzed and then integrated and compared to a baseline statistical model predicting risk for a combined outcome (suicide death, suicide attempt and overdose). Results: First-time LE mentions, with a PPV of 0.8 or higher, showed a temporal correlation to suicide-related events (SRE) (suicide ideation, attempt and/or death). A significant increase of LE occurrences was observed starting 2.5 months prior to an SRE. Predictive models integrating NLP-derived LE show an improved AUC of 0.81 vs. a 0.79 obtained with the baseline and novel patient identification of up to 57%. Discussion: Our analysis shows that: 1) performance metrics, specifically PPV, improved significantly from previous work and outperform related works; 2) the mentions of LE in the unstructured data increase as time to a SRE approaches; 3) LE identified from the notes in the weeks prior to a SRE were not associated with administrative bias caused by outreach; and 4) LE improved the AUC of predictive models and identified novel patients at risk for suicide. Conclusion: The resulting person-period longitudinal data demonstrated that NLP-derived LE served as acute predictors for suicide-related events. NLP integration into predictive models may help improve clinician decision support. Future work is necessary to better define these LE.
Article
Full-text available
Conversational artificial intelligence (AI), particularly AI-based conversational agents (CAs), is gaining traction in mental health care. Despite their growing usage, there is a scarcity of comprehensive evaluations of their impact on mental health and well-being. This systematic review and meta-analysis aims to fill this gap by synthesizing evidence on the effectiveness of AI-based CAs in improving mental health and factors influencing their effectiveness and user experience. Twelve databases were searched for experimental studies of AI-based CAs’ effects on mental illnesses and psychological well-being published before May 26, 2023. Out of 7834 records, 35 eligible studies were identified for systematic review, out of which 15 randomized controlled trials were included for meta-analysis. The meta-analysis revealed that AI-based CAs significantly reduce symptoms of depression (Hedge’s g 0.64 [95% CI 0.17–1.12]) and distress (Hedge’s g 0.7 [95% CI 0.18–1.22]). These effects were more pronounced in CAs that are multimodal, generative AI-based, integrated with mobile/instant messaging apps, and targeting clinical/subclinical and elderly populations. However, CA-based interventions showed no significant improvement in overall psychological well-being (Hedge’s g 0.32 [95% CI –0.13 to 0.78]). User experience with AI-based CAs was largely shaped by the quality of human-AI therapeutic relationships, content engagement, and effective communication. These findings underscore the potential of AI-based CAs in addressing mental health issues. Future research should investigate the underlying mechanisms of their effectiveness, assess long-term effects across various mental health outcomes, and evaluate the safe integration of large language models (LLMs) in mental health care.
Article
Full-text available
Background This study aimed to determine the global prevalence of suicidal ideation and attempts among men who have sex with men (MSM) as a systematic review, and meta-analysis. Methods For this meta-analysis, a search in four international databases (PubMed, Scopus, Web of Science, and EMBASE) was designed, and performed. In the next step, the information extraction checklist was prepared based on the study authors’ opinions, and the quality of the articles was evaluated using the Newcastle–Ottawa scale (NOS) checklist. Data meta-analysis was performed using STATA16 software with a significance level below 0.05. Results The results showed the prevalence of suicidal ideation, and suicide attempts among MSM was 21% (95% CI 17%-26%), and 12% (95% CI 8%-17%), respectively. The results of the subgroup analysis showed that the prevalence of suicidal ideation in the population of MSM living with Human immunodeficiency virus (HIV) was 40% (95% CI 35%–45%), and the prevalence of suicide attempts among MSM with HIV was 10% (95% CI 1%–27%). The prevalence of suicidal ideation in European MSM, and the prevalence of suicide attempts among American MSM were higher than other MSM in other geographical areas. Conclusion Considering that the prevalence of suicidal ideation and attempts among these people is many times higher than that among men in the general population, developing programs for the prevention of mental disorders with special attention to suicide is necessary for these people. Screening programs are also recommended for early diagnosis and prevention of suicide among these people.
Article
Full-text available
Background and Aims In deep learning, a major difficulty in identifying suicidality and its risk factors in clinical notes is the lack of training samples given the small number of true positive instances among the number of patients screened. This paper describes a novel methodology that identifies suicidality in clinical notes by addressing this data sparsity issue through zero‐shot learning. Our general aim was to develop a tool that leveraged zero‐shot learning to effectively identify suicidality documentation in all types of clinical notes. Methods US Veterans Affairs clinical notes served as data. The training data set label was determined using diagnostic codes of suicide attempt and self‐harm. We used a base string associated with the target label of suicidality to provide auxiliary information by narrowing the positive training cases to those containing the base string. We trained a deep neural network by mapping the training documents’ contents to a semantic space. For comparison, we trained another deep neural network using the identical training data set labels, and bag‐of‐words features. Results The zero‐shot learning model outperformed the baseline model in terms of area under the curve, sensitivity, specificity, and positive predictive value at multiple probability thresholds. In applying a 0.90 probability threshold, the methodology identified notes documenting suicidality but not associated with a relevant ICD‐10‐CM code, with 94% accuracy. Conclusion This method can effectively identify suicidality without manual annotation.
Article
Full-text available
Suicide is a leading cause of death that demands cross-disciplinary research efforts to develop and deploy suicide risk screening tools. Such tools, partly informed by influential suicide theories, can help identify individuals at the greatest risk of suicide and should be able to predict the transition from suicidal thoughts to suicide attempts. Advances in artificial intelligence have revolutionized the development of suicide screening tools and suicide risk detection systems. Thus, various types of AI systems, including text-based systems, have been proposed to identify individuals at risk of suicide. Although these systems have shown acceptable performance, most of them have not incorporated suicide theories in their design. Furthermore, directly applying suicide theories may be difficult because of the diversity and complexity of these theories. To address these challenges, we propose an approach to develop speech- and language-based suicide risk detection systems. We highlight the promise of establishing a benchmark textual and vocal dataset using a standardized speech and language assessment procedure, and research designs that distinguish between the risk factors for suicide attempt above and beyond those for suicidal ideation alone. The benchmark dataset could be used to develop trustworthy machine learning or deep learning-based suicide risk detection systems, ultimately constructing a foundation for vocal and textual-based suicide risk detection systems.
Article
Full-text available
Objectives To describe and compare the association between suicidality and subsequent readmission for patients hospitalized for eating disorder treatment, within 2 years of discharge, at two large academic medical centers in two different countries. Methods Over an 8‐year study window from January 2009 to March 2017, we identified all inpatient eating disorder admissions at Weill Cornell Medicine, New York, USA (WCM) and South London and Maudsley Foundation NHS Trust, London, UK (SLaM). To establish each patient's—suicidality profile, we applied two natural language processing (NLP) algorithms, independently developed at the two institutions, and detected suicidality in clinical notes documented in the first week of admission. We calculated the odds ratios (OR) for any subsequent readmission within 2 years postdischarge and determined whether this was to another eating disorder unit, other psychiatric unit, a general medical hospital admission or emergency room attendance. Results We identified 1126 and 420 eating disorder inpatient admissions at WCM and SLaM, respectively. In the WCM cohort, evidence of above average suicidality during the first week of admission was significantly associated with an increased risk of noneating disorder‐related psychiatric readmission (OR 3.48 95% CI = 2.03–5.99, p‐value < .001), but a similar pattern was not observed in the SLaM cohort (OR 1.34, 95% CI = 0.75–2.37, p = .32), there was no significant increase in risk of admission. In both cohorts, personality disorder increased the risk of any psychiatric readmission within 2 years. Discussion Patterns of increased risk of psychiatric readmission from above average suicidality detected via NLP during inpatient eating disorder admissions differed in our two patient cohorts. However, comorbid diagnoses such as personality disorder increased the risk of any psychiatric readmission across both cohorts. Public Significance Suicidality amongst is eating disorders is an extremely common presentation and it is important we further our understanding of identifying those most at risk. This research also provides a novel study design, comparing two NLP algorithms on electronic health record data based in the United States and United Kingdom on eating disorder inpatients. Studies researching both UK and US mental health patients are sparse therefore this study provides novel data.
Article
Full-text available
Suicide is a major global health concern and a prominent cause of death in adolescents. Previous research on suicide prediction has mainly focused on clinical or adult samples. To prevent suicides at an early stage, however, it is important to screen for risk factors in a community sample of adolescents. We compared the accuracy of logistic regressions, elastic net regressions, and gradient boosting machines in predicting suicide attempts by 17-year-olds in the Millennium Cohort Study (N = 7,347), combining a large set of self- and other-reported variables from different categories. Both machine learning algorithms outperformed logistic regressions and achieved similar balanced accuracies (.76 when using data 3 years before the self-reported lifetime suicide attempts and .85 when using data from the same measurement wave). We identified essential variables that should be considered when screening for suicidal behavior. Finally, we discuss the usefulness of complex machine learning models in suicide prediction.
Article
Importance: Advancements in technology, including mobile-based ecological momentary assessments (EMAs) and passive sensing, have immense potential to identify short-term suicide risk. However, the extent to which EMA and passive data, particularly in combination, have utility in detecting short-term risk in everyday life remains poorly understood. Objective: To examine whether and what combinations of self-reported EMA and sensor-based assessments identify next-day suicidal ideation. Design, setting, and participants: In this intensive longitudinal prognostic study, participants completed EMAs 4 times daily and wore a sensor wristband (Fitbit Charge 3) for 8 weeks. Multilevel machine learning methods, including penalized generalized estimating equations and classification and regression trees (CARTs) with repeated 5-fold cross-validation, were used to optimize prediction of next-day suicidal ideation based on time-varying features from EMAs (affective, cognitive, behavioral risk factors) and sensor data (sleep, activity, heart rate). Young adult patients who visited an emergency department with recent suicidal ideation and/or suicide attempt were recruited. Identified via electronic health record screening, eligible individuals were contacted remotely to complete enrollment procedures. Participants (aged 18 to 25 years) completed 14 708 EMA observations (64.4% adherence) and wore a sensor wristband approximately half the time (55.6% adherence). Data were collected between June 2020 and July 2021. Statistical analysis was performed from January to March 2023. Main outcomes and measures: The outcome was presence of next-day suicidal ideation. Results: Among 102 enrolled participants, 83 (81.4%) were female; 6 (5.9%) were Asian, 5 (4.9%) were Black or African American, 9 (8.8%) were more than 1 race, and 76 (74.5%) were White; mean (SD) age was 20.9 (2.1) years. The best-performing model incorporated features from EMAs and showed good predictive accuracy (mean [SE] cross-validated area under the receiver operating characteristic curve [AUC], 0.84 [0.02]), whereas the model that incorporated features from sensor data alone showed poor prediction (mean [SE] cross-validated AUC, 0.56 [0.02]). Sensor-based features did not improve prediction when combined with EMAs. Suicidal ideation-related features were the strongest predictors of next-day ideation. When suicidal ideation features were excluded, an alternative EMA model had acceptable predictive accuracy (mean [SE] cross-validated AUC, 0.76 [0.02]). Both EMA models included features at different timescales reflecting within-day, end-of-day, and time-varying cumulative effects. Conclusions and relevance: In this prognostic study, self-reported risk factors showed utility in identifying near-term suicidal thoughts. Best-performing models required self-reported information, derived from EMAs, whereas sensor-based data had negligible predictive accuracy. These results may have implications for developing decision algorithms identifying near-term suicidal thoughts to guide risk monitoring and intervention delivery in everyday life.
Article
Objective: To identify potential clinical utility of polygenic risk scores (PRS) and exposomic risk scores (ERS) for psychosis and suicide attempt in youth and assess the ethical implications of these tools. Study design: We conducted a narrative literature review of emerging findings on PRS and ERS scores for suicide and psychosis as well as a literature review on the ethics of PRS. We discuss the ethical implications of the emerging findings for the clinical potential of PRS and ERS. Results: Emerging evidence suggests that PRS and ERS may offer clinical utility in the relatively near future, but that this utility will be limited to specific, narrow clinical questions, in contrast to the suggestion that population-level screening will have sweeping impact. Combining PRS and ERS might optimize prediction. This clinical utility would change the risk-benefit balance of PRS, and further empirical assessment of proposed risks would be necessary. Some concerns for PRS, such as those about counselling, privacy, and inequities, will apply to ERS. ERS raises distinct ethical challenges as well, including some that involve informed consent and direct-to-consumer advertising. Both may face questions about the ethics of machine learning/artificial intelligence approaches. Conclusion: Predictive analytics using PRS and ERS may soon play a role in youth mental health settings. Our findings help educate clinicians about potential capabilities, limitations, and ethical implications of these tools. We suggest that a broader discussion with the public is needed to avoid overenthusiasm and determine regulations and guidelines for use of predictive scores.
Article
Introduction: Suicide rates have risen in Hispanic communities since 2015, and poverty rates among Hispanics often exceed the national average. Suicidality is a complex phenomenon. Mental illness may not alone explain whether suicidal thoughts or behaviors will occur; it remains uncertain how poverty affects suicidality among Hispanic persons with known mental health conditions. Our objective was to examine whether poverty was associated with suicidal ideation among Hispanic mental healthcare patients from 2016 to 2019. Methods: We used de-identified electronic health record (EHR) data from Holmusk, captured using the MindLinc EHR system. Our analytic sample included 4,718 Hispanic patient-year observations from 13 states. Holmusk uses deep-learning natural language processing (NLP) algorithms to quantify free-text patient assessment data and poverty for mental health patients. We conducted a pooled cross-sectional analysis and estimated logistic regression models. Results: Hispanic mental health patients who experienced poverty had 1.55 greater odds of having suicidal thoughts in a given year than patients who did not experience poverty. Conclusion: Poverty may put Hispanic patients at greater risk for suicidal thoughts even when they are already receiving treatment for psychiatric conditions. NLP appears to be a promising approach for categorizing free-text information on social circumstances affecting suicidality in clinical settings.
Article
Background: Non-suicidal self-injury and suicide attempt represent significant public health concerns. While these outcomes are related, there is prior evidence that their etiology does not entirely overlap. Efforts to directly differentiate risk across outcomes are uncommon, particularly among older, population-based cohorts. Methods: This research has been conducted using the UK Biobank. Data on individuals' self-reported history of non-suicidal self-injury only versus suicide attempt (maximum N = 6643) were analyzed. Applying LASSO and standard logistic regression, participants reporting one of these outcomes were assessed for differences across a range of sociodemographic, behavioral, and environmental features. Results: Sociodemographic features most strongly differentiated between the outcomes of non-suicidal self-injury only versus suicide attempt. Specifically, Black individuals were more likely to report a suicide attempt, as were those of mixed race, those endorsing higher levels of depressive symptoms or trauma history, and those who had experienced financial problems (odds ratios 1.02-3.92). Those more likely to engage in non-suicidal self-injury only were younger, female, had higher levels of education, those who resided with a partner, and those who had a recently injured relative. Limitations: Differences in timing across correlates and outcomes preclude the ability to establish causal pathways. Conclusions: The factors identified in the current study as differentially associated with non-suicidal self-injury only versus suicide attempt provide further evidence of at least partially distinct correlates, and warrant follow-up in independent samples to investigate causality.