ArticlePDF Available

Predicting health-related social needs in Medicaid and Medicare populations using machine learning

Springer Nature
Scientific Reports
Authors:

Abstract and Figures

Providers currently rely on universal screening to identify health-related social needs (HRSNs). Predicting HRSNs using EHR and community-level data could be more efficient and less resource intensive. Using machine learning models, we evaluated the predictive performance of HRSN status from EHR and community-level social determinants of health (SDOH) data for Medicare and Medicaid beneficiaries participating in the Accountable Health Communities Model. We hypothesized that Medicaid insurance coverage would predict HRSN status. All models significantly outperformed the baseline Medicaid hypothesis. AUCs ranged from 0.59 to 0.68. The top performance (AUC = 0.68 CI 0.66–0.70) was achieved by the “any HRSNs” outcome, which is the most useful for screening prioritization. Community-level SDOH features had lower predictive performance than EHR features. Machine learning models can be used to prioritize patients for screening. However, screening only patients identified by our current model(s) would miss many patients. Future studies are warranted to optimize prediction of HRSNs.
Content may be subject to copyright.
1
Vol.:(0123456789)
Scientic Reports | (2022) 12:4554 | https://doi.org/10.1038/s41598-022-08344-4
www.nature.com/scientificreports
Predicting health‑related social
needs in Medicaid and Medicare
populations using machine learning
Jennifer Holcomb1,2, Luis C. Oliveira3,4, Linda Higheld5,6, Kevin O. Hwang7,
Luca Giancardo8 & Elmer Victor Bernstam3,6*
Providers currently rely on universal screening to identify health‑related social needs (HRSNs).
Predicting HRSNs using EHR and community‑level data could be more ecient and less resource
intensive. Using machine learning models, we evaluated the predictive performance of HRSN status
from EHR and community‑level social determinants of health (SDOH) data for Medicare and Medicaid
beneciaries participating in the Accountable Health Communities Model. We hypothesized that
Medicaid insurance coverage would predict HRSN status. All models signicantly outperformed
the baseline Medicaid hypothesis. AUCs ranged from 0.59 to 0.68. The top performance (AUC = 0.68
CI 0.66–0.70) was achieved by the “any HRSNs” outcome, which is the most useful for screening
prioritization. Community‑level SDOH features had lower predictive performance than EHR features.
Machine learning models can be used to prioritize patients for screening. However, screening only
patients identied by our current model(s) would miss many patients. Future studies are warranted to
optimize prediction of HRSNs.
e association of social determinants of health (SDOHs) and social needs with health outcomes has been rec-
ognized internationally and in the United States. While oen used interchangeably, these are distinct concepts.
SDOHs are broader upstream social conditions in which people are born, live, and work while social needs are
more immediate and downstream individual or family needs impacted by the conditions1,2. Social needs such
as food insecurity havebeen associated with depression3, diabetes distress3, and chronic health conditions47.
Similarly, children who experience energy insecurity (i.e., inability to obtain energy to heat or cool one’s home) in
their household are at an increased odds of food insecurity, hospitalization, and poor health8. Unmet social needs
have also been associated with missed medical appointments, more frequent emergency department (ED) use
and hospital readmission9,10. ere is increasing evidence of the impact of social interventions to increase access
to preventive healthcare11, improve management of chronic conditions11, and reduce hospital admissions12,13,
reducing healthcare costs.1416.
To achieve more equitable health outcomes at lower costs17, healthcare systems should prioritize individual
patients for social interventions18. Screening patients, particularly those who are low-income and those at highest
risk for adverse health outcomes, is an important step in addressing social needs19. Current approaches to screen-
ing for social needs in U.S. healthcare settings rely on universal screening of patients. Various universal screening
approaches have been tested through the Protocol for Responding to and Assessing Patient Assets, Risks, and
Experiences (PRAPARE)20, Your Current Life Situation screening tool developed by the Kaiser Permanente Care
Management Institute21, and the Accountable Health Communities (AHC) Model screening tool developed by
OPEN
1Department of Management, Policy, and Community Health, The University of Texas Health Science Center
at Houston (UTHealth) School of Public Health, 1200 Pressler St, Houston, TX 77030, USA. 2Sinai Urban Health
Institute, 1500 South Faireld Avenue, Chicago, IL 60608, USA. 3The University of Texas Health Science Center
at Houston (UTHealth) School of Biomedical Informatics, 7000 Fannin, Houston, TX 77030, USA. 4Houston
Methodist Academic Institute, 6670 Bertner Ave, Houston, TX 77030, USA. 5Departments of Management, Policy,
and Community Health and Epidemiology, Human Genetics and Environmental Sciences, The University of Texas
Health Science Center at Houston (UTHealth) School of Public Health, 1200 Pressler St, Houston, TX 77030,
USA. 6Department of Internal Medicine, The University of Texas Health Science Center at Houston (UTHealth)
John P and Katherine G McGovern Medical School, 6410 Fannin, Houston, TX 77030, USA. 7Center for Healthcare
Quality and Safety at UTHealth/Memorial Hermann, The University of Texas Health Science Center at Houston
(UTHealth) John P and Katherine G McGovern Medical School, 6410 Fannin, Houston, TX 77030, USA. 8Center
for Precision Health, The University of Texas Health Science Center at Houston (UTHealth) School of Biomedical
Informatics, 7000 Fannin, Houston, TX 77030, USA. *email: elmer.v.bernstam@uth.tmc.edu
Content courtesy of Springer Nature, terms of use apply. Rights reserved
2
Vol:.(1234567890)
Scientic Reports | (2022) 12:4554 | https://doi.org/10.1038/s41598-022-08344-4
www.nature.com/scientificreports/
the Centers for Medicare & Medicaid Services (CMS) Innovation Center (CMMI)22,23. e PRAPARE social
needs assessment has been used frequently across healthcare settings and aligns with national data systems (e.g.,
Uniform Data System used by the Health Resources and Services Administration with Federally-Qualied Com-
munity Health Centers (FQHCs)24. A pilot approach to universal health-related social need (HRSN) screening
through CMMI’s AHC Model22,23 is currently being implemented by 28 organizations across the U.S. However,
the U.S. currently lacks standards and guidelines related to the collection of social needs screening data, par-
ticularly in healthcare settings2528. Surveying patients requires healthcare sta to build trust with patients and
for healthcare systems to increase healthcare spending to ensure dedicated healthcare sta, electronic health
record (EHR) infrastructure, other resources (e.g., funding, sta training, screening materials) and time2931.
Additionally, studies have shown that healthcare sta, including primary care physicians, do not feel condent
screening for and responding to social needs, leading to low screening rates30,32. Integration of HRSN data into
the EHR in an actionable format continues to present a challenge for healthcare providers and limits screening
as a pathway to addressing social needs33.
An alternative approach to universal screening is to utilize patient risk scores or risk prediction models
to identify and prioritize patients who are most likely to have HRSNs. Risk scores are already widely used in
healthcare settings to predict a range of outcomes from specic disease conditions (e.g., cardiovascular disease)
to hospital readmissions, healthcare cost, and ED utilization3438. Recently, there has been increasing interest in
using SDOH and social needs data to improve risk prediction models. Risk prediction eorts linking community-
level geocoded data with EHRs and other patient-level data sources (e.g., claims/administrative data) are nascent
and to date have primarily focused on predicting healthcare utilization, such as hospital readmission and ED
visits34,3941. Studies have been limited by lack of data on individuallevel social needs and in most cases limiting
to a single healthcare provider or system34,41,42. To date, few studies attempted to predict individual patient social
needs43. ese studies attempted to predict social service referrals rather than whether the patient reported a
social need.
An opportunity exists to better understand the potential for integrating risk prediction to proactively identify
patients in need of further social need assessment or social intervention outside of the healthcare model. Predict-
ing HRSN status is a novel application of predictive models and highly relevant and actionable as screening is the
rst step in the social intervention pathway. Risk prediction could also help address the structural and logistical
barriers to universal HRSN screening implementation that have been identied in recent research, including the
low level of uptake by providers, lack of time, EHR integration, availability of trained or skilled sta to conduct
screening (and intervention), patient preference, and increased costs, which are oen not reimbursed2226. To our
knowledge, there are not currently studies available comparing universal versus targeted screening approaches
for HRSN. However, research in other health domains such as HIV indicates that targeted screening can be
benecial when implementation barriers such as those noted above are present44. e objective of this study
was to predict HRSNs of patients in the CMMI AHC Model from patient-level EHR data and publicly available
community-level SDOH data. We evaluated the predictive performance versus a baseline method using Medicaid
status to assume existence of HRSNs. Our hypothesis was that using a combined dataset would outperform any
single data source alone. We further hypothesized that patients insured by Medicaid would be likely to have a
HRSN and that the combined dataset would more accurately predict social needs status (e.g., positive or nega-
tive) than the Medicaid assumption.
Methods
Study design. Patient-level HRSN screening data were collected from September 2018 through Decem-
ber 2020 in the Greater Houston area, Texas in a cross-sectional study design. e AHC Model implementa-
tion in the Greater Houston area is a part of a national randomized controlled trial funded by CMMI to test a
systematic approach to HRSN screening, community resource referral, and community resource navigation of
CMS beneciaries22,23,45. Any community-dwelling CMS beneciaries accessing care across 13 clinical delivery
sites including Emergency Departments (ED), Labor and Delivery Departments and ambulatory clinics in three
large health systems were eligible to be screened. e three health systems included a nonprot, private hospital
system (Health System A), a network of outpatient clinics at an academic health university (Health System B),
and a safety net hospital (Health System C). Patient EHR data and community-level SDOH data were combined
to predict the HRSN status of those patients in the AHC Model. Eligible patients for analysis were those with a
completed screening survey in the AHC Model, EHR data from two years prior to HRSN screening date, and
an address for geocoding to facilitate linkage of community-level SDOH data. is study has been approved by
the Committee for the Protection of Human Subjects (CPHS, the UTHSCH Institutional Review Board) under
protocol HSCSBMI130549. All methods were performed in accordance with the relevant guidelines and regu-
lations. Informed consent for this study was waived by the CPHS as part of protocol HSC-SBMI-13-0549.
Data features. Individual patient-level EHR data included demographics, diagnosis codes (ICD-10), pro-
cedure codes (CPT codes for ambulatory and hospital), and insurance type (Medicare, Medicaid or dually cov-
ered). For ICD-10 codes, only part of the code describing the disease category was used in order to create
clinically relevant "clusters". For community-level SDOH features, we reviewed the existing literature to identify
potential SDOH factors associated with known health outcomes, healthcare utilization, and HRSNs13,19,40,4650.
Community-level SDOH data at the Texas state Census Tract level were derived from the 5-year (2015 to 2019)
estimates from U.S. Census Bureaus American Community Survey (ACS) and the Centers for Disease Control
and Preventions Social Vulnerability Index (SVI) (2018). e 12 SDOH features include median household
income, poverty level, educational attainment, unemployment, health insurance coverage including uninsured
and public insurance, car ownership, home ownership, Supplemental Nutrition Assistance Program (SNAP)
Content courtesy of Springer Nature, terms of use apply. Rights reserved
3
Vol.:(0123456789)
Scientic Reports | (2022) 12:4554 | https://doi.org/10.1038/s41598-022-08344-4
www.nature.com/scientificreports/
benets, overcrowding, and disability from the ACS and from the SVI, minority status and language. A descrip-
tion of the EHR and Census data can be found in the Supplementary Material.
As part of the HRSN survey patients were asked about four indicators of social needs, and for this study, we
developed models predicting the need for each of these indicators based on a patient’s EHR and associated ACS
and SVI Census data23. We used the following indicators based on 4 of the 5 core social needs screened for in
the AHC Model23: (1) Core need: housing situation—Identies whether respondent has HRSN related to hous-
ing stability and/or housing quality. (2) Core need: food insecurity—Identies whether respondent has HRSN
related to purchasing food. (3) Core need: transportation—Identies whether respondent has HRSN related to
accessing reliable transportation. (4) Core need: utilities—Identies whether respondent has HRSN related to
diculty paying utility bills. (5) Any core need—is indicator is true if at least one of the four core needs is
true. (6) All core needs—is indicator is true if all of the four core needs are true. In addition to these metrics,
we used the Medicaid ID on the survey to indicate whether the respondent was a Medicaid beneciary. is
metric was used as a baseline for testing the predictive model, under the assumption that respondents who are
Medicaid beneciaries would have HRSNs.
Data linkage. Figure1 illustrates how the data sources were combined to create the combined dataset.
We used a table specially created in the Master Patient Index database (MPI) to map patient IDs from the
HRSN survey to the corresponding patient ID in the EHRs51. e Match Analysis Methodology in the MPI
uses key information from the HRSN surveys like survey patient ID, rst name, last name, middle name, date
of birth, sex, address (city, state, zip) Medicare Beneciary Identier (Medicare), Medicare eective date, and
Medicaid eective date to link the records to the EHR data. We used the address provided in the survey and the
EHR to geocode each patient’s address and then determined the corresponding Census tract for the address. At
each stage of matching, exclusion criteria were applied. HRSN surveys without corresponding EHR data for the
patient were excluded (n = 2418). Any records whose geocode did not match between the HRSN survey data
and the EHR data as were excluded (n = 2814). ese records had a greater than 1-km dierence between the
address provided in the HRSN survey and in the EHR. e corresponding Census Tract was then used to match
the SDOH information from Census data. Any records matched with Census Tracts located outside of the state
of Texas were excluded because they could not be matched with SDOH information (n = 40). A Consort ow
diagram52 was used to depict sample size at each step in Fig.2.
Data analysis. First, we randomly allocated the samples into three datasets: 20% of the samples (n = 1960)
were allocated for the test set (not used during the training process); 80% of the remaining samples (n = 6272)
were allocated for the training set (64% of the entire data set), with the remaining samples (n = 1568) allocated
HRSN Outcomes
Community level
SDoH
Demographics
Procedures
Diagnoses
Survey id
Address
Census tract
Paent id
AHC Survey
Electronic Health
Record
(EHR)
enterprise
Master Person Index
(eMPI)
Census data
Geoloca
on
Figure1. Data sources and linkage for modeling. Flow chart showing the data sources combined to create the
dataset displayed. e data sources are displayed as three cylinders displaying the data linking between sources.
e measures from each source are displayed as a rectangle linking to other cylinders. Patient-level HRSNs were
collected in the AHC screening survey. Using survey demographic data, patients were mapped using a Master
Patient Index database (MPI) to patient ID, demographics, diagnosis, and procedures in the EHR. Patients
addresses provided in the survey and the EHR were geocoded to each patients address and corresponding
Census tract. e geocoding is displayed as a diamond connected to the HRSN survey data measures.
Content courtesy of Springer Nature, terms of use apply. Rights reserved
4
Vol:.(1234567890)
Scientic Reports | (2022) 12:4554 | https://doi.org/10.1038/s41598-022-08344-4
www.nature.com/scientificreports/
for the validation set (16% of the entire data set. ese datasets were stratied such that each group contained
approximately the same percentage of samples of each target class as the complete set. e test set was reserved
for testing and was neither used for model training, nor for manually or automatically evaluating feature selec-
tion or any of the other model parameters. A Gradient Boosting Decision Tree Machine Learning algorithm
(LightGBM) was used to predict HRSN status using the individual and combined data sets53. ese types of
algorithms oer some degree of interpretability and work particularly well in machine learning problems with a
high dimensionality, large number of features, and large sparsity of data, which is one of the main hurdles when
dealing with EHR data. LightGBM is inherently able to handle missing data which allows us to avoid any type
of articial data imputation. e LightGBM model hyperparameters were tuned using a Bayesian optimiza-
tion, which allowed for an unbiased search of the best performing model without direct trial and error which
could lead to overtting54. Specically, we used the scikit-optimize library55 for a crossed validated Bayesian
search (implemented in the BayesianSearchCV scikit-optimize class) on the training set. e best combination
of hyperparameters was selected by maximizing the accuracy on the validation set. e test set remained com-
pletely independent from the hyperparameter search, thereby avoiding any risk of overtting and data leakage.
For a full description of LightGBM and the Bayesian hyperparameter search we refer the readers to the papers
referenced5355.
Area Under the Receiving Operating Characteristic Curve (AUC)56 and comparison to a baseline decision
using Medicaid status were used to evaluate model performance on the test set. Analysis was performed using the
Python scikit-learn and lightGBM libraries. P-values were also computed with a non-parametric Mann–Whitney
U test, under the null hypothesis that, for each HRSN, the distribution of the ordinal real value output of the
models is equal when HRSN = False or HRSN = True.
Results
Table1 summarizes the demographic and HRSN characteristics of the patients included in the nal modeling.
Patients were primarily female (52.7%), Black or African American (40.6%), single marital status (59.3%), cov-
ered by Medicaid (85.4%), and screened at Health System A, a nonprot, private hospital system (83.8%). Over
half of patients (57%) screened positive for at least one HRSN. Food insecurity was the highest frequency need,
reported at 39%. Housing, transportation and utility needs were reported with similar frequencies (26–29%).
In Fig.3, we compare and contrast the predictive performance of the ML model trained with the set of features
(EHR, Census, or EHR + Census) and the baseline Medicaid status to determine a HRSN. All models trained
with EHR and Census features signicantly outperformed the baseline Medicaid insurance status to determine
presence of a HRSN as shown by the 95% condence intervals (CI) of the Receiver Operating Characteristic
(ROC) curves (shaded areas) when compared to the baseline Medicaid decision (shown as a red cross). When
all features were used, AUCs ranged from 0.59 to 0.68. e top performance (AUC = 0.68, CI 0.66–0.70) was
achieved by the “any HRSNs” outcome, which is the most useful for patient HRSN screening prioritization. In the
majority of experiments, models trained with community-level SDOH features had lower predictive performance
Health System A
surveys (n=12,961)
Health System C
surveys (n=1,469)
Health System B
surveys (n=642)
Total paents
(n=15,072)
Paents with EHR
data (n=12,654)
Excluded (n=2,418)
No EHR data
Paents with
accurate locaon
(n=9,840)
Excluded (n=2,814)
Locaon not aligned (EHR
and survey)
Paents with census
data (n=9,800)
Excluded (n=40)
Missing census data
Training dataset
(n=6,272)
Test dataset
(n=1,960)
Validaon dataset
(n=1,568)
Figure2. Consort ow diagram. Consort ow diagram depicting the sample size at each step of the data
linkage. e diagram moves downward with each step displayed as a rectangle. Patients were excluded from
the nal datasets if they had no EHR data, if there was not alignment with the EHR and survey addresses, and
if their geocoded location was missing corresponding Census data. ese exclusions are depicted as rectangles
with arrows along the diagram indicating where a patient sample was excluded. From these exclusions, the
bottom and nal three rectangles depict the training, validation, and test datasets included in the data analysis.
Content courtesy of Springer Nature, terms of use apply. Rights reserved
5
Vol.:(0123456789)
Scientic Reports | (2022) 12:4554 | https://doi.org/10.1038/s41598-022-08344-4
www.nature.com/scientificreports/
than EHR features alone. e only exception was the “Diculty Paying Utilities” HRSN, where the main drivers
for predictive performance were Census features. In order to aid the reproducibility of our ndings, all model
hyperparameters automatically identied by the Bayesian search are shown in the Supplementary Material.
Discussion
We found that the addition of readily available SDOH data at the community-level did not improve performance
over data typically available in the EHR for predicting patient social needs status. Of the models, “any HRSN”
had the best predictive power at 0.68. Our AUC values were slightly lower than some previous studies43, but it
is important to note that our outcome (whether the patient reported a HRSN) is dierent than other currently
published studies (referral to a social service), limiting our ability to compare. Use of patient Medicaid insur-
ance status signicantly under-predicted social needs status, indicating that use of Medicaid insurance coverage
alone is not predictive and we caution providers against using this factor to determine need or who should be
screened. Previous studies have shown that patients of lower income status have high rates of social needs, poorer
self-rated health, and higher rates of chronic conditions19, particularly those with Medicaid and those dually
covered by Medicare and Medicaid5759. Given that Medicaid in Texas covers low-income populations and our
Table 1. Demographic Characteristics and Health-Related Social Needs (HRSNs) of CMS Beneciaries in the
Accountable Health Communities (AHC) Model in the Greater Houston Area, September 2018 to December
2020. a Includes "Unknown", "Declined", "Not Answered” responses and records that had no response. b Patients
could be in multiple categories so numbers do not sum to total.
Characteristics
Sample (n = 9800)
Patients, No. (%)
Age, mean (SD), years 35.5 (26.3)
Race
Black or African American 3978 (40.6)
Other 2857 (29.2)
White 1763 (18.0)
Latin American 612 (6.2)
Hispanic or Latino 337 (3.4)
American Indian or Alaska Native 23 (0.2)
Asian or Pacic Islander 15 (0.2)
Unknowna215 (2.2)
Marital status
Single 5809 (59.3)
Married 1411 (14.4)
Widowed 368 (3.8)
Divorced 365 (3.7)
Separated 67 (0.7)
Life Partner 6 (0.1)
Legally Separated 6 (0.1)
Unknown 1768 (18.0)
Sex
Female 5162 (52.7)
Male 3049 (31.1)
Unknown 1589 (16.2)
Insurance typeb
Medicaid 8370 (85.4)
Medicare 2231 (22.8)
Health Related Social Needs (HRSNs)b
Housing instability and/or quality 2876 (29.3)
Food insecurity 3780 (38.6)
Transportation 2722 (27.8)
Diculty paying utility bills 2582 (26.3)
Any core need 5588 (57.0)
All core needs 813 (8.3)
Health System
Health System A 8211 (83.8)
Health System B 1108 (11.3)
Health System C 481 (4.9)
Content courtesy of Springer Nature, terms of use apply. Rights reserved
6
Vol:.(1234567890)
Scientic Reports | (2022) 12:4554 | https://doi.org/10.1038/s41598-022-08344-4
www.nature.com/scientificreports/
sample of dually covered beneciaries was too low to allow splitting between test, training and validation datasets
(< 5% of the overall sample), we felt using Medicaid status represented a reasonable baseline hypothesis to apply.
What our ndings indicate is that the relationship of social needs to insurance status may be more sensitive than
previous literature has been able to detect without screening tools. e AHC screening tool recently underwent
psychometric testing and was found to have concurrent validity and be sensitive for detecting social needs across
a wide swath of patients60. When compared to other tools, AHC was more sensitive to detecting certain social
needs including housing instability.
Our study adds to the growing literature on the application of machine learning and integration of commu-
nity SDOH data for use in healthcare settings39. Studies to date have found mixed results when adding SDOH
data, with some reporting minimal to no improvements in model prediction43. Similar to these studies, we
found that the addition of SDOH data led to very little improvements to model performance for HRSN status
with the exception of diculty paying utility bills. As other authors have identied, there may be a number
of reasons why community-level SDOH are not good predictors of HRSNs. First, while our study included a
large population of patients, their geographic locations were clustered into a small number of Census Tracts.
e similarity of demographic and SDOH factors resulting from relatively few Census Tracts may have limited
discriminatory power43. As noted in previous studies40, the SDOH factors could be correlated with the patient-
level demographics and health status in existing EHR data, therefore, adding limited predictive power in the
models from community-level SDOH.
Additionally, we only examined four of the domains of HRSNs, thus it is possible that the SDOH and EHR
data might have predictive power in other social need domains such as nancial health, social isolation, com-
munity safety, and health literacy6163. While the SDOH data from community sources andHRSN screening data
measure dierent constructs, dierent levels of the associated constructs or dierent time periods for associated
constructs, might impact discriminatory power. For example, diculty paying utility bills conceptually aligns
with SDOH community-level socioeconomic status, particularly income and poverty. e AHC Model survey
question asks about diculty paying bills for the previous 12months. Whereas, questions from the survey for
housing ask about today. Future studies using survey design methods to consider variation in constructs, levels of
measurement, and impact of time period assessed are warranted. Lastly, a larger, more geographically and socially
diverse sample could be valuable to future modeling eorts to determine if SDOH are truly predictive or not for
HRSN status. It is also possible that the high rate of social needs observed in our population limited discrimi-
natory power. is is similar to previously published studies showing high rates of HRSN in ED populations34.
Figure3. Machine learning model predictive value by HRSN status.
Content courtesy of Springer Nature, terms of use apply. Rights reserved
7
Vol.:(0123456789)
Scientic Reports | (2022) 12:4554 | https://doi.org/10.1038/s41598-022-08344-4
www.nature.com/scientificreports/
Our study oers a number of strengths to the existing literature on the application of machine learning models
for predicting HRSNs. To our knowledge this is the rst study to directly model HRSN status using publicly
available data, EHR data, and individual level HRSN screening data. We utilized data from three large health
systems representing patients from the largest medical center in the world. e EHR included both ambulatory
and inpatient visit data in addition to patient demographics. We used individual level data on a large number of
patients who were screened for HRSN through a universal oer to screen. We used readily available EHR and
public SDOH data to model HRSN status making our approach easily replicable by other researchers and health
systems. We also compared our ndings with Medicaid insurance status as a baseline assumption and potential
proxy for HRSN status. Previous studies have found strong associations between Medicaid coverage, social
needs, and healthcare utilization and outcomes64. Finally, we used state of the art Gradient Boosting Decision
Tree ML approaches whose hyperparameters where automatically tuned with Bayesian optimization without
using a non-overlapping test set. is allowed for a fully unbiased ne tuning of the algorithm to each HRSN
without direct trial and error which could lead to overtting.
Using community-level SDOH data to predict individual HRSN status collected via screening is prone to
limitations and potential biases. First is the risk of ecologic fallacy, where assumptions made about individuals
using aggregate-level (area) data yield incorrect results65. Despite the value of using such data to predict HRSN
status, our study adds to previously published ndings that the ecological fallacy may be a limitation to the
utility of such eorts.
Second, a potential limitation is the use of individual level HRSN screening data collected via self-report.
Self-reported data are prone to bias. Characteristics may dier from those who agreed to answer the HRSNs
questions versus those who declined, though our high survey completion rate (~ 45%) lessens this likelihood.
Patients might also underreport HRSNs because of social stigma, social desirability bias, or lack of perceived
benet of reporting needs (i.e., access to navigation services66).
Lastly, additional limitations relate to the selected SDOH data used in our study. We selected community-
level SDOH factors based on previously published studies. However, there is a vast diversity of secondary data
available and it is possible that there is an unmeasured and un-modeled SDOH factor, which could improve
predictive performance40.
ere are implications from this study for healthcare providers and institutions. First, targeting those patients
with the highest social and health needs could help improve patient health and healthcare utilization. A previous
study has shown that social interventions targeting high-utilizing patient populations decreased overall health-
care utilization with more signicant eects seen in low–socioeconomic status patients67. In terms of hospital
utilization, a dose–response relationship has been reported between HRSNs and hospital readmission68. is
further highlights the need to understand and intervene on high-utilizing populations with social needs.
Second, there is a need to identify how to best screen patients for social needs while reducing clinic burden
across healthcare setting types. Machine learning methods can help prioritize patients for HRSN screening
while reducing clinic burden39. Predicting HRSNs could reduce the need for additional data collection, EHR
infrastructure, sta time, and training needed to oer the screening69. However, we did not have the ability to
screen out any patient group or target people for future intervention without the risk of missing or excluding
people. It is dicult to dene a threshold for predictive accuracy that would be acceptable. Dierent accuracy
thresholds may be acceptable depending on multiple factors including the specic use case (e.g., prioritizing
screening vs. excluding a subpopulation from screening), institutional resources, and other factors. Based on our
model prediction and AUC, our results indicate that providers need to continue to use universal oer to screen
approaches while more research is conducted on how to best model social needs status and on the clinical and
cost eectiveness of social needs screening across healthcare settings27. Ideally, a future analysis would apply data
from all 28 AHC Model sites in the US coupled with their EHR data to provide a large and geographically diverse
enough sample to test the potential predictive power and application of risk modeling for HRSNs.
ird, the integration of SDOH with EHR data has implications for healthcare institutions with the shi to
value-based care in the U.S.39. e use of community and individual level data could help identify factors associ-
ated with social needs to improve healthcare utilization and health outcomes.
We examined the predictive power of HRSN status using community-level SDOH data with individual patient
EHR data. We found the addition of SDOH data led to very little improvement in model performance, with the
exception of the presence of a utility need. Models trained with EHR and SDOH data performed better than
Medicaid insurance statusalone. However, screening only these patients identied by the better performing
models would miss many patients with HRSNs. Future studies should examine variation of SDOH and EHR data
in a geographically broader patient sample to identify possible model enhancements to predict HRSN status and
prioritize patients for social interventions.
Data availability
e AHC and EHR datasets generated during and/or analyzed during the current study are not publicly available
due to identifying beneciaries and clinical site information, but are available from the corresponding author
on reasonable request.
Received: 24 September 2021; Accepted: 3 March 2022
References
1. Social Determinants of Health (SDOH). NEJM Catalyst (2017).
2. Green, K. & Zook, M. When Talking About Social Determinants, Precision Matters | Health Aairs Forefront. https:// doi. org/ 10.
1377/ foref ront. 20191 025. 776011/ full/.
Content courtesy of Springer Nature, terms of use apply. Rights reserved
8
Vol:.(1234567890)
Scientic Reports | (2022) 12:4554 | https://doi.org/10.1038/s41598-022-08344-4
www.nature.com/scientificreports/
3. Silverman, J. et al. e relationship between food insecurity and depression, diabetes distress and medication adherence among
low-income patients with poorly-controlled diabetes. J Gen Intern Med 30, 1476–1480 (2015).
4. Chambers, E. C., McAuli, K. E., Heller, C. G., Fiori, K. & Hollingsworth, N. Toward understanding social needs among primary
care patients with uncontrolled diabetes. J. Prim. Care Community Health 12, 2150132720985044 (2021).
5. Nagata, J. M. et al. Food insecurity and chronic disease in US young adults: Findings from the national longitudinal study of
adolescent to adult health. J. Gen. Intern. Med. 34, 2756–2762 (2019).
6. Venci, B. J. & Lee, S.-Y. Functional limitation and chronic diseases are associated with food insecurity among U.S. adults. Ann.
Epidemiol. 28, 182–188 (2018).
7. Jih, J. et al. Chronic disease burden predicts food insecurity among older adults. Public Health Nutr. 21, 1737–1742 (2018).
8. Cook, J. T. et al. A brief indicator of household energy security: associations with food security, child health, and child development
in US infants and toddlers. Pediatrics 122, e867-875 (2008).
9. Berkowitz, S. A., Seligman, H. K., Meigs, J. B. & Basu, S. Food insecurity, healthcare utilization, and high cost: a longitudinal cohort
study. Am. J. Manag. Care 24, 399–404 (2018).
10. McQueen, A. et al. Social needs, chronic conditions, and health care utilization among medicaid beneciaries. Popul Health Manag
24, 681–690 (2021).
11. Hill-Briggs, F. et al. Social determinants of health and diabetes: A scientic review. Diabetes Care 44, 258–279. https:// doi. org/ 10.
2337/ dci20- 0053 (2020).
12. Cassarino, M. et al. Impact of early assessment and intervention by teams involving health and social care professionals in the
emergency department: A systematic review. PLoS ONE 14, e0220709 (2019).
13. Hatef, E. et al. Assessing the impact of social needs and social determinants of health on health care utilization: Using patient- and
community-level data. Popul. Health Manag. 24, 222–230 (2021).
14. Knighton, A. J., Stephenson, B. & Savitz, L. A. Measuring the eect of social determinants on patient outcomes: A systematic
literature review. J. Health Care Poor Underserved 29, 81–106 (2018).
15. Berkowitz, S. A., Baggett, T. P. & Edwards, S. T. Addressing health-related social needs: Value-based care or values-based care?. J.
Gen. Intern. Med. 34, 1916–1918 (2019).
16. Gurewich, D., Garg, A. & Kressin, N. R. Addressing social determinants of health within healthcare delivery systems: A framework
to ground and inform health outcomes. J. Gen. Intern. Med. 35, 1571–1575 (2020).
17. National Academies of Sciences Engineering, and Medicine. Integrating Social Care into the Delivery of Health Care: Moving
Upstream to Improve the Nation’s Health. Integrating Social Care into the Delivery of Health Care: Moving Upstream to Improve the
Nation’s Health (National Academies Press (US), 2019).
18. Byho, E., Freund, K. M. & Garg, A. Accelerating the implementation of social determinants of health interventions in internal
medicine. J. Gen. Intern. Med. 33, 223–225 (2018).
19. Cole, M. B. & Nguyen, K. H. Unmet social needs among low-income adults in the United States: Associations with health care
access and quality. Health Serv. Res. 55(Suppl 2), 873–882 (2020).
20. Kusnoor, S. V. et al. Collection of social determinants of health in the community clinic setting: A cross-sectional study. BMC
Public Health 18, 550 (2018).
21. LaForge, K. et al. How 6 organizations developed tools and processes for social determinants of health screening in primary care:
An overview. J. Ambul. Care Manag. 41, 2–14 (2018).
22. Alley, D. E., Asomugha, C. N., Conway, P. H. & Sanghavi, D. M. Accountable health communities-addressing social needs through
medicare and medicaid. N. Engl. J. Med. 374, 8–11 (2016).
23. Billioux, A., Verlander, K., Anthony, S. & Alley, D. Standardized screening for health-related social needs in clinical settings: e
accountable health communities screening tool. NAM Perspectives https:// doi. org/ 10. 31478/ 20170 5b (2017).
24. Weir, R. C. et al. Collecting social determinants of health data in the clinical setting: Findings from national PRAPARE implemen-
tation. J. Health Care Poor Underserved 31, 1018–1035 (2020).
25. Cottrell, E. K. et al. Comparison of community-level and patient-level social risk data in a network of community health centers.
JAMA Netw. Open 3, e2016852 (2020).
26. Cantor, M. N. & orpe, L. Integrating data on social determinants of health into electronic health records. Health A. (Millwood)
37, 585–590 (2018).
27. Andermann, A. Screening for social determinants of health in clinical care: Moving from the margins to the mainstream. Public
Health Rev 39, 19 (2018).
28. O’Gurek, D. T. & Henke, C. A practical approach to screening for social determinants of health. Fam. Pract. Manag. 25, 7–12
(2018).
29. Fraze, T. K. et al. Prevalence of Screening for Food Insecurity, Housing Instability, Utility Needs, Transportation Needs, and
Interpersonal Violence by US Physician Practices and Hospitals. JAMA Netw. Open 2, e1911514 (2019).
30. Schickedanz, A., Hamity, C., Rogers, A., Sharp, A. L. & Jackson, A. Clinician experiences and attitudes regarding screening for
social determinants of health in a large integrated health system. Med. Care 57(Suppl 6 Suppl 2), S197–S201 (2019).
31. Samuels-Kalow, M. E. et al. Screening for health-related social needs of emergency department patients. Ann. Emerg. Med. 77,
62–68 (2021).
32. Fenton. Health care’s blind side: e overlooked connection between social needs and good health, summary of ndings from
a survey of America’s physicians | SIREN. https:// siren etwork. ucsf. edu/ tools- resou rces/ resou rces/ health- cares- blind- side- overl
ooked- conne ction- betwe en- social- needs- and.
33. Palacio, A. et al. Provider perspectives on the collection of social determinants of health. Popul. Health Manag. 21, 501–508 (2018).
34. Vest, J. R. & Ben-Assuli, O. Prediction of emergency department revisits using area-level social determinants of health measures
and health information exchange information. Int. J. Med. Inform. 129, 205–210 (2019).
35. Hao, S. et al. Development, validation and deployment of a real time 30 day hospital readmission risk assessment tool in the Maine
healthcare information exchange. PLoS ONE 10, e0140271 (2015).
36. Jin, B. et al. Prospective stratication of patients at risk for emergency department revisit: Resource utilization and population
management strategy implications. BMC Emerg. Med. 16, 10 (2016).
37. Ye, C. et al. Prediction of incident hypertension within the next year: prospective study using statewide electronic health records
and machine learning. J. Med Internet Res. 20, e22 (2018).
38. Nijhawan, A. E. et al. An electronic medical record-based model to predict 30-day risk of readmission and death among HIV-
infected inpatients. J. Acquir. Immune Dec. Syndr. 61, 349–358 (2012).
39. Chen, S. et al. Using applied machine learning to predict healthcare utilization based on socioeconomic determinants of care. Am.
J. Manag. Care 26, 26–31 (2020).
40. Zhang, Y. et al. Assessing the impact of social determinants of health on predictive models for potentially avoidable 30-day read-
mission or death. PLoS ONE 15, e0235064 (2020).
41. Bhavsar, N. A., Gao, A., Phelan, M., Pagidipati, N. J. & Goldstein, B. A. Value of neighborhood socioeconomic status in predicting
risk of outcomes in studies that use electronic health record data. JAMA Netw. Open 1, e182716 (2018).
42. Assessment of Social Factors Impacting Health Care Quality in Texas Medicaid. 22.
Content courtesy of Springer Nature, terms of use apply. Rights reserved
9
Vol.:(0123456789)
Scientic Reports | (2022) 12:4554 | https://doi.org/10.1038/s41598-022-08344-4
www.nature.com/scientificreports/
43. Kasthurirathne, S. N., Vest, J. R., Menachemi, N., Halverson, P. K. & Grannis, S. J. Assessing the capacity of social determinants of
health data to augment predictive models identifying patients in need of wraparound social services. J. Am. Med. Inform. Assoc.
25, 47–53 (2018).
44. Miller, R. L. et al. Evaluating testing strategies for identifying youths with HIV infection and linking youths to biomedical and
other prevention services. JAMA Pediatr. 171, 532–537 (2017).
45. Linda Higheld, P. et al. A conceptual framework for addressing social needs through the accountable health communities model.
(2020).
46. Hammond, G. & Joynt Maddox, K. E. A theoretical framework for clinical implementation of social determinants of health. JAMA
Cardiol. 4, 1189–1190 (2019).
47. Kolak, M., Bhatt, J., Park, Y. H., Padrón, N. A. & Molefe, A. Quantication of neighborhood-level social determinants of health in
the continental United States. JAMA Netw. Open 3, e1919928 (2020).
48. Krause, T. M., Schaefer, C. & Higheld, L. e association of social determinants of health with health outcomes. Am. J. Manag.
Care 27, e89–e96 (2021).
49. Lee, J. S. & Frongillo, E. A. Factors associated with food insecurity among U.S. elderly persons: Importance of functional impair-
ments. J. Gerontol. B Psychol. Sci. Soc. Sci. 56, S94-99 (2001).
50. Meddings, J. et al. e impact of disability and social determinants of health on condition-specic readmissions beyond medicare
risk adjustments: A cohort study. J Gen Intern Med 32, 71–80 (2017).
51. Joe, E. et al. A benchmark comparison of deterministic and probabilistic methods for dening manual review datasets in duplicate
records reconciliation. J. Am. Med. Inform. Assoc. 21, 97–104 (2014).
52. Consort - Welcome to the CONSORT Website. http:// www. conso rt- state ment. org/.
53. Ke, G. et al. LightGBM: A highly ecient gradient boosting decision tree. In Advances in Neural Information Processing Systems
Vol. 30 (Curran Associates, Inc., 2017).
54. Hosseini, M. et al. I tried a bunch of things: e dangers of unexpected overtting in classication of brain data. Neurosci. Biobehav.
Rev. 119, 456–467 (2020).
55. GitHub - scikit-optimize/scikit-optimize: Sequential model-based optimization with a `scipy.optimize` interface. https:// github.
com/ scikit- optim ize/ scikit- optim ize.
56. Sokolova, M. & Lapalme, G. A systematic analysis of performance measures for classication tasks. Inf. Process. Manage. 45,
427–437 (2009).
57. Alberti, P. M. & Baker, M. C. Dual eligible patients are not the same: How social risk may impact quality measurement’s ability to
reduce inequities. Medicine 99, e22245 (2020).
58. Roberts, E. T., Mellor, J. M., McInerney, M. & Sabik, L. M. State variation in the characteristics of Medicare-Medicaid dual enrollees:
Implications for risk adjustment. Health Serv Res 54, 1233–1245 (2019).
59. Hwang, A., Keohane, L. & Sharma, L. Improving Care for Individuals Dually Eligible for Medicare and Medicaid: Preliminary
Findings from Recent Evaluations of the Financial Alignment Initiative. 8 (2019).
60. Lewis, C. C. et al. Comparing the performance of two social risk screening tools in a vulnerable subpopulation. J. Family Med.
Prim. Care 9, 5026–5034 (2020).
61. Weida, E. B., Phojanakong, P., Patel, F. & Chilton, M. Financial health as a measurable social determinant of health. PLoS ONE 15,
e0233359 (2020).
62. Payne, R. et al. Evaluating perceptions of social determinants of health and Part D star performance of Medicare Advantage-
contracted primary care providers serving a South Texas market. J. Manag. Care Spec. Pharm. 27, 544–553 (2021).
63. Ancker, J. S., Kim, M.-H., Zhang, Y., Zhang, Y. & Pathak, J. e potential value of social determinants of health in predicting health
outcomes. J. Am. Med. Inform. Assoc. 25, 1109–1110 (2018).
64. Kreuter, M. W. et al. How do social needs cluster among low-income individuals?. Popul Health Manag 24, 322–332 (2021).
65. Gottlieb, L. M., Francis, D. E. & Beck, A. F. Uses and misuses of patient- and neighborhood-level social determinants of health
data. Perm. J. 22, 18–078 (2018).
66. Fiori, K. P. et al. Unmet social needs and no-show visits in primary care in a US Northeastern Urban Health System, 2018–2019.
Am. J. Public Health 110, S242–S250 (2020).
67. Schickedanz, A. et al. Impact of social needs navigation on utilization among high utilizers in a large integrated health system: A
quasi-experimental study. J. Gen. Intern. Med. 34, 2382–2389 (2019).
68. Bensken, W. P., Alberti, P. M. & Koroukian, S. M. Health-related social needs and increased readmission rates: Findings from the
nationwide readmissions database. J. Gen. Intern. Med. 36, 1173–1180 (2021).
69. Why More Evidence is Needed on the Eectiveness Of Screening For Social Needs Among High-Use Patients In Acute Care Set-
tings | Health Aairs Forefront. https:// doi. org/ 10. 1377/ foref ront. 20190 520. 243444/ full/.
Acknowledgements
e authors would like to acknowledge the participating AHC Greater Houston area clinical delivery sites.
is work was supported by CMS,the Cullen Trust for Healthcare, the National Center for Advancing Transla-
tional Sciences (NCATS) under awards UL1TR000371 and U01TR002393; the Cancer Prevention and Research
Institute of Texas (CPRIT), under award RP170668, and the Reynolds and Reynolds Professorship in Clinical
Informatics.
Author contributions
All authors contributed to the design of the work. J.H and L.H wrote the main manuscript. L.G and L.C.O con-
ducted the data analysis and wrote the results section of the manuscript. All authors revised the manuscript and
approved the nal version.
Funding
is publication was supported by the Centers for Medicare and Medicaid Services (CMS) of the U.S. Depart-
ment of Health and Human Services (HHS) as part of a nancial assistance award totaling $529,632 with 10%
percentage funded by CMS/HHS and 90 percentage funded by non-government source(s), the Cullen Trust
for Healthcare. is work was supported in part by the National Center for Advancing Translational Sciences
(NCATS) under awards UL1TR000371 and U01TR002393; the Cancer Prevention and Research Institute of
Texas (CPRIT), under award RP170668, and the Reynolds and Reynolds Professorship in Clinical Informatics.
Competing interests
e authors declare no competing interests.
Content courtesy of Springer Nature, terms of use apply. Rights reserved
10
Vol:.(1234567890)
Scientic Reports | (2022) 12:4554 | https://doi.org/10.1038/s41598-022-08344-4
www.nature.com/scientificreports/
Additional information
Supplementary Information e online version contains supplementary material available at https:// doi. org/
10. 1038/ s41598- 022- 08344-4.
Correspondence and requests for materials should be addressed to E.V.B.
Reprints and permissions information is available at www.nature.com/reprints.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and
institutional aliations.
Open Access is article is licensed under a Creative Commons Attribution 4.0 International
License, which permits use, sharing, adaptation, distribution and reproduction in any medium or
format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the
Creative Commons licence, and indicate if changes were made. e images or other third party material in this
article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the
material. If material is not included in the article’s Creative Commons licence and your intended use is not
permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder. To view a copy of this licence, visit http:// creat iveco mmons. org/ licen ses/ by/4. 0/.
© e Author(s) 2022
Content courtesy of Springer Nature, terms of use apply. Rights reserved
1.
2.
3.
4.
5.
6.
Terms and Conditions
Springer Nature journal content, brought to you courtesy of Springer Nature Customer Service Center GmbH (“Springer Nature”).
Springer Nature supports a reasonable amount of sharing of research papers by authors, subscribers and authorised users (“Users”), for small-
scale personal, non-commercial use provided that all copyright, trade and service marks and other proprietary notices are maintained. By
accessing, sharing, receiving or otherwise using the Springer Nature journal content you agree to these terms of use (“Terms”). For these
purposes, Springer Nature considers academic use (by researchers and students) to be non-commercial.
These Terms are supplementary and will apply in addition to any applicable website terms and conditions, a relevant site licence or a personal
subscription. These Terms will prevail over any conflict or ambiguity with regards to the relevant terms, a site licence or a personal subscription
(to the extent of the conflict or ambiguity only). For Creative Commons-licensed articles, the terms of the Creative Commons license used will
apply.
We collect and use personal data to provide access to the Springer Nature journal content. We may also use these personal data internally within
ResearchGate and Springer Nature and as agreed share it, in an anonymised way, for purposes of tracking, analysis and reporting. We will not
otherwise disclose your personal data outside the ResearchGate or the Springer Nature group of companies unless we have your permission as
detailed in the Privacy Policy.
While Users may use the Springer Nature journal content for small scale, personal non-commercial use, it is important to note that Users may
not:
use such content for the purpose of providing other users with access on a regular or large scale basis or as a means to circumvent access
control;
use such content where to do so would be considered a criminal or statutory offence in any jurisdiction, or gives rise to civil liability, or is
otherwise unlawful;
falsely or misleadingly imply or suggest endorsement, approval , sponsorship, or association unless explicitly agreed to by Springer Nature in
writing;
use bots or other automated methods to access the content or redirect messages
override any security feature or exclusionary protocol; or
share the content in order to create substitute for Springer Nature products or services or a systematic database of Springer Nature journal
content.
In line with the restriction against commercial use, Springer Nature does not permit the creation of a product or service that creates revenue,
royalties, rent or income from our content or its inclusion as part of a paid for service or for other commercial gain. Springer Nature journal
content cannot be used for inter-library loans and librarians may not upload Springer Nature journal content on a large scale into their, or any
other, institutional repository.
These terms of use are reviewed regularly and may be amended at any time. Springer Nature is not obligated to publish any information or
content on this website and may remove it or features or functionality at our sole discretion, at any time with or without notice. Springer Nature
may revoke this licence to you at any time and remove access to any copies of the Springer Nature journal content which have been saved.
To the fullest extent permitted by law, Springer Nature makes no warranties, representations or guarantees to Users, either express or implied
with respect to the Springer nature journal content and all parties disclaim and waive any implied warranties or warranties imposed by law,
including merchantability or fitness for any particular purpose.
Please note that these rights do not automatically extend to content, data or other material published by Springer Nature that may be licensed
from third parties.
If you would like to use or distribute our Springer Nature journal content to a wider audience or on a regular basis or in any other manner not
expressly permitted by these Terms, please contact Springer Nature at
onlineservice@springernature.com
... These data can provide a longitudinal and comprehensive view of the patient and are not dependent upon a single organization for data collection. ML predictive modeling has already demonstrated promise in identifying patients with HRSNs [19][20][21]. Nevertheless, the performance of ML predictive models has not yet been compared to other methods of identifying patients needing HRSN services. ...
... Overall, the ML EHR model performed better than the screening questionnaire model. The ML EHR modeling approach has the obvious advantage of drawing upon more information and, thus, is an increasingly preferred approach to HRSN measurement [19][20][21]. While model performance still requires significant improvement, our findings highlight the potential value of several EHR data elements already accessible to healthcare organizations. ...
Article
Full-text available
Background Health-related social needs (HRSNs), such as housing instability, food insecurity, and financial strain, are increasingly prevalent among patients. Healthcare organizations must first correctly identify patients with HRSNs to refer them to appropriate services or offer resources to address their HRSNs. Yet, current identification methods are suboptimal, inconsistently applied, and cost prohibitive. Machine learning (ML) predictive modeling applied to existing data sources may be a solution to systematically and effectively identify patients with HRSNs. The performance of ML predictive models using data from electronic health records (EHRs) and other sources has not been compared to other methods of identifying patients needing HRSN services. Methods A screening questionnaire that included housing instability, food insecurity, transportation barriers, legal issues, and financial strain was administered to adult ED patients at a large safety-net hospital in the mid-Western United States (n = 1,101). We identified those patients likely in need of HRSN-related services within the next 30 days using positive indications from referrals, encounters, scheduling data, orders, or clinical notes. We built an XGBoost classification algorithm using responses from the screening questionnaire to predict HRSN needs (screening questionnaire model). Additionally, we extracted features from the past 12 months of existing EHR, administrative, and health information exchange data for the survey respondents. We built ML predictive models with these EHR data using XGBoost (ML EHR model). Out of concerns of potential bias, we built both the screening question model and the ML EHR model with and without demographic features. Models were assessed on the validation set using sensitivity, specificity, and Area Under the Curve (AUC) values. Models were compared using the Delong test. Results Almost half (41%) of the patients had a positive indicator for a likely HRSN service need within the next 30 days, as identified through referrals, encounters, scheduling data, orders, or clinical notes. The screening question model had suboptimal performance, with an AUC = 0.580 (95%CI = 0.546, 0.611). Including gender and age resulted in higher performance in the screening question model (AUC = 0.640; 95%CI = 0.609, 0.672). The ML EHR models had higher performance. Without including age and gender, the ML EHR model had an AUC = 0.765 (95%CI = 0.737, 0.792). Adding age and gender did not improve the model (AUC = 0.722; 95%CI = 0.744, 0.800). The screening questionnaire models indicated bias with the highest performance for White non-Hispanic patients. The performance of the ML EHR-based model also differed by race and ethnicity. Conclusion ML predictive models leveraging several robust EHR data sources outperformed models using screening questions only. Nevertheless, all models indicated biases. Additional work is needed to design predictive models for effectively identifying all patients with HRSNs.
... Previous studies have used SDOH to predict various health-related outcomes. For example, a study predicted health-related social needs among Medicare and Medicaid beneficiaries using SDOH, achieving moderate prediction accuracy [6]. A recent study utilized SDOH in the All of Us (AoU) dataset to predict depression, delayed medical care, and emergency room visits [7]. ...
Article
Full-text available
This study applied machine learning (ML) algorithms to predict health-related quality of life (HRQOL) using comprehensive social determinants of health (SDOH) features. Data from the All of Us dataset, comprising participants with complete HRQOL and SDOH records, were analyzed. The primary outcome was HRQOL, which encompassed physical and mental health components, while SDOH features included social, educational, economic, environmental, and healthcare access factors. Three ML algorithms, namely logistic regression, XGBoost, and Random Forest, were tested. The models achieved accuracy ranges of 0.73–0.77 for HRQOL, 0.70–0.71 for physical health, and 0.72–0.77 for mental health, with corresponding area under the curve ranges of 0.81–0.84, 0.74–0.76, and 0.83–0.85, respectively. Emotional stability, activity management, spiritual beliefs, and comorbidity were identified as key predictors. These findings underscore the critical role of SDOH in predicting HRQOL and suggests future research to focus on applying such models to diverse patient populations and specific clinical conditions.
... 68 A growing body of research is exploring the potential of ML algorithms to classify patients' HRSNs. These studies have used composite measures of service utilization, 69,70 positive responses to screening tools, 71 or chart review as the prediction targets. 72 This study relied on patient reported HRSN status on validated instruments as the reference standard for the prediction target, which is a potential advantage over other targets. ...
Article
Full-text available
Objective Measurement of health-related social needs (HRSNs) is complex. We sought to develop and validate computable phenotypes (CPs) using structured electronic health record (EHR) data for food insecurity, housing instability, financial insecurity, transportation barriers, and a composite-type measure of these, using human-defined rule-based and machine learning (ML) classifier approaches. Materials and Methods We collected HRSN surveys as the reference standard and obtained EHR data from 1550 patients in 3 health systems from 2 states. We followed a Delphi-like approach to develop the human-defined rule-based CP. For the ML classifier approach, we trained supervised ML (XGBoost) models using 78 features. Using surveys as the reference standard, we calculated sensitivity, specificity, positive predictive values, and area under the curve (AUC). We compared AUCs using the Delong test and other performance measures using McNemar's test, and checked for differential performance. Results Most patients (63%) reported at least one HRSN on the reference standard survey. Human-defined rule-based CPs exhibited poor performance (AUCs=.52 to .68). ML classifier CPs performed significantly better, but still poor-to-fair (AUCs = .68 to .75). Significant differences for race/ethnicity were found for ML classifier CPs (higher AUCs for White non-Hispanic patients). Important features included number of encounters and Medicaid insurance. Discussion Using a supervised ML classifier approach, HRSN CPs approached thresholds of fair performance, but exhibited differential performance by race/ethnicity. Conclusion CPs may help to identify patients who may benefit from additional social needs screening. Future work should explore the use of area-level features via geospatial data and natural language processing to improve model performance.
... Social risks were based on items from the SDOH-HE module comprised of 10 questions assessing life satisfaction, social and emotional support, social isolation, employment stability, food security (2 questions), housing security, utility security, transportation access, and mental well-being. [26][27][28] For consistency, we used the CDC's suggested recoding from participants' original response options to each question (eTable 2 in Supplement 1). 29 We also included a variable for cost as a barrier to health care access. ...
Article
Full-text available
Importance Research indicates that social drivers of health are associated with cancer screening adherence, although the exact magnitude of these associations remains unclear. Objective To investigate the associations between individual-level social risks and nonadherence to guideline-recommended cancer screenings. Design, Setting, and Participants This cross-sectional study used 2022 Behavioral Risk Factor Surveillance System data from 39 US states and Washington, DC. Analyses for each specific cancer screening subsample were limited to screening-eligible participants according to the latest US Preventive Services Task Force (USPSTF) guidelines. Data were analyzed from February 22 to June 5, 2024. Exposures Ten social risk items, including life satisfaction, social and emotional support, social isolation, employment stability, food security (2 questions), housing security, utility security, transportation access, and mental well-being. Main Outcomes and Measures Up-to-date status (adherence) was assessed using USPSTF definitions. Adjusted risk ratios (ARRs) and 95% CIs were estimated using modified Poisson regression with robust variance estimator. Results A total of 147 922 individuals, representing a weighted sample of 78 784 149 US adults, were included in the analysis (65.8% women; mean [SD] age, 56.1 [13.3] years). The subsamples included 119 113 individuals eligible for colorectal cancer screening (CRCS), 7398 eligible for lung cancer screening (LCS), 56 585 eligible for cervical cancer screening (CCS), and 54 506 eligible for breast cancer screening (BCS). Findings revealed slight differences in effect size magnitude and in some cases direction; therefore results were stratified by sex, although precision was reduced for LCS. For the social contextual variables, life dissatisfaction was associated with nonadherence for CCS (ARR, 1.08; 95% CI, 1.01-1.16) and BCS (ARR, 1.22; 95% CI, 1.15-1.29). Lack of support was associated with nonadherence in CRCS in men and women and BCS, as was feeling isolated in CRCS in women and BCS. An association with feeling mentally distressed was seen in BCS. Under economic stability, food insecurity was associated with increased risk of nonadherence in CRCS in both men and women, CCS, and BCS; the direction of effect sizes for LCS were the same, but were not statistically significant. Under built environment, transportation insecurity was associated with nonadherence in CRCS in women and BCS, and cost barriers to health care access were associated with increased risk of nonadherence in CRCS for both men and women, LCS in women, and BCS, with the greatest risk and with reduced precision seen in LCS in women (ARR, 1.54; 95% CI, 1.01-2.33). Conclusions and Relevance In this cross-sectional study of adults eligible for cancer screening, findings revealed notable variations in screening patterns by both screening type and sex. Given that these risks may not always align with patient-centered social needs, further research focusing on specific target populations is essential before effective interventions can be implemented.
... Our logistic regression (Models I and II) and GEE models (Models III and IV) had satisfactory performance across the overall populations and subpopulations of interest. For the models using EHR structured data (Models I and III), the model performance in our study was comparable with those in the study by Holcomb et al, 27 where they predicted healthrelated social needs using EHR structured data and community-level data and machine learning modeling for Medicare and Medicaid beneficiaries participating in the Accountable Health Communities project. Their models performed relatively well, with AUCs ranging from 0.59 to 0.68 for patients with different domains of social needs. ...
Article
Full-text available
Objective To improve the performance of a social risk score (a predictive risk model) using electronic health record (EHR) structured and unstructured data. Materials and Methods We used EPIC-based EHR data from July 2016 to June 2021 and linked it to community-level data from the US Census American Community Survey. We identified predictors of interest within the EHR structured data and applied natural language processing (NLP) techniques to identify patients’ social needs in the EHR unstructured data. We performed logistic regression models with and without information from the unstructured data (Models I and II) and compared their performance with generalized estimating equation (GEE) models with and without the unstructured data (Models III and IV). Results The logistic model (Model I) performed well (Area Under the Curve [AUC] 0.703, 95% confidence interval [CI] 0.701:0.705) and the addition of EHR unstructured data (Model II) resulted in a slight change in the AUC (0.701, 95% CI 0.699:0.703). In the logistic models, the addition of EHR unstructured data resulted in an increase in the area under the precision-recall curve (PRC 0.255, 95% CI 0.254:0.256 in Model I versus 0.378, 95% CI 0.375:0.38 in Model II). The GEE models performed similarly to the logistic models and the addition of EHR unstructured data resulted in a slight change in the AUC (0.702, 95% CI 0.699:0.705 in Model III versus 0.699, 95% CI 0.698:0.702 in Model IV). Discussion Our work presents the enhancement of a novel social risk score that integrates community-level data with patient-level data to systematically identify patients at increased risk of having future social needs for in-depth assessment of their social needs and potential referral to community-based organizations to address these needs. Conclusion The addition of information on social needs extracted from unstructured EHR resulted in an improved prediction of positive cases presented by the improvement in the PRC.
... Our secondary exposure of interest was a social needs measure. The questions on social needs measure were based on the Center for Medicare and Medicaid Innovation Social Needs Assessment Tool, and asked about employment/economic stability, housing stability and quality, food security, transportation access, utilities security, loneliness, social and emotional support, life satisfaction, and mental stress (De Marchis et al., 2020;Holcomb et al., 2022;Thomas-Henkel and Schulman, 2023). We used ten questions from the SDOH/ health equity module to calculate the summarizing score of social needs measure based on the BRFSS statistical brief report (Supplementary Table S1)(Centers for Disease Control and Prevention, 2023b). ...
Article
Full-text available
Objective We sought to examine the influence of social needs on the relationship between cancer history and colorectal cancer (CRC) screening utilization among adults in the United States. Methods We conducted a cross-sectional analysis using data from the 2022 Behavioral Risk Factor Surveillance System. Our outcome of interest was utilization of guideline-concordant CRC screening and exposures of interest were cancer history/levels of social needs. Multivariable logistic regression was performed to examine the association. Results Among 74,743 eligible adults, a majority did not have a personal history of cancer (87.9 %), had at least one social need (58.4 %), and had undergone CRC screening (72.2 %). In multivariable analysis, a history of cancer was positively associated with use of CRC screening (OR = 1.59, 95 %CI, 1.35 – 1.87). Having at least one social need was associated with lower likelihood of being screened (one social need: OR = 0.85 95 %CI, 0.76 – 0.95; two + social needs: OR = 0.77, 95 % CI, 0.69 – 0.87). When exploring the effects of social needs, adults without a history of cancer who reported at least one need were 12–20 % less likely to be screened for CRC. Conclusions A personal history of cancer was associated with greater utilization of CRC screening, whilst having at least one social need had lower screening use. Having social needs plays an important role in reducing screening uptake among adults without a history of cancer. Integrated care that considers both cancer history and social needs may have implications for improved adherence of CRC screening recommendations.
Article
Many social need screening to advance population health and reduce health disparities, but barriers to screening remain. Improved knowledge of patient populations at risk for social needs based on administrative data could facilitate more targeted practices, and by extension, feasible social need screening and referral efforts. To illustrate the use of cluster analysis to identify patient population segments at risk for social needs. We used clustering analysis to identify population segments among Veterans (N=2010) who participated in a survey assessing nine social needs (food, housing, utility, financial, employment, social disconnection, legal, transportation, and neighborhood safety). Clusters were based on eight variables (age, race, gender, comorbidity, region, no-show rate, rurality, and VA priority group). We used weighted logistic regression to assess association of clusters with the risk of experiencing social needs. National random sample of Veterans with and at risk for cardiovascular disease who responded to a mail survey (N=2010). Self-reported social needs defined as the risk of endorsing (1) each individual social need, (2) one or more needs, and (3) a higher total count of needs. From the clustering analysis process with sensitivity analysis, we identified a consistent population segment of Veterans. From regression modeling, we found that this cluster, with lower average age and higher proportions of women and racial minorities, was at higher risk of experiencing ≥ 1 unmet need (OR 1.74, CI 1.17–2.56). This cluster was also at a higher risk for several individual needs, especially utility needs (OR 3.78, CI 2.11–6.78). The identification of characteristics associated with increased unmet social needs may provide opportunities for targeted screenings. As this cluster was also younger and had fewer comorbidities, they may be less likely to be identified as experiencing need through interactions with healthcare providers.
Article
Objective We sought to examine the associations between a social needs measure and physical, and mental health among cancer survivors in the United States. Methods We conducted a cross‐sectional analysis using the 2022 Behavioral Risk Factor Surveillance System survey involving 16,930 eligible cancer survivors. The primary outcomes of interest were self‐reported physical and mental health in the past 30 days. A social needs measure was our primary exposure of interest. Multivariate logistic regression was used to examine the associations of interest. Results Overall, 56% and 73% survivors with several days of poor physical and mental health, respectively, reported having two or more social needs. In multivariate analysis, those having at least one social need were more likely to report several days of poor physical (one need: OR, 1.62; 95% CI, 1.31–2.00, two or more needs: OR, 3.52; 95% CI, 2.84–4.35) and mental health (one need: OR, 3.07; 95% CI, 2.07–4.57, two or more needs: OR, 9.69; 95% CI, 6.83–13.74). Among survivors with two or more social needs, having exercised in the past 30 days were 41% and 59% less likely to experience poor physical and mental health, respectively ( p ‐value < 0.05). However, having at least one chronic disease was associated with several days of poor physical/mental health among those with two or more needs ( p ‐value < 0.05). Conclusion Having social needs was associated with self‐reporting of several days of poor physical and mental health among cancer survivors. Integrated care should include mental/physical health management addressing cancer survivors' various social needs.
Conference Paper
Machine learning (ML) is crucial in modern healthcare because it can quickly and accurately detect hidden patterns that would otherwise go unnoticed. Current healthcare systems have only implemented a small number of the many ML applications that have been found or are currently being studied. Healthcare professionals face significant challenges when trying to implement ML applications due to dispersed data and a lack of well-structured, easily-explained documentation in the relevant sector, despite the fact that ML presents a huge potential in the healthcare system. In order to provide quick access to important information, this research sought to consolidate ML applications across healthcare domains in a more efficient and effective manner. Our research was organised into five main categories: community-based initiatives, healthcare operations management, early detection, remote care, risk management/preventive care, and healthcare operations. In order to facilitate easy access, we have divided these groupings into subgroups and included pertinent references and descriptions in tabular form. We want to drive healthcare professionals towards a healthcare system that is more based on machine learning, educate the public about the benefits of machine learning in healthcare, and close the knowledge gap among clinicians on the uses of machine learning.
Article
Full-text available
Objectives: This study explored the contributions of social determinants of health (SDOH) to measures of population health-specifically cost, hospitalization rates, rate of emergency department utilization, and health status-in Texas. Study design: The study associated common SDOH metrics from public data sources (county specific) with health plan enrollment data (including demographics, counties, and zip codes) and medical and pharmaceutical annual claims data. Methods: Following correlation analyses to reduce variables, the contribution of each SDOH individually and by category to the health outcomes was evaluated. Separate matrices for age populations (under age 19, general population [all ages], and ≥ 65 years) were created with assigned weights of influence for categories and the factors within each category. Results: The contributions of the categories varied by population, confirming that different SDOH influence populations to varying degrees. This was reflected in each model. The largest contributor to cost for the general population and for the group 65 years and older was factors grouped as health outcomes (such as perceived health), at 43.5% contribution and 37.7% contribution, respectively. Yet for the population younger than 19 years, the largest contributor to cost was socioeconomic factors (such as unemployment rate), at 40.2%. The other performance measures also varied by population and the mix and weight of determinants. Conclusions: This study and the developed population-based matrices can provide a valuable framework for reporting the impact of SDOH on health care quality. The variation suggests the need for further research on how age groups react to the social environment.
Article
Full-text available
Introduction/Objectives Uncontrolled diabetes can lead to major health complications, and significantly contributes to diabetes-related morbidity, mortality, and healthcare costs. Few studies have examined the relationship between unmet social needs and diabetes control among predominantly Black and Hispanic patient populations. Methods In a large urban hospital system in the Bronx, NY, 5846 unique patients with diabetes seen at a primary care visit between April 2018 and December 2019 completed a social needs screener. Measures included diabetes control (categorized as Hemoglobin (Hb) A1c <9.0 as controlled and Hb A1C ≥9.0 as uncontrolled), social needs (10-item screen), and demographic covariates, including age, sex, race/ethnicity, insurance status, percentage of block-group poverty, patient’s preferred language, and the Elixhauser Comorbidity Index. Results Twenty-two percent (22%) of the patient sample had at least 1 unmet social need, and the most prevalent unmet social needs were housing issues (including housing quality and insecurity), food insecurity, and lack of healthcare transportation. Logistic regression analysis showed a significant relationship between social needs and uncontrolled diabetes, with more social needs indicating a greater likelihood of uncontrolled diabetes (Adjusted Odds Ratio (AOR) for ≥3 needs: 1.59, 95% CI: 1.26, 2.00). Of the patients with most frequently occurring unmet social needs, lack of healthcare transportation (AOR: 1.54, 95% CI: 1.22, 1.95) and food insecurity (AOR: 1.50, 95% CI: 1.19, 1.89) had the greatest likelihood of having uncontrolled diabetes, after adjusting for covariates. Conclusion Unmet social needs appear to be linked to a greater likelihood of uncontrolled diabetes. Implications for healthcare systems to screen and address social needs for patients with diabetes are discussed.
Article
Full-text available
Machine learning has enhanced the abilities of neuroscientists to interpret information collected through EEG, fMRI, and MEG data. With these powerful techniques comes the danger of overfitting of hyperparameters which can render results invalid. We refer to this problem as ‘overhyping’ and show that it is pernicious despite commonly used precautions. Overhyping occurs when analysis decisions are made after observing analysis outcomes and can produce results that are partially or even completely spurious. It is commonly assumed that cross-validation is an effective protection against overfitting or overhyping, but this is not actually true. In this article, we show that spurious results can be obtained on random data by modifying hyperparameters in seemingly innocuous ways, despite the use of cross-validation. We recommend a number of techniques for limiting overhyping, such as lock boxes, blind analyses, pre-registrations, and nested cross-validation. These techniques, are common in other fields that use machine learning, including computer science and physics. Adopting similar safeguards is critical for ensuring the robustness of machine-learning techniques in the neurosciences.
Article
Full-text available
Objectives:To present a conceptual framework to address social needs through accountable health programs in the United States. We present our implementation of the Accountable Health Communities (AHC) Model as an example.Study Design:Conceptual framework and case study.Methods:Our conceptual framework of the AHC Model is adapted from the UK Rainbow Model of Integrated Care for the US health care context based on current literature. Our approach is further underpinned by Medical Research Council guidance on intervention development stages, recent framework development in intervention coproduction, and quality improvement methods.Results:Our team used the adapted framework coupled with Medical Resource Council stages for intervention development to create a program scope and sequence. Standard operating procedures and an implemen-tation plan were created and approved by the Center for Medicare and Medicaid Innovation. Implementation was successfully launched at 13 clinical delivery sites and completed screenings of more than 10,000 patients have been done to date.Conclusions:Our conceptual framework, which we are applying as a bridge organization in the AHC Model, can serve as a model that organizations can use to successfully design interventions to address social needs in clin-ical care settings in the United States. Further application and testing of the framework are warranted to advance understanding of social needs interven-tions in the United States.
Article
Full-text available
Decades of research have demonstrated that diabetes affects racial and ethnic minority and low-income adult populations in the U.S. disproportionately, with relatively intractable patterns seen in these populations’ higher risk of diabetes and rates of diabetes complications and mortality (1). With a health care shift toward greater emphasis on population health outcomes and value-based care, social determinants of health (SDOH) have risen to the forefront as essential intervention targets to achieve health equity (2-4). Most recently, the COVID-19 pandemic has highlighted unequal vulnerabilities borne by racial and ethnic minority groups and by disadvantaged communities. In the wake of concurrent pandemic and racial injustice events in the U.S., the American College of Physicians, American Academy of Pediatrics, Society of General Internal Medicine, National Academy of Medicine, and other professional organizations have published statements on SDOH (5-8), and calls to action focus on amelioration of these determinants at individual, organizational, and policy levels (9-11). In diabetes, understanding and mitigating the impact of SDOH are priorities due to disease prevalence, economic costs, and disproportionate population burden (12-14). In 2013, the American Diabetes Association (ADA) published a scientific statement on socioecological determinants of prediabetes and type 2 diabetes (15). Toward the goal of understanding and advancing opportunities for health improvement among the population with diabetes through addressing SDOH, ADA convened the current SDOH and diabetes writing committee, prepandemic, to review the literature on 1) associations of SDOH with diabetes risk and outcomes and 2) impact of interventions targeting amelioration of SDOH on diabetes outcomes. This article begins with an overview of key definitions and SDOH frameworks. The literature review focuses primarily on U.S.-based studies of adults with diabetes and on five SDOH: Socioeconomic status (education, income, occupation); neighborhood and physical environment (housing, built environment, toxic environmental exposures); food environment (food insecurity, food access); health care (access, affordability, quality); and social context (social cohesion, social capital, social support). This review concludes with recommendations for linkages across health care and community sectors from national advisory committees, recommendations for diabetes research, and recommendations for research to inform practice.
Article
Full-text available
Importance: Responding to the substantial research on the relationship between social risk factors and health, enthusiasm has grown around social risk screening in health care settings, and numerous US health systems are experimenting with social risk screening initiatives. In the absence of standard social risk screening recommendations, some health systems are exploring using publicly available community-level data to identify patients who live in the most vulnerable communities as a way to characterize patient social and economic contexts, identify patients with potential social risks, and/or to target social risk screening efforts. Objective: To explore the utility of community-level data for accurately identifying patients with social risks by comparing the social deprivation index score for the census tract where a patient lives with patient-level social risk screening data. Design, setting, and participants: Cross-sectional study using patient-level social risk screening data from the electronic health records of a national network of community health centers between June 24, 2016, and November 15, 2018, linked to geocoded community-level data from publicly available sources. Eligible patients were those with a recorded response to social risk screening questions about food, housing, and/or financial resource strain, and a valid address of sufficient quality for geocoding. Exposures: Social risk screening documented in the electronic health record. Main outcomes and measures: Community-level social risk was assessed using census tract-level social deprivation index score stratified by quartile. Patient-level social risks were identified using food insecurity, housing insecurity, and financial resource strain screening responses. Results: The final study sample included 36 578 patients from 13 US states; 22 113 (60.5%) received public insurance, 21 181 (57.9%) were female, 17 578 (48.1%) were White, and 10 918 (29.8%) were Black. Although 6516 (60.0%) of those with at least 1 social risk factor were in the most deprived quartile of census tracts, patients with social risk factors lived in all census tracts. Overall, the accuracy of the community-level data for identifying patients with and without social risks was 48.0%. Conclusions and relevance: Although there is overlap, patient-level and community-level approaches for assessing patient social risks are not equivalent. Using community-level data to guide patient-level activities may mean that some patients who could benefit from targeted interventions or care adjustments would not be identified.
Article
Health care organizations are increasingly assessing patients' social needs (eg, food, utilities, transportation) using various measures and methods. Prior studies have assessed social needs at the point of care and many studies have focused on correlates of 1 specific need (eg, food). This comprehensive study examined multiple social needs and medical and pharmacy claims data. Medicaid beneficiaries in Louisiana (n = 10,275) completed a self-report assessment of 10 social needs during July 2018 to June 2019. Chronic health conditions, unique medications, and health care utilization were coded from claims data. The sample was predominantly female (72%), Black (45%) or White (32%), had a mean age of 42 years, and at least 1 social need (55%). In bivariate analyses, having greater social needs was associated with greater comorbidity across conditions, and each social need was consistently associated with mental health and substance use disorders. In multivariable logistic analyses, having ≥2 social needs was positively associated with emergency department (ED) visits (OR = 1.39, CI = 1.23 - 1.57) and negatively associated with wellness visits (OR = 0.87, CI = 0.77 - 0.98), inpatient visits (OR = 0.87, CI = 0.76 - 0.99), and 30-day rehospitalization (OR = 0.66, CI = 0.50 - 0.87). Findings highlight the greater concomitant risk of social needs, mental health, and substance use. Admission policies may reduce the impact of social needs on hospitalization. Chronic disease management programs offered by health plans may benefit from systematically assessing and addressing social needs outside point-of-care interactions to impact health outcomes and ED utilization. Behavioral health care management programs would benefit from integrating interventions for multiple social needs.
Article
BACKGROUND: Socioeconomic factors can have a significant impact on a patient's health status and could be responsible for as much as 70%-80% of a patient's overall health. These factors, called the social determinants of health (SDoH), define a patient's day-to-day experiences. While the influence of such factors is well recognized, who ultimately is responsible for addressing SDoH in health care remains unclear. Physicians and other clinicians are suitably placed to assess SDoH factors that can impact clinical decision making. Understanding Medicare Advantage (MA)-contracted primary care provider (PCP) SDoH perceptions has yet to be fully explored. OBJECTIVES: To (a) understand MA-contracted PCP perceptions of SDoH and (b) investigate correlations between PCP perceptions and their CMS Part D star performances, as well as their hospital admissions and emergency room admissions. METHODS: Survey data were collected from MA-contracted PCPs serving a South Texas market during 2019. An 8-item survey consisting of short answer, ranking, and multiple-choice questions was deployed at attendance-mandatory provider meetings from August to October. Analyses were conducted to understand the providers' SDoH perceptions. PCP responses were first summarized as frequencies and percentages. Baseline descriptive characteristics of the providers were compared by Medicare star ratings using chi-square tests (for categorical variables) and t-tests (for continuous variables). Group differences in physician beliefs on how SDoH affects patients' overall health (question 1), as well as provider beliefs regarding how SDoH affects patients' medication adherence practices (question 2), were assessed using chi-square and t-tests. Associations of provider SDoH perceptions with hospital admissions and emergency room admissions were also assessed. A Fischer's chi-square test was used to examine associations between how PCPs answered the question regarding lack of consistent transportation (question 3) and emergency room admissions. The relationships between PCP perceptions of whose job it is to address SDoH (question 7) and hospital admissions were also evaluated. RESULTS: The response rate for returned surveys was 89%. Analysis revealed that the top 3 barriers were financial insecurity (24.87%), low health literacy (18.65%), and social isolation (15.03%). However, about 36% of PCPs felt they should be the primary addressor of SDoH. There was a significant association between years of practice and CMS Part D star ratings (P = 0.005). A significant association between responses in belief towards patients' overall health and CMS Part D star ratings was examined (P = 0.047). There was a statistically significant difference in mean hospital admissions with PCP perception of who should address SDOH (P = 0.03). Emergency room admissions was significantly associated with perceptions regarding lack of consistent transportation (P = 0.04). No differences with star ratings were observed. CONCLUSIONS: Previous literature recognize safety and food insecurity as key SDoH barriers. However, they were not among the top SDoH barriers in our survey. Future research should examine patient perceptions of SDoH in this population to identify ways providers can better serve their patients. DISCLOSURES: Funding for this study was provided by CareAllies, a Cigna business. Statistical analysis was completed in partnership with the University of Houston. Payne, Esse, Qian, Serna, Villarreal, and Becho-Dominguez are employees of CareAllies. Mohan and Abughosh are employed by the University of Houston College of Pharmacy. Abughosh reports grants from Valeant and Regeneron/Sanofi, unrelated to this work. Vadhariya has nothing to disclose. This research was presented virtually at the AMCP Pharmacist Virtual Learning Days event, April 2020, as well as the American College of Clinical Pharmacy Virtual Poster Symposium, May 26-27, 2020.
Article
Background: While health-related social needs (HRSN) are known to compromise health, work to date has not clearly demonstrated the relationship between clinically acknowledged social needs, via ICD-10 Z-codes, and readmission. Objective: Assess the rate of 30-, 60-, and 90-day readmission by the level of ICD-10-identified social need. In addition, we examined the associations between demographics, social need, hospital characteristics, and comorbidities on 30-day readmission. Design: Retrospective study using the 2017 Nationwide Readmission Database PARTICIPANTS: We identified 5 domains of HRSN from ICD-10 diagnosis codes including employment, family, housing, psychosocial, and socioeconomic status (SES) and identified how many and which an individual was coded with during the year. Main measures: The proportion of patients with 30-, 60-, and 90-day readmission stratified by the number of HRSN domains with a multivariable logistic regression to examine the relationship between the number/type of and readmission adjusting for sex, age, payer, hospital characteristics, functional limitations, and comorbidities. Key results: From 13,217,506 patients, only 2.4% had at least one HRSN diagnosis. Among patients without HRSN, 11.5% had a 30-day readmission, compared to 27.0% of those with 1 domain, increasing to 63.5% for patients with codes in 5 domains. Similar trends were observed for 60- and 90-day readmission; 78.7% of patients with documented HRSN in all 5 domains were hospitalized again within 90 days. The adjusted odds ratio for readmission for individuals with all 5 domains was 12.55 (95% CI: 9.04, 17.43). Housing and employment emerged as two of the most commonly documented HRSN, as well as having the largest adjusted odds ratio. Conclusions: There is a dose-response relationship between the number of HRSN diagnoses and hospital readmission. This work calls attention to the need to develop interventions to reduce readmissions for those at social risk and demonstrates the significance of ICD-10 Z-codes in health outcomes studies.
Article
Study objective: There has been increasing attention to screening for health-related social needs. However, little is known about the screening practices of emergency departments (EDs). Within New England, we seek to identify the prevalence of ED screening for health-related social needs, understand the factors associated with screening, and understand how screening patterns for health-related social needs differ from those for violence, substance use, and mental health needs. Methods: We analyzed data from the 2018 National Emergency Department Inventory-New England survey, which was administered to all 194 New England EDs during 2019. We used descriptive statistics to compare ED characteristics by screening practices, and multivariable logistic regression models to identify factors associated with screening. Results: Among the 166 (86%) responding EDs, 64 (39%) reported screening for at least one health-related social need, 160 (96%) for violence (including intimate partner violence or other violent exposures), 148 (89%) for substance use disorder, and 159 (96%) for mental health needs. EDs reported a wide range of social work resources to address identified needs, with 155 (93%) reporting any social worker availability and 41 (27%) reporting continuous availability. Conclusion: New England EDs are screening for health-related social needs at a markedly lower rate than for violence, substance use, and mental health needs. EDs have relatively limited resources available to address health-related social needs. We encourage research on the development of scalable solutions for identifying and addressing health-related social needs in the ED.