ArticlePDF Available

Real-time imputation of missing predictor values improved the application of prediction models in daily practice

Authors:

Abstract and Figures

Objectives – In clinical practice, many prediction models cannot be used when predictor values are missing. We therefore propose and evaluate methods for real-time imputation. Study design and Setting – We describe (i) mean imputation (where missing values are replaced by the sample mean), (ii) joint modeling imputation (JMI, where we use a multivariate normal approximation to generate patient-specific imputations) and (iii) conditional modeling imputation (CMI, where a multivariable imputation model is derived for each predictor from a population). We compared these methods in a case study evaluating the root mean squared error (RMSE) and coverage of the 95% confidence intervals (i.e. the proportion of confidence intervals that contain the true predictor value) of imputed predictor values. Results –RMSE was lowest when adopting JMI or CMI, although imputation of individual predictors did not always lead to substantial improvements as compared to mean imputation. JMI and CMI appeared particularly useful when the values of multiple predictors of the model were missing. Coverage reached the nominal level (i.e. 95%) for both CMI and JMI. Conclusion – Multiple imputation using, either CMI or JMI, is recommended when dealing with missing predictor values in real time settings.
Content may be subject to copyright.
ORIGINAL ARTICLE
Real-time imputation of missing predictor values improved the
application of prediction models in daily practice
Steven Willem Joost Nijman
a,1,
*, T. Katrien J. Groenhof
a,1
, Jeroen Hoogland
a
, Michiel L. Bots
a
,
Menno Brandjes
b
, John J.L. Jacobs
b
, Folkert W. Asselbergs
c,d,e
, Karel G.M. Moons
a
,
Thomas P.A. Debray
a,d
a
Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Utrecht, the Netherlands
b
LogiqCare, Ortec B.V., Zoetermeer, the Netherlands
c
Division Heart & Lungs, Department of Cardiology, University Medical Center Utrecht, Utrecht University, Utrecht, the Netherlands
d
Institute of Cardiovascular Science, Faculty of Population Health Sciences, University College London, London, UK
e
Health Data Research UK, Institute of Health Informatics, University College London, London, UK
Accepted 12 January 2021; Published online 19 January 2021
Abstract
Objectives: In clinical practice, many prediction models cannot be used when predictor values are missing. We, therefore, propose and
evaluate methods for real-time imputation.
Study Design and Setting: We describe (i) mean imputation (where missing values are replaced by the sample mean), (ii) joint
modeling imputation (JMI, where we use a multivariate normal approximation to generate patient-specific imputations), and (iii) condi-
tional modeling imputation (CMI, where a multivariable imputation model is derived for each predictor from a population). We compared
these methods in a case study evaluating the root mean squared error (RMSE) and coverage of the 95% confidence intervals (i.e., the pro-
portion of confidence intervals that contain the true predictor value) of imputed predictor values.
Results: eRMSE was lowest when adopting JMI or CMI, although imputation of individual predictors did not always lead to substan-
tial improvements as compared to mean imputation. JMI and CMI appeared particularly useful when the values of multiple predictors of the
model were missing. Coverage reached the nominal level (i.e., 95%) for both CMI and JMI.
Conclusion: Multiple imputations using either CMI or JMI is recommended when dealing with missing predictor values in real-time
settings. Ó2021 The Authors. Published by Elsevier Inc. This is an open access article under the CC BY license (http://creativecommons.
org/licenses/by/4.0/).
Keywords: Missing data; Multiple imputations; Real-time imputation; Prediction; Computerized decision support system; Electronic health records
Funding sources: This work was supported by the Netherlands Heart
Foundation (public-private study grant, number: #2018B006); and the Top
Sector Life Sciences and health (PPP allowance made available to
Netherlands Heart Foundation to stimulate public-private partnerships).
TD and JH acknowledge financial support from the Netherlands Organiza-
tion for Health Research and Development (VENI grant 91617050, and
TOP grant 91215058, respectively).
Competing interests: The authors declare no competing interests.
Author contributions: TD, KM, KG, and SN conceived of the presented
idea, in correspondence with earlier work by FA, MB and KG. SN, TD, JH,
and KG derived the models and analyzed the data. TD and JH verified the
analytical methods and R scripts. SN, KG, TD, and JH contributed to the
interpretation of the results. KG and SN wrote the initial version of the
manuscript. All authors contributed to subsequent revisions of the manu-
script and provided critical feedback.
Data availability statement: The data that support the findings of this
study are available from the UCC upon reasonable request (https://www.
umcutrecht.nl/en/Research/Strategic-themes/Circulatory-Health/Facilities/
UCC).
Author statement: Steven Nijman contributed to Conceptualization,
Methodology, Software, Investigation, Writing - Original draft, Writing e
Review & Editing, Visualization. Katrien Groenhof contributed to Concep-
tualization, Methodology, Investigation, Writing - Original draft, Writing e
Review & Editing, Visualization. Jeroen Hoogland contributed to Software,
Validation, Writing eReview & Editing. Michiel Bots contributed to Fund-
ing acquisition, Writing eReview & Editing, Project administration. Menno
Brandjes contributed to Writing eReview & Editing. John Jacobs contrib-
uted to Writing eReview & Editing. Folkert Asselbergs contributed to
Conceptualization, Writing eReview & Editing. Carl Moons contributed
to Conceptualization, Methodology, Writing eReview & Editing. Thomas
Debray contributed to Conceptualization, Methodology, Validation, Writing
eReview & Editing, Supervision.
1
Contributed equally.
* Corresponding author. Julius Center for Health Sciences and Primary
Care, University Medical Center Utrecht, Heidelberglaan 100, 3584 CX
Utrecht, the Netherlands. Tel.: þ31-(0)88-75 68012; fax: þ31(0)887568099.
E-mail address: s.w.j.nijman@umcutrecht.nl (S.W.J. Nijman).
https://doi.org/10.1016/j.jclinepi.2021.01.003
0895-4356/Ó2021 The Authors. Published by Elsevier Inc. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/
4.0/).
Journal of Clinical Epidemiology 134 (2021) 22e34
What is new?
Key findings
Multiple imputation approaches can be adapted
without much difficulty to allow for real-time
imputation of missing predictor variables.
Both conditional modeling imputation (CMI??)
and joint modeling imputation (JMI??) give more
accurate estimates of missing predictor values
when compared to mean imputation.
What this adds to what was known?
Imputation of missing predictor values does not
require ‘‘live’’ access to a source dataset. Simple
population characteristics (such as the mean and
covariance) can be used to generate imputations
that are tailored to a specific individual.
What is the implication, and what should change
now?
Real-time multiple imputations using either CMI
or JMI should be made available in clinical prac-
tice (e.g., via a computerized decision support sys-
tem) to support guideline-recommended use of
prediction models and to be more transparent about
uncertainty
When developing or validating a prediction model,
researchers should report the mean and covariance
of the study population, as this information can
directly be used to impute missing values in
routine care.
1. Introduction
In present-day medical practice, characterized by an ag-
ing population, multimorbidity, and high complexity of dis-
eases, attention has grown toward personalized medicine
aiming to administer the most applicable treatment to the
individual patient given their risk profile [1e3]. In cardio-
vascular disease management, guidelines advocate the use
of prediction models to assess the patients’ risk of devel-
oping a certain cardiovascular disease to guide treatment
decision making [1]. For integrating risk-guided care in
daily practice, technological solutions such as computer-
ized decision support systems (CDSS) are increasingly
developed [4,5]. Using predictor values directly extracted
from the electronic health record (EHR), CDSS can provide
an immediate risk assessment of each encountered patient
at a glance [6e8].
The use of prediction models in daily practice in an in-
dividual patient requires real-time availability of the pa-
tient’s values of the predictors in the model. Most
prediction models cannot provide a risk estimate in the
presence of missing predictor values, which hampers im-
plementation and may ultimately limit guideline adherence
[9]. Therefore, predictor values should be measured and
registered (e.g., in the Electronic Health Record; EHR) in
such a way that they are available in real-time. Yet, routine
clinical care data is often incomplete because certain mea-
surements are deemed unnecessary, time-consuming, or
expensive, or because they cannot directly be extracted
from the EHR (e.g., registered as free text) [10].
Missing data is a well-known challenge in (medical)
research, for which several scalable solutions exist [11].
Multiple imputations by chained equations has often been
recommended to handle missing data in a research setting
where data from multiple patients are available for study
analysis purposes [12,13]. This approach, however, is not
directly applicable when applying a prediction model in
real-time to a single patient in the consulting room. In
particular, the models used for imputation cannot be gener-
ated ‘‘live’’ in clinical practice, and therefore, need to be
derived elsewhere and beforehand [14].
One option is to replace missing predictor values by their
respective mean/median, which, in turn, is estimated from
another data set or training sample [15,16]. While straight-
forward to implement, mean imputation may be insufficient
when the predictor with missing values is a strong predictor
or exhibits large variability such that assigning an overall
mean may lead to the less predictive accuracy of the predic-
tion model and misinformed treatment decisions. Mean
imputation does not distinguish between patients and may,
therefore, likely impute values that are unrealistic given the
patient’s observed predictor values. Also, mean imputation
obfuscates any uncertainty about the imputed values.
To address these issues, we expand on two well-known
methods that may also be used in real-time imputation of
missing predictor values [14]: joint modeling imputation
(JMI) [17] and conditional modeling imputation (CMI,
known for its common use in multiple imputations by
chained equations) [13]. As opposed to mean imputation,
these methods are able to incorporate the relationship be-
tween multiple patient characteristics, and therefore, allow
imputations to be adjusted for observed patient specific char-
acteristics. Similar to mean imputation, these relations can be
learned from training data, and in real time, applied on new
patients that are not part of the training sample. Additionally,
both methods allow for multiple imputations to be estimated,
reflecting the uncertainty with respect to the imputed value.
Using a real-world example and empirical data set on
cardiovascular risk prediction, we compared the accuracy
and usability of three imputation methods (mean imputa-
tion, JMI, and CMI) to deal with missing values of predic-
tors in the prediction model in real time. Although mean
imputation has been known to be problematic during model
development, it was chosen as a comparison due to its com-
mon use during model application in routine clinical prac-
tice or in decision support [18e21].
23S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
2. Methods
2.1. Imputation methods
For facilitating the live imputation of missing values in
routine care, it is essential to obtain information on the dis-
tribution of the target population. This summary informa-
tion can, for instance, be derived in an epidemiologic
(e.g., cohort) study and then be utilized for training live
imputation models. A key constraint given is that after be-
ing trained, all methods are independent and stand-alone,
which means that they can directly be used for live imputa-
tion in a new, single, patient without requiring the need for
any additional procedures.
The three methods under evaluation are mean imputa-
tion, joint modeling imputation (JMI), and conditional
modeling imputation (CMI) [13,14,17]. All methods were
implemented in R and facilitate live imputation of missing
values in individual patients. Source code is available from
the supplementary information (Appendix D).
2.1.1. Mean imputation
The training sample is used to derive the means of all
predictors in the model (Fig. 1). Missing predictor values
are then imputed by their respective mean (or proportion
in the case of binary variables). This method is relatively
straightforward to implement, and can be extended to
subgroup-specific means (i.e., creating subdivisions based
on certain parameters of a population of which multiple
means are calculated).
2.1.2. Joint modeling imputation
The training sample is used to derive the means and
covariance of all predictor variables (Fig. 2). It is assumed
that all predictor variables of the training sample are nor-
mally distributed, such that imputations for an individual
patient can directly be generated from the mean and covari-
ance of the training sample and the observed predictor
values [14,17]. In contrast to overall mean imputation, the
use of covariances between all predictors incorporates the
relation between the predictors, and therefore, allows impu-
tations to be tailored to an individual patient’s own charac-
teristics. A more detailed description is provided in
Appendix A [14].
2.1.3. Conditional modeling imputation
The training sample is used to derive a flexible (e.g.,
regression) model for each predictor (as dependent vari-
able) with all other predictor variables as independent vari-
ables (Fig. 3). These models describe the conditional
distribution of each predictor and usually need to be esti-
mated using a Gibbs sampling procedure (as predictor
values may also be missing in the training sample). Due
to the flexible nature of these conditional models, it is no
longer assumed that predictor variables of the training sam-
ple are normally distributed (as does JMI). For instance, a
logistic regression model can be used to estimate the con-
ditional distribution of a binary predictor variable (e.g., cur-
rent smoker). Subsequently, when the smoking status for a
new patient is unknown, the logistic regression model can
be used to generate a probability that they are a current
smoker. This probability can directly be used as an imputed
value (in case only 1 imputation is needed). Alternatively, if
multiple imputations are required, a Bernoulli distribution
(with the aforementioned probability) can be used to sam-
ple multiple (discrete) values for the patient’s current smok-
ing status. If multiple predictor values are missing, the
conditional models need to be used successively using an
iterative Monte Carlo procedure (Appendix A).
2.2. Simulation study
Cardiovascular disease prevention is an example of a
setting where risk-guided management of predictorsd
smoking, blood pressure, cholesteroldis common practice
Fig. 1. Mean imputation.
24 S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
[22]. Numerous risk prediction models have been devel-
oped, and the (international) guidelines advocate the use
of risk classification to inform treatment decisions
[23,24]. These models are typically implemented in a
CDSS, where a patient’s characteristics of the predictors
can be entered manually or are automatically retrieved from
the patient’s EHR [4,6,25].
For this study, we used a data set of the ongoing Utrecht
Cardiovascular cohort initiative (UCC). This cohort in-
cludes all patients who come for a first-time visit to the
Center for Circulatory Health at the UMC Utrecht for the
evaluation of a symptomatic vascular disease or an asymp-
tomatic vascular condition. A minimum set of predictors,
according to the Dutch Cardiovascular Risk Management
Guidelines, is collected in all patients. No data on outcomes
(i.e., time-to-event data) was recorded. UCC has been
approved by the Institutional Review Board of the UMC
Utrecht (Biobank Ethics committee). For the present ana-
lyses, an anonymized dataset was used of the UCC cohort
up to November 2018 [26,22].
The sample consisted of 3,880 patients with information
on 23 variables, measured during the patient’s visit (Table 1
and Appendix B). For clarity of exposition, we completed
this dataset using all 23 variables in k-nearest neighbor
imputation, which aggregates the values of the knearest
neighbors to an imputation [27]. In practice, regular multi-
ple imputation techniques can be used in case of incomplete
training data.
For evaluating the quality of the three selected imputa-
tion methods in individual patients, a leave-one-out-cross-
validation (LOOCV) procedure was used in the completed
UCC dataset. In LOOCV, all but one patient are used as the
training sample from which the overall mean or proportion
(method 1) or imputation models (methods 2 and 3) are
derived (Fig. 4). In the remaining hold-out patient, missing
values are introduced for one or more predictor variables.
As we apply each scenario to each patient exactly once,
the missing data mechanism is essentially missing-
completely-at-random (MCAR) [18]. The summary infor-
mation from the training sample is then used to impute
the missing predictor values in the hold-out patient. For
CMI and JMI, we generated 50 imputations for each
missing predictor value. This process is repeated until all
patients have been taken from the dataset exactly once.
We consider eight scenarios where missing values occur
for one predictor variable, and eight scenarios where multiple
predictor variables are simultaneously missing (Fig. 5). A
detailed description of how the scenarios were selected and
of the R code are listed in Appendix C and D, respectively.
2.3. Measures of performance
To evaluate the performance of the three imputation
methods, we used four performance metrics:
1. We calculated the root-mean-squared error (RMSE)
between the average of the multiple imputed
Fig. 2. Joint modeling imputation.
25S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
predictor values (i.e., 50 imputations) and the true,
original (i.e., before the simulation of missing) pre-
dictor value to evaluate the accuracy of the imputa-
tions. The RMSE is a performance measure that
aggregates error due to bias and variability. Generally,
an RMSE of zero means perfect imputation and an
increasing RMSE means decreasing performance of
the imputation. The clinical relevance of an RMSE
depends on the natural range of the predictor. For
example, an RMSE of 0.5 is large for LDL-c (mean
3.0 SD 1.3 mmol/L) but not for SBP (mean 143 SD
24 mmHg).
2. For each hold-out patient, we assessed whether the
original predictor value was in the 95% confidence in-
terval around the imputed predictor value. Subse-
quently, we calculated the proportion of confidence
intervals that consisted the original value (coverage).
For a 95% CI, the coverage should ideally be equal to
95% [28]. A lower coverage translates to imputed
predictor values that are too precise (which, in turn,
Fig. 3. Conditional modeling imputation.
26 S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
may lead to estimates of predicted risk that are too
precise), whereas a coverage above 95% indicates
that imputed predictor values are too imprecise
[13]. We assessed coverage only for continuous pre-
dictor variables.
3. We assessed the effect on treatment decision support
for blood pressure in patients with manifest cardio-
vascular disease (n51,971) to evaluate the clinical
implications of the imputed predictor values. Guide-
lines indicate that all patients with a history of
CVD should receive blood pressure-lowering treat-
ment when their blood pressure is higher than 140/
90 mmHg [1,22]. We adopted the LOOCV approach
and set values for SBP missing in the hold-out pa-
tient. Subsequently, we imputed the missing value
and compared the treatment decision for the true
value with the treatment decision for the imputed
value (SBP !O 140 mmHg). Afterward, we calcu-
lated the sensitivity, specificity, positive predictive
value, and negative predictive value. Also, we illus-
trated the importance of reporting confidence inter-
vals based on imputed values to inform the
discussion around treatment commencement.
4. We compared the risk predictions that were obtained
in the absence of missing values (i.e., in the original
data) with the risk predictions that are based on impu-
tations to evaluate the impact of the imputed values
on the precision of predicted risk. Ideally, the predic-
tions that are based on imputed values should have a
similar distribution as the predictions that are derived
from the complete original data. To explore any devi-
ation, we assessed the interquartile range of predicted
risk for a single missing predictor scenario and a mul-
tiple missing predictor scenario. Rather than devel-
oping a new prediction model ourselves, we applied
the previously developed SMART prediction model
for the risk of 10-year recurrent vascular disease as
reported in the original development study [23][.
Table 1. Descriptive statistics (after imputation)
Variables (unit) Part of missing data scenarios Mean (SD) or n/total (%)
a
Original missing %
Age (yr) No 61.7 (18.2) 0.00
Sex (1 5female; 0 5male) No 1,987/3,880 (51.2) 0.00
Smoking (1 5yes; 0 5no) No 363/3,880 (9.4) 24.07
SBP (mmHg) Yes 142.8 (24.2) 10.54
TC (mmol/L) Yes 5.1 (1.2) 24.54
LDL-c (mmol/L) Yes 3.1 (1.3) 26.01
HDL-c (mmol/L) No 1.4 (0.4) 25.39
eGFR (mL/min/1.73 m
2
) Yes 81.8 (24.6) 15.98
History of CVD (1 5yes; 0 5no) Yes 1,971/3,880 (50.8) 23.45
History of PAD (1 5yes; 0 5no) No 335/3,880 (8.6) 23.45
History of CHD (1 5yes; 0 5no) No 591/3,880 (15.2) 23.45
History of CHF (1 5yes; 0 5no) No 284/3,880 (7.3) 23.45
History of CVA (1 5yes; 0 5no) No 579/3,880 (14.9) 23.45
History of DM (1 5yes; 0 5no) No 607/3,880 (15.6) 23.45
Polyvascular disease No 0.6 (0.7) 23.45
Number of medications No 0.8 (1.7) 27.24
BP lowering medication (1 5yes; 0 5no) No 705/3,880 (18.2) 27.24
Statin (1 5yes; 0 5no) No 415/3,880 (10.7) 27.24
HbA1c (mmol/mol) No 40 (10.7) 26.37
Years after first CVD (yr) Yes 4.6 (8.1) 26.21
Diabetes (1 5yes; 0 5no) Yes 755/3,880 (19.5) 8.12
Diabetes duration (yr) No 11.3 (7.3) 86.11
Pulse pressure (mmHg) No 61.7 (18.9) 10.54
Abbreviations: SBP, systolic blood pressure; TC, total cholesterol; LDL-c, low-density lipoprotein cholesterol; HDL-c, high-density lipoprotein
cholesterol; eGFR, estimated glomerular filtration rate according to the CKD epi formula; CVD, cardiovascular disease; PAD, peripheral artery dis-
ease; CHD, coronary heart disease; CHF, chronic heart failure; CVA, cerebrovascular accident; DM, diabetes mellitus; BP, blood pressure; HbA1c,
glycated hemoglobin.
a
After KNN-imputation.
27S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
The prediction model includes 11 variables: age, sex,
current smoker, SBP, diabetes, history of cerebrovas-
cular disease, aortic aneurysm or peripheral vascular
disease, polyvascular disease, HDL-cholesterol, and
total cholesterol.
3. Results
3.1. Root-mean-squared error
With the exception of smoking, all predictor variables in
single missing predictor scenarios had a lower RMSE when
using JMI or CMI as compared to mean imputation
(Table 2). For most multiple missing predictor scenarios,
the RMSE is consistently lower when using JMI or CMI
as compared to mean imputation. The exceptions being
the history of CVD and smoking. Performance diminished
as more variables were missing. For example, the RMSEs
of years after the 1st CVD event are 6.30 and 6.26 for
JMI and CMI respectively when univariately missing, while
mean imputation has an RMSE of 8.06. When additional
variables (e.g., SBP, history of CVD, and smoking) are
missing, the RMSE for years after the 1st CVD event for
both JMI and CMI increases to 7.58 and 7.84, respectively.
3.2. Coverage rate
For JMI, the coverage reached nominal levels for all sin-
gle missing predictor scenarios and multiple missing pre-
dictor scenarios (Table 3). For CMI, the coverage reached
nominal levels for all single missing predictor scenarios
and multiple missing predictor scenarios. For mean imputa-
tion, coverage was 0% by definition for all imputed predic-
tors because no uncertainty is taken into account.
3.3. Clinical decision accuracy
When assessing the treatment decision for blood pres-
sure management according to the prevailing clinical
guidelines (see above), we selected 1,971 out of the total
3,880 patients with manifest cardiovascular disease. We
found that 1,134 patients (57.53%) should be treated. How-
ever, when blood pressure values were set to miss, the
Fig. 4. Missing data simulation procedure.
28 S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
overall mean imputed value was 142 mmHg (Table 1),
which is just above the treatment threshold of 140 mmHg.
As a result, everyone would have been treated when adopt-
ing overall mean imputation, such that 837 patients
(42.47%) would have been treated unnecessarily. When
adopting JMI or CMI, only 16.08% or, respectively,
15.98% of patients would have been treated unnecessarily
(Table 4). Hence, the imputation of missing blood pressure
values using CMI or JMI was more than adequate than the
mean imputation in terms of decision making.
To illustrate the importance of measuring uncertainty,
we provided an example in which we compare the use of
imputation in a real-life situation (Table 5). In the example,
a patient with an imputed SBP of 144 mmHg was given an
indication for blood pressure-lowering treatment according
to the guidelines [1]. However, given that the uncertainty
around the imputed predictor value crosses the treatment
line of 140 mmHG (scenario A), there is reasonable doubt
this imputation is too uncertain to be used for treatment de-
cision making.
3.4. Effect on risk predictions
The predicted risks, given each method, did not seem to
deviate much from the originally predicted risk, given the
complete data (Table 6). When assessing the single missing
predictor scenario, there was a difference between overall
mean imputation (median difference of 1.713% to the
originally predicted risk) and the combination of JMI and
CMI (median difference of respectively 0.301% and
0.399% to the originally predicted risk). Further, we found
that predicted risks for mean imputation were more similar
when compared to the complete data (standard devia-
tion 515.12 vs. the reference of 18.91). In contrast, the
standard deviations of JMI and CMI were 17.87 and
17.86, respectively.
In the multiple missing predictor scenario, there was a
similar difference between mean imputation (median differ-
ence of 2.064% to the originally predicted risk) and JMI
and CMI (median difference of respectively 0.375% and
0.390% to the originally predicted risk). With multiple
missing predictors, the predicted risks for mean imputation
were again more similar than the predicted risk given the
complete data (standard deviation 514.42 vs. the reference
of 18.91). The standard deviations of JMI and CMI were
17.67 and 17.68, respectively.
The difference between mean imputation and both JMI
and CMI is especially apparent in high-risk patients (i.e.,
75% IQR) where mean imputation, as expected, underesti-
mates the risk. This is because mean imputation pulls the
risk predictions of patients with missing values toward
the prediction for an ‘‘average’’ patient. As such, JMI and
CMI perform much better with regards to their impact on
prediction in high-risk patients when compared to mean
imputation.
4. Discussion
This project described the development and performance
of three imputation methods to handle missing data on an
individual patient level in real-life clinical decision making.
Fig. 5. Multivariable missing data scenarios.
29S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
As expected, both JMIeusing draws from a normal distri-
bution constructed from means and covariance in the
training sample and observed values in the patienteand
CMIeusing a conditional distribution of each variable
based on regression models fitted on all other variables,
were more accurate and showed better coverage as
compared to mean imputation, resulting in fewer inappro-
priate treatment decisions and lower impact on predicted
risk.
The accuracy measureseRMSE, coverage, and clinical
decision accuracyewere comparable for JMI and CMI.
Hence, both methods can be used for generating live impu-
tations in routine care. Based on usability, we recommend
JMI, as its implementation in decision support systems is
fairly straightforward and only requires information on
the mean and covariance of the target population. Although
its assumption of multivariate normality may be unrealistic
for real-life clinical data, simulation studies have demon-
strated that this rarely affects the performance of imputa-
tion [29e31].
Previous studies on imputation methods to handle
missing data on an individual patient level have focused
on the impact of missing values on the performance of a
prediction model and evaluated the use of mean imputation,
as well as the (re)development of a simplified prediction
model [15,16]. Mean imputation was recommended due
to its applicability in practice and relatively good perfor-
mance compared to other models but was considered insuf-
ficient when strong predictors were missing. For this
reason, our proposed multiple imputation models appear
particularly relevant when strong or multiple predictors
are missing. This was confirmed in our simulation study:
RMSE and coverage did not deteriorate much with the
increasing number of predictor values that were
Table 2. RMSE for each combination of, individual or multiple, missing predictor values
Single missing variable scenarios
Variable name
(type of data)
Diabetes
(binary)
SBP
(continuous)
EGFR
(continuous)
History of CVD
(binary)
Years after 1st
CVD
(continuous)
Smoking
(binary)
Total
cholesterol
(continuous)
HDL-
cholesterol
(continuous)
Mean imputation
a
0.40 24.24 24.56 0.50 8.06 0.29 1.24 0.36
JMI 0.17 22.31 19.60 0.39 6.30 0.30 1.19 0.34
CMI 0.21 22.29 19.69 0.39 6.26 0.29 1.19 0.34
Multiple missing variables scenarios
Method Scenario Diabetes SBP eGFR History of
CVD
Years after
1st CVD
Smoking Total
cholesterol
HDL-
cholesterol
JMI 1 0.46 7.59 0.30
CMI 1 0.51 7.78 0.29
JMI 2 19.68 1.19 0.35
CMI 2 19.69 1.20 0.35
JMI 3 1.19 0.33
CMI 3 1.19 0.35
JMI 4 0.17 0.48 7.65 0.30 1.22 0.35
CMI 4 0.20 0.50 7.83 0.28 1.21 0.35
JMI 5 0.17 22.62 19.86 0.47 7.66 0.30 1.23 0.35
CMI 5 0.21 22.48 19.87 0.51 7.86 0.29 1.22 0.35
JMI 6 22.45 19.61 1.19 0.34
CMI 6 22.50 19.59 1.20 0.34
JMI 7 0.17 19.83 0.48 7.69 0.30 1.22 0.35
CMI 7 0.21 19.75 0.50 7.84 0.29 1.23 0.35
JMI 8 22.36 0.46 7.58 0.30
CMI 8 22.35 0.51 7.84 0.29
Abbreviations: JMI, joint modeling imputation; CMI, conditional modeling imputation; SBP, systolic blood pressure; eGFR, estimated glomer-
ular filtration rate according to the CKD epi formula; CVD, cardiovascular disease.
The RMSE should ideally be 0. Multiple missing predictor scenarios: (1) history of CVD, years after 1st CVD event & smoking, (2) eGFR, total
cholesterol & HDL-cholesterol, (3) total cholesterol & HDL-cholesterol, (4) all variables but SBP & eGFR, (5) all variables, (6) SBP, eGFR, total
cholesterol & HDL-cholesterol, (7) all variables but SBP and (8) SBP, history of CVD, years after 1st CVD event and smoking.
a
Mean imputation is only included in the single missing variable scenarios, as the performance of the model when multiple variables are
missing, is equivalent.
30 S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
simultaneously missing for the individual patient. Because
of the way missing data was introduced, it is noted that our
simulations were not able to distinguish between various
mechanisms by which data can be missing, for example,
data that is missing at random (MAR) vs. data that is
missing-completely-at-random (MCAR??) [18].
Furthermore, because the described imputation methods
can accommodate numerous patient characteristics that are
not necessarily disease-specific, they are highly scalable to
other settings and populations. However, it is likely that
some local tailoring is necessary when imputation models
are derived from specific studies or settings that do not fully
match the intended target population. For JMI, the means
and covariances could, for instance, simply be replaced
by their respective values in a local ‘‘training’’ sample.
For CMI, the regression coefficients can be revised using
recently described updating methods [32]. When the (local)
training data are affected by missing predictor values,
advanced methods exist to estimate the mean and the
covariance [33]. All methods can be potentially incorpo-
rated within an EHR based computerized decision support
system and generate imputations based on observed data
from individual patients extracted from the EHR.
Evidently, before implementing imputation models in clin-
ical practice, it is of the utmost importance to assess their
validity, likely impact on treatment decisions, patient out-
comes, as well as any practical, security, and ethical
constraints.
Although multiple imputations offer a computational
framework to account for missing values, we always
recommend optimizing data collection first and avoid hav-
ing missing values: clinical decision making should never
be based solely on imputed values. However, imputed
values can serve as a proxy for prior risk, setting an indica-
tion for more (advanced) diagnostic tests. This is especially
useful for expensive tests, tests associated with complica-
tions, or when tests are unavailable. Additional diagnostic
testing should preferably only be performed when it is ex-
pected to change treatment, and the potential clinical
benefit outweighs the risk of the tests). Note that in this
study, we do not take into account the (un)certainty around
imputed values when assessing treatment decision support.
Additionally, due to the limited data at our disposal, a full
evaluation of the impact on predicted risk was not possible.
Table 3. Coverage for each combination of individual or multiple imputations
Coverage: Single missing variable scenarios
SBP eGFR Years after1st CVD Total cholesterol HDL-cholesterol
JMI 0.945 0.948 0.952 0.952 0.950
CMI 0.945 0.948 0.954 0.953 0.948
Coverage: Multiple missing variable scenarios
Method Scenario SBP eGFR Years after1st CVD Total cholesterol HDL-cholesterol
JMI 1 0.951
CMI 1 0.948
JMI 2 0.947 0.951 0.951
CMI 2 0.946 0.955 0.949
JMI 3 0.949 0.951
CMI 3 0.950 0.949
JMI 4 0.951 0.950 0.951
CMI 4 0.949 0.952 0.952
JMI 5 0.944 0.947 0.951 0.952 0.951
CMI 5 0.948 0.948 0.946 0.953 0.953
JMI 6 0.945 0.950 0.951 0.948
CMI 6 0.948 0.948 0.949 0.949
JMI 7 0.950 0.951 0.951 0.951
CMI 7 0.947 0.950 0.948 0.951
JMI 8 0.945 0.952
CMI 8 0.945 0.950
Abbreviations: JMI, joint modeling imputation; CMI, conditional modeling imputation; SBP, systolic blood pressure; eGFR, estimated glomer-
ular filtration rate according to the CKD epi formula; CVD, cardiovascular disease.
The presented values depict the coverage of 95% confidence intervals (hence, the reference value is 0.95). Multiple missing predictor sce-
narios: (1) history of CVD, years after 1st CVD event & smoking, (2) eGFR, total cholesterol & HDL-cholesterol, (3) total cholesterol & HDL-
cholesterol, (4) all variables but SBP & eGFR, (5) all variables, (6) SBP, eGFR, total cholesterol & HDL-cholesterol, (7) all variables but SBP
and (8) SBP, history of CVD, years after 1st CVD event and smoking.
31S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
Ideally, uncertainty in imputed values should be propagated
to (additional) uncertainty in predicted risk and evaluated
with presented confidence intervals. The predicted risk in
this paper primarily serves as a way to illustrate how impu-
tations could influence the predicted risk.
In cardiovascular risk management, the decision to start
treatment of a risk factor is based on (i) the predicted risk
for a cardiovascular disease or patient characteristics that
are per definition associated with a high risk for cardiovas-
cular disease and (ii) the absolute value of the risk factor
itself. We focused on imputation models to recover the
missing value and to quantify its uncertainty. We demon-
strated that the choice of imputation method might impact
risk predictions and decision making. While the magnitude
of this effect was not always substantial, it may vary ac-
cording to the number of missing predictors and their
weight in the decision-making process and should, there-
fore, be evaluated when applying these models in different
settings and populations.
Last, traditional (e.g., regression-based) prediction
models assume complete input data, which is often not real-
istic in routine clinical practice. Although we developed
models for imputing the missing values, which can subse-
quently be used to generate predictions, it is also possible
to develop prediction models that do not require complete
information on the predictors. Well-known examples are
the use of decision trees with surrogate or sparsity-aware
splits [34e36], the use of submodels [37], or the use of
missing indicator variables [38]. More research is war-
ranted to evaluate whether these methods may offer any
improvement in model predictions, as well as facilitate
their implementation in routine care.
In summary, this study describes three imputation
methods to handle missing values in the context of comput-
erized decision support systems in clinical practice. We
found that JMI and CMI provide imputations that are closer
to the original value (as compared to mean imputation) and
able to reflect uncertainty due to missing data. We,
Table 4. 22 tables of guideline adherence to treatment threshold given the point estimate of each method
Mean imputation
True value
TotalsTreatment advised (140 mmHg) Treatment not advised (!140 mmHg)
Point estimate
Treatment advised (O140 mmHg) 1,134 837 1,971
Treatment not advised (!140 mmHg) 0 0 0
Totals 1,134 837 1,971
Joint modeling imputation
True value
TotalsTreatment advised (140 mmHg) Treatment not advised (!140 mmHg)
Point estimate
Treatment advised (O140 mmHg) 946 317 1,263
Treatment not advised (!140 mmHg) 188 520 708
Totals 1,134 837 1,971
Conditional modeling imputation
True value
TotalsTreatment advised (140 mmHg) Treatment not advised (!140 mmHg)
Point estimate
Treatment advised (O140 mmHg) 960 315 1,275
Treatment not advised (!140 mmHg) 174 522 696
Totals 1,134 837 1,971
Sensitivity 100%, specificity 0%, Positive Predictive Value 58%, Negative Predictive Value (cannot be calculated) %.
Sensitivity 83%, specificity 62%, Positive Predictive Value 75%, Negative Predictive Value 73%.
Sensitivity 85%, specificity 62%, Positive Predictive Value 75%, Negative Predictive Value 75%.
Table 5. Clinical interpretation of imputed SBP values and 95% confidence intervals from a patient with a history of CVD
True Scenario A Scenario B
SBP (95%CI) 144 144 (138e150) 144 (142e146)
Treatment based on point estimate O140 mmHg, Start treatment O140 mmHg, Start treatment O140 mmHg, Start treatment
Treatment based on 95% CI NA Uncertain O140 mmHg, Start treatment
Abbreviations: SBP, systolic blood pressure; 95% CI, 95% confidence interval; A, hypothetical situation where imputed value interval contains
treatment threshold; B, hypothetical situation where imputed value interval does not contain treatment threshold.
32 S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
therefore, recommend their implementation in situations
where information on relevant predictors is often incom-
plete due to practical constraints.
Acknowledgments
This study was conducted on behalf of the Utrecht Car-
diovascular Cohort- Cardiovascular Risk Management
(UCC- CVRM) study group. Members of the UCC- CVRM
Study group: F.W. Asselbergs, Department of Cardiology;
G.J. de Borst, Department of Vascular Surgery; M.L. Bots
(chair), Julius Center for Health Sciences and Primary
Care; S. Dieleman, Division of Vital Functions (anesthesi-
ology and intensive care); M.H. Emmelot, Department of
Geriatrics; P.A. de Jong, Department of Radiology; A.T.
Lely, Department of Obstetrics/Gynecology; I.E. Hoefer,
Laboratory of Clinical Chemistry and Hematology; N.P.
van der Kaaij, Department of Cardiothoracic Surgery;
Y.M. Ruigrok, Department of Neurology; M.C. Verhaar,
Department of Nephrology & Hypertension, F.L.J. Visse-
ren, Department of Vascular Medicine, University Medical
Center Utrecht and Utrecht University.
Supplementary data
Supplementary data to this article can be found online at
https://doi.org/10.1016/j.jclinepi.2021.01.003. References
[39,40] are cited in the appendix section.
References
[1] Piepoli MF, Hoes AW, Agewall S, Albus C, Brotons C, Catapano AL,
et al. 2016 European guidelines on cardiovascular disease prevention
in clinical practice: the sixth joint task force of the European society
of Cardiology and other societies on cardiovascular disease preven-
tion in clinical practice (constituted by representatives of 10 societies
and by invited experts) developed with the special contribution of the
European association for cardiovascular prevention & rehabilitation
(EACPR). Eur Heart J 2016;37:2315e81.
[2] Hoffman MA, Williams MS. Electronic medical records and person-
alized medicine. Hum Genet 2011;130(1):33e9.
[3] Ginsburg G. Personalized medicine: revolutionizing drug discovery
and patient care. Trends Biotechnol 2001;19(12):491e6.
[4] Groenhof TKJ, Groenwold RHH, Grobbee DE, Visseren FLJ,
Bots ML, On Behalf of the UCC-SMART Study Group. The effect
of computerized decision support systems on cardiovascular risk fac-
tors: a systematic review and meta-analysis. BMC Med Inform Decis
Mak 2019;19(1):108.
[5] Bezemer T, de Groot MC, Blasse E, ten Berg MJ, Kappen TH,
Bredenoord AL, et al. A Human(e) factor in clinical decision support
systems. J Med Internet Res 2019;21(3):e11732.
[6] Groenhof TKJ, Bots ML, Brandjes M, Jacobs JJL, Grobbee DE, van
Solinge WW, et al. A computerised decision support system for car-
diovascular risk management ‘live’ in the electronic health record
environment: development, validation and implementationdthe
Utrecht Cardiovascular Cohort Initiative. Neth Heart J 2019;27(9):
435e42.
[7] Kawamoto K, Houlihan CA, Balas EA, Lobach DF. Improving clin-
ical practice using clinical decision support systems: a systematic re-
view of trials to identify features critical to success. BMJ 2005;330:
765.
[8] Sitapati A, Kim H, Berkovich B, Marmor R, Singh S, El-Kareh R,
et al. Integrated precision medicine: the role of electronic health re-
cords in delivering personalized treatment: Integrated precision med-
icine. Wiley Interdiscip Rev Syst Biol Med 2017;9(3):e1378.
Table 6. Differences in predicted 10-yr risk of CVD for both a single missing predictor scenario and a multiple missing predictor scenario
Single missing
predictor: eGFR 25% IQR
Absolute risk
difference to
completed data Median
Absolute risk
difference to
completed data 75% IQR
Absolute risk
difference to
completed data
Predicted risk
complete data
8.382% e13.711% e28.170% e
Predicted risk
(mean)
7.287% 1.095% 11.997% 1.713% 23.035% 5.135%
Predicted risk (joint) 8.767% 0.385% 14.012% 0.301% 27.734% 0.435%
Predicted risk
(conditional)
8.786% 0.404% 14.110% 0.399% 27.783% 0.387%
Multiple missing
predictors:
SBP, TC, LDL-c
and eGFR
Predicted risk
complete data
8.382% e13.711% e28.170% e
Predicted risk
(mean)
7.473% 0.909% 11.647% 2.064% 22.692% 5.478%
Predicted risk
(joint)
8.809% 0.427% 14.085% 0.375% 28.410% 0.240%
Predicted risk
(conditional)
8.786% 0.404% 14.100% 0.390% 28.267% 0.097%
33S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
[9] Kotseva K, Wood D, De Bacquer D, De Backer G, Ryd
en L,
Jennings C, et al. Euroaspire IV: a European Society of Cardiology
survey on the lifestyle, risk factor and therapeutic management of
coronary patients from 24 European countries. Eur J Prev Cardiol
2016;23(6):636e48.
[10] Rumsfeld JS, Joynt KE, Maddox TM. Big data analytics to improve
cardiovascular care: promise and challenges. Nat Rev Cardiol 2016;
13(6):350e9.
[11] Perkins NJ, Cole SR, Harel O, Tchetgen Tchetgen EJ, Sun B,
Mitchell EM, et al. Principled approaches to missing data in epidemi-
ologic studies. Am J Epidemiol 2018;187:568e75.
[12] Azur MJ, Stuart EA, Frangakis C, Leaf PJ. Multiple imputation by
chained equations: what is it and how does it work?: multiple impu-
tation by chained equations. Int J Methods Psychiatr Res 2011;20(1):
40e9.
[13] Van Buuren S. Flexible imputation of missing data. 2nd ed. Boca Ra-
ton, FL: CRC Press; 2018.
[14] Hoogland J, van Barreveld M, Debray TPA, Reitsma JB,
Verstraelen TE, Dijkgraaf MGW, et al. Handling missing predictor
values when validating and applying a prediction model to new pa-
tients. Stat Med 2020;17:3591e607.
[15] Berkelmans G. Dealing with missing patient characteristics when us-
ing cardiovascular prediction models in clinical practice. Oxford, En-
gland: European Heart Journal, Oxford University Press; 2018:
110e33. https://doi.org/10.1093/eurheartj/ehy565.P1533. Accessed
January 4, 2019.
[16] Janssen KJM, Vergouwe Y, Donders ART, Harrell FE, Chen Q, Grob-
bee DE, et al. Dealing with missing predictor values when applying
clinical prediction models. 8. Clinical chemistry, American Associa-
tion for Clinical Chemistry: Washington, DC. Available at https://doi.
org/10.1373/clinchem.2008.115345. Accessed 1 February 2019.
[17] Carpenter JR, Kenward MG. Multiple imputation and its application.
Chichester: Wiley; 2013:345. : (Statistics in practice).
[18] Donders ART, van der Heijden GJMG, Stijnen T, Moons KGM. Re-
view: a gentle introduction to imputation of missing values. J Clin
Epidemiol 2006;59:1087e91.
[19] G
okc¸ay D, Eken A, Baltacı S. Binary classification using neural and
clinical features: an application in fibromyalgia with likelihood based
decision level fusion. 10. IEEE Journal of Biomedical and Health
Informatics: Piscataway, NJ. Available at https://doi.org/10.1109/
JBHI.2018.2844300. Accessed 5 December 2020.
[20] Deb
edat J, Sokolovska N, Coupaye M, Panunzi S, Chakaroun R,
Genser L, et al. Long-term relapse of type 2 diabetes after roux-en-
Y Gastric Bypass: prediction and clinical relevance. Diabetes Care
2018;41:2086e95.
[21] Chen R, Stewart WF, Sun J, Ng K, Yan X. Recurrent neural networks
for early detection of heart failure from longitudinal electronic health
record data. Circ Cardiovasc Qual Outcomes 2019;15:e005114.
[22] Nederlands Huisartsen Genootschap. Multidisciplinaire richtlijn car-
diovasculair risicomanagement. Houten: Bohn Stafleu van Loghum;
2011.
[23] Dorresteijn JAN, Visseren FLJ, Wassink AMJ, Gondrie MJA,
Steyerberg EW, Ridker PM, et al. Development and validation of a
prediction rule for recurrent vascular events based on a cohort study
of patients with arterial disease: the SMART risk score. Heart 2013;
99:866e72.
[24] Damen JAAG, Hooft L, Schuit E, Debray TPA, Collins GS,
Tzoulaki I, et al. Prediction models for cardiovascular disease risk
in the general population: systematic review. BMJ 2016i2416.
[25] Visseren FLJ, Dorresteijn JAN, van der Graaf Y. U-prevent u bent ‘in
control’ [Internet] 2018. Available at https://www.u-prevent.nl/nl-NL.
Accessed October 7, 2019.
[26] Asselbergs FW, Visseren FL, Bots ML, de Borst GJ, Buijsrogge MP,
Dieleman JM, et al. Uniform data collection in routine clinical prac-
tice in cardiovascular patients for optimal care, quality control and
research: the Utrecht Cardiovascular Cohort. Eur J Prev Cardiol
2017;24(8):840e7.
[27] Kowarik A, Templ M. Imputation with the R Package VIM. J Stat
Softw 2016;74. Available at http://www.jstatsoft.org/v74/i07/. Ac-
cessed September 26, 2019.
[28] Burton A, Altman DG, Royston P, Holder RL. The design of simula-
tion studies in medical statistics. Stat Med 2006;25:4279e92.
[29] Hughes RA, White IR, Seaman SR, Carpenter JR, Tilling K,
Sterne JA. Joint modelling rationale for chained equations. BMC
Med Res Methodol 2014;14:28.
[30] Demirtas H, Hedeker D. An imputation strategy for incomplete lon-
gitudinal ordinal data. Stat Med 2008;27:4086e93.
[31] Murray JS. Multiple imputation: a review of practical and theoretical
findings. Arxiv180104058 Stat 2018. Available at http://arxiv.org/abs/
1801.04058. Accessed September 26, 2019.
[32] Vergouwe Y, Nieboer D, Oostenbrink R, Debray TPA, Murray GD,
Kattan MW, et al. A closed testing procedure to select an appropriate
method for updating prediction models: method selection to update a
prediction model. Stat Med 2017;36:4529e39.
[33] Varadhan R. condMVNorm: conditional multivariate normal distribu-
tion. 2015; R package version 2015.2-1. Available at http://CRAN.R-
project.org/package5condMVNorm. Accessed April 3, 2019.
[34] Hapfelmeier A. Analysis of missing data with random forests
[Internet]. 2012. Available at https://edoc.ub.uni-muenchen.de/
15058/1/Hapfelmeier_Alexander.pdf. Accessed September 4, 2019.
[35] Feelders A. Handling missing data in trees: surrogate splits or statis-
tical imputation?. In:
_
Zytkow JM, Rauch J, editors. Principles of data
mining and knowledge discovery [Internet]. Berlin, Heidelberg:
Springer Berlin Heidelberg; 1999:329e34. Available at http://link.
springer.com/10.1007/978-3-540-48247-5_38. Accessed October 2,
2019.
[36] Burgette LF, Reiter JP. Multiple imputation for missing data via
sequential regression trees. Am J Epidemiol 2010;172:1070e6.
[37] Hoogland J, van Barreveld M, Debray T, Reitsma J, Verstraelen T,
Dijkgraaf M, et al. Handling missing predictor values when vali-
dating and applying a prediction model to new patients. Stat Med
2019;39:3591e607. [Under review].
[38] Groenwold RHH, White IR, Donders ART, Carpenter JR,
Altman DG, Moons KGM. Missing covariate data in clinical
research: when and when not to use the missing-indicator method
for analysis. Can Med Assoc J 2012;184(11):1265e9.
[39] Chen M-H. Monte carlo methods in bayesian computation. Berlin,
Germany: Springer; 2013.
[40] Genz A, Bretz F, Miwa T, Mi X, Leisch F, Scheipl F, et al. Mvtnorm:
multivariate normal and t distributions. J Stat Softw 2018. Available
at http://CRAN.R-project.org/package5mvtnorm. Accessed April 3,
2019.
34 S.W.J. Nijman et al. / Journal of Clinical Epidemiology 134 (2021) 22e34
... Although using mean imputation (applicable under MCAR [55]) to deal with predictors missing data might worsen the subsequent prediction [56,57], we still include it for methods comparison since it is simply implemented and also commonly used. The mean imputation method is a single imputation method, denoted as SI for simplicity. ...
... Therefore, to ensure the reliability and validity of the prediction models in similar cases, we suggest that researchers deal with the ME of predictors by MI methods, like MI-ME or other calibration methods. Of course, we recommend optimizing data collection and avoiding missing values and ME initially because essential clinical decision-making needs supportable evidence induced from high-quality data [57]. In the scenario where tests are expensive, complicated, or even unavailable, high-quality data is difficult to obtain. ...
... In the scenario where tests are expensive, complicated, or even unavailable, high-quality data is difficult to obtain. In this case, statistical methods or ML algorithms to address missing data and ME are good tools for researchers to acquire information about the relationship of predictors-outcome as much as possible from the available data [57]. Additionally, the measurement of AL has been encouraged to be used as an important indicator of myopia prevention in myopia studies [10] since it is highly correlated with myopia progression [19] and its measurement is accurate and less timeconsuming than cycloplegic autorefraction [87]. ...
Article
Full-text available
Background Early identification of children at high risk of developing myopia is essential to prevent myopia progression by introducing timely interventions. However, missing data and measurement error (ME) are common challenges in risk prediction modelling that can introduce bias in myopia prediction. Methods We explore four imputation methods to address missing data and ME: single imputation (SI), multiple imputation under missing at random (MI-MAR), multiple imputation with calibration procedure (MI-ME), and multiple imputation under missing not at random (MI-MNAR). We compare four machine-learning models (Decision Tree, Naive Bayes, Random Forest, and Xgboost) and three statistical models (logistic regression, stepwise logistic regression, and least absolute shrinkage and selection operator logistic regression) in myopia risk prediction. We apply these models to the Shanghai Jinshan Myopia Cohort Study and also conduct a simulation study to investigate the impact of missing mechanisms, the degree of ME, and the importance of predictors on model performance. Model performance is evaluated using the receiver operating characteristic curve (AUROC) and the area under the precision-recall curve (AUPRC). Results Our findings indicate that in scenarios with missing data and ME, using MI-ME in combination with logistic regression yields the best prediction results. In scenarios without ME, employing MI-MAR to handle missing data outperforms SI regardless of the missing mechanisms. When ME has a greater impact on prediction than missing data, the relative advantage of MI-MAR diminishes, and MI-ME becomes more superior. Furthermore, our results demonstrate that statistical models exhibit better prediction performance than machine-learning models. Conclusion MI-ME emerges as a reliable method for handling missing data and ME in important predictors for early-onset myopia risk prediction.
... 14 In brief, JMI is suitable for individual predictions, does not need large amounts of raw data, and can achieve near-real time handling of missing data that makes it attractive for use in CDSS. [14][15][16] Joint modelling imputation uses a two-step approach: first, population characteristics (i.e. means and covariance) are estimated in raw individual patient data and stored in the EHR-system; and second, the CDSS handles any missing data by drawing imputations using the stored population characteristics. ...
... We used methods for performing JMI as developed in previous studies, though separately implemented in R to be used on clinical data. [14][15][16] In general, JMI assumes multivariate normality. Any imputations are randomly sampled from a (multivariate) normal distribution, constructed from means and covariance in a training sample, which is conditioned on the observed predictor values (i.e. ...
Article
Full-text available
Introduction A major challenge of the use of prediction models in clinical care is missing data. Real-time imputation may alleviate this. However, to what extent clinicians accept this solution remains unknown. We aimed to assess acceptance of real-time imputation for missing patient data in a clinical decision support system (CDSS) including 10-year cardiovascular absolute risk for the individual patient. Methods We performed a vignette study extending an existing CDSS with the real-time imputation method Joint Modelling Imputation (JMI). We included 17 clinicians to use the CDSS with three different vignettes, describing potential use cases (missing data, no risk estimate; imputed values, risk estimate based on imputed data; complete information). In each vignette missing data was introduced to mimic a situation as could occur in clinical practice. Acceptance of end-users was assessed on three different axes: clinical realism, comfortableness and added clinical value. Results Overall, the imputed predictor values were found to be clinically reasonable and according to the expectations. However, for binary variables, use of a probability scale to express uncertainty was deemed inconvenient. The perceived comfortableness with imputed risk prediction was low and confidence intervals were deemed too wide for reliable decision making. The clinicians acknowledged added value for using JMI in clinical practice when used for educational, research or informative purposes. Conclusion Handling missing data in CDSS via JMI is useful, but more accurate imputations are needed to generate comfort in clinicians for use in routine care. Only then CDSS can create clinical value by improving decision making.
... Depending on the method used to handle missing data, some patients may lack predictive points, leading to overestimations of patient numbers. If multiple imputation is employed, future information that should not be used at earlier points may be employed when handling missing data [72][73][74]. Figure 4 shows an example of how such situations can lead to overfitting and exaggerated patient numbers, attributable to data leakage during handling of time-series data. ...
Article
Acute kidney injury (AKI) is a significant health challenge associated with adverse patient outcomes and substantial economic burdens. Many authors have sought to prevent and predict AKI. Here, we comprehensively review recent advances in the use of artificial intelligence (AI) to predict AKI, and the associated challenges. Although AI may detect AKI early and predict prognosis, integration of AI-based systems into clinical practice remains challenging. It is difficult to identify AKI patients using retrospective data; information preprocessing and the limitations of existing models pose problems. It is essential to embrace standardized labeling criteria and to form international multi-institutional collaborations that foster high-quality data collection. Additionally, existing constraints on the deployment of evolving AI technologies in real-world healthcare settings and enhancement of the reliabilities of AI outputs are crucial. Such efforts will improve the clinical applicability, performance, and reliability of AKI Clinical Support Systems, ultimately enhancing patient prognoses.
... Another thirteen percent (i.e., 3/23) of studies employed hybrid imputation methods; for instance, study [44] used fuzzy principal component analysis (FPCA), SVM, and fuzzy c-means (FCM) clustering methods, while study [59] applied utility-based regression (UBR) with synthetic minority over-sampling technique for regression (SMOTER). Finally, thirteen percent (i.e., 3/23) of studies implemented deep learning-based imputation methods, including clinical condition generative adversarial network (CCGAN) [43,56] and partial multiple imputation with variational auto-encoders (PMI-VAE) [75]. In this category, 65% of the studies utilized simulation approaches to implement algorithms for data imputation. ...
Article
Full-text available
Background and objectives Comprehending the research dataset is crucial for obtaining reliable and valid outcomes. Health analysts must have a deep comprehension of the data being analyzed. This comprehension allows them to suggest practical solutions for handling missing data, in a clinical data source. Accurate handling of missing values is critical for producing precise estimates and making informed decisions, especially in crucial areas like clinical research. With data's increasing diversity and complexity, numerous scholars have developed a range of imputation techniques. To address this, we conducted a systematic review to introduce various imputation techniques based on tabular dataset characteristics, including the mechanism, pattern, and ratio of missingness, to identify the most appropriate imputation methods in the healthcare field. Materials and methods We searched four information databases namely PubMed, Web of Science, Scopus, and IEEE Xplore, for articles published up to September 20, 2023, that discussed imputation methods for addressing missing values in a clinically structured dataset. Our investigation of selected articles focused on four key aspects: the mechanism, pattern, ratio of missingness, and various imputation strategies. By synthesizing insights from these perspectives, we constructed an evidence map to recommend suitable imputation methods for handling missing values in a tabular dataset. Results Out of 2955 articles, 58 were included in the analysis. The findings from the development of the evidence map, based on the structure of the missing values and the types of imputation methods used in the extracted items from these studies, revealed that 45% of the studies employed conventional statistical methods, 31% utilized machine learning and deep learning methods, and 24% applied hybrid imputation techniques for handling missing values. Conclusion Considering the structure and characteristics of missing values in a clinical dataset is essential for choosing the most appropriate data imputation technique, especially within conventional statistical methods. Accurately estimating missing values to reflect reality enhances the likelihood of obtaining high-quality and reusable data, contributing significantly to precise medical decision-making processes. Performing this review study creates a guideline for choosing the most appropriate imputation methods in data preprocessing stages to perform analytical processes on structured clinical datasets.
... The difference was computed using χ 2 or Fisher's exact test for categorical variables, Student's t test for normally distributed continuous variables, or the Mann-Whitney U test for nonnormally distributed data. Mean imputation was used to address six missing PCT data points [17,18]. Least absolute shrinkage and selection operator (LASSO) regression (glmnet package) and logistic regression were used to construct the model. ...
Article
Full-text available
Background Vasovagal syncope (VVS) is the most common type of orthostatic intolerance in children. We investigated whether platelet-related factors related to treatment efficacy in children suffering from VVS treated with metoprolol. Methods Metoprolol-treated VVS patients were recruited. The median duration of therapy was three months. Patients were followed and divided into two groups, treament-effective group and treatment-ineffective group. Logistic and least absolute shrinkage selection operator regressions were used to examine treatment outcome variables. Receiver-operating characteristic (ROC) curves, precision–recall (PR) curves, calibration plots, and decision curve analyses were used to evaluate the nomogram model. Results Among the 72 patients who complete the follow-up, treatment-effective group and treatment-ineffective group included 42 (58.3%) and 30 (41.7%) cases, respectively. The patients in the treatment-effective group exhibited higher mean platelet volume (MPV) [(11.0 ± 1.0) fl vs. (9.8 ± 1.0) fl, P < 0.01] and platelet distribution width [12.7% (12.3%, 14.3%) vs. 11.3% (10.2%, 12.2%), P < 0.01] than those in the treatment-ineffective group. The sex ratio was significantly different ( P = 0.046). A fit model comprising age [odds ratio (OR) = 0.766, 95% confidence interval (CI) = 0.594–0.987] and MPV (OR = 5.613, 95% CI = 2.297–13.711) might predict therapeutic efficacy. The area under the curve of the ROC and PR curves was computed to be 0.85 and 0.9, respectively. The P value of the Hosmer–Lemeshow test was 0.27. The decision curve analysis confirmed that managing children with VVS based on the predictive model led to a net advantage ranging from 0.01 to 0.58. The nomogram is convenient for clinical applications. Conclusion A novel nomogram based on age and MPV can predict the therapeutic benefits of metoprolol in children with VVS.
... A given procedure or test can be described with different terms across systems and there can be inconsistencies in the meaning associated with a given term 13,23 . Different systems also collect different data, leading to missing predictor variables at deployment 24,25 . Even when these variables are available, they might be measured using different procedures, which has been shown to result in model biases 26 . ...
Preprint
Full-text available
In the period immediately following surgery, patients are at high risk of various negative outcomes such as Acute Kidney Injury (AKI) and Myocardial Infarction (MI). Identifying patients at increased risk of developing these complications assists in their prevention and management. During surgery, rich time series data of vital signs and ventilator parameters are collected. This data holds enormous potential for the prediction of postoperative outcomes. There is, however, minimal work exploring this potential. Moreover, existing approaches rely on deep learning, which is computationally expensive, often requiring specialized hardware and significant energy consumption. We demonstrate that it is possible to extract substantial value from intraoperative time series using techniques that are extremely computationally efficient. We used recordings from 66 300 procedures at the Lyell McEwin Hospital (Adelaide, South Australia), occurring in 2013 through 2020. The procedures associated with 80% of the patients were used for model training, with the remainder held out for testing. A combination of techniques including MultiRocket, Multitask and logistic regression were used to predict Rapid Response Team (RRT) calls within 48 hours of surgery and mortality, AKI and elevated troponin levels within 30 days of surgery. This approach achieved an Area Under the Receiver Operating Characteristic curve (AUROC) (95% CI) on the test data of 0.96 (0.95-0.97) for mortality, 0.85 (0.84-0.87) for AKI, 0.89 (0.87-0.91) for elevated troponin levels and 0.80 (0.78-0.83) for RRT calls, outperforming the ASA score and Charlson comorbidity index on the test population for all outcomes. These results show that roughly equivalent accuracy to computationally expensive modelling approaches using diverse sources of clinical data can be achieved using highly computationally efficient techniques and only a small set of automatically recorded intraoperative time series. This implies substantial potential in the analysis of these time series for the improvement of perioperative patient care. We also performed an analysis of the measurement sampling rate required to achieve these results, demonstrating the advantage of high-frequency patient vitals monitoring.
Article
Clinical prediction models have been widely acknowledged as informative tools providing evidence‐based support for clinical decision making. However, prediction models are often underused in clinical practice due to many reasons including missing information upon real‐time risk calculation in electronic health records (EHR) system. Existing literature to address this challenge focuses on statistical comparison of various approaches while overlooking the feasibility of their implementation in EHR. In this article, we propose a novel and feasible submodel approach to address this challenge for prediction models developed using the model approximation (also termed “preconditioning”) method. The proposed submodel coefficients are equivalent to the corresponding original prediction model coefficients plus a correction factor. Comprehensive simulations were conducted to assess the performance of the proposed method and compared with the existing “one‐step‐sweep” approach as well as the imputation approach. In general, the simulation results show the preconditioning‐based submodel approach is robust to various heterogeneity scenarios and is comparable to the imputation‐based approach, while the “one‐step‐sweep” approach is less robust under certain heterogeneity scenarios. The proposed method was applied to facilitate real‐time implementation of a prediction model to identify emergency department patients with acute heart failure who can be safely discharged home.
Article
Whereas several decades of research have documented behavior during couples' conflict discussions that is destructive to the couple relationship, the effectiveness of post‐conflict reconciliation efforts has only recently received research attention. The current study aimed to discover how couples' negative and positive communication during conflict discussions was associated with the helpfulness of two strategies for reconciling: Active Repair and Letting Go. The sample comprised 217 heterosexual cohabiting couples who completed (1) questionnaires assessing their relationship satisfaction and tendency to engage in demand/withdraw conflict management, (2) a recorded in‐lab conflict discussion, and (3) fourteen daily diaries reporting on relationship conflict and Active Repair and Letting Go reconciliation strategies. Daily negative conflict behavior (i.e., conflict expansion and contemptuous communication) did not predict helpfulness of reconciliatory attempts; nor did habitual negative conflict behavior (i.e., demand/withdraw). However, daily positive conflict behavior (i.e., attentive listening) did predict more helpful reconciliatory attempts. Observers' one‐time ratings of couples' communication competence during conflict discussions revealed an unexpected inverse association with helpfulness of post‐conflict reconciliatory attempts for men. Results indicate that couples' ability to maintain positive behavior during day‐to‐day conflict may be the best strategy for supporting subsequent efforts to reconcile and reconnect afterward.
Article
Background: The effect of a liberal transfusion strategy as compared with a restrictive strategy on outcomes in critically ill patients with traumatic brain injury is unclear. Methods: We randomly assigned adults with moderate or severe traumatic brain injury and anemia to receive transfusion of red cells according to a liberal strategy (transfusions initiated at a hemoglobin level of ≤10 g per deciliter) or a restrictive strategy (transfusions initiated at ≤7 g per deciliter). The primary outcome was an unfavorable outcome as assessed by the score on the Glasgow Outcome Scale-Extended at 6 months, which we categorized with the use of a sliding dichotomy that was based on the prognosis of each patient at baseline. Secondary outcomes included mortality, functional independence, quality of life, and depression at 6 months. Results: A total of 742 patients underwent randomization, with 371 assigned to each group. The analysis of the primary outcome included 722 patients. The median hemoglobin level in the intensive care unit was 10.8 g per deciliter in the group assigned to the liberal strategy and 8.8 g per deciliter in the group assigned to the restrictive strategy. An unfavorable outcome occurred in 249 of 364 patients (68.4%) in the liberal-strategy group and in 263 of 358 (73.5%) in the restrictive-strategy group (adjusted absolute difference, restrictive strategy vs. liberal strategy, 5.4 percentage points; 95% confidence interval, -2.9 to 13.7). Among survivors, a liberal strategy was associated with higher scores on some but not all the scales assessing functional independence and quality of life. No association was observed between the transfusion strategy and mortality or depression. Venous thromboembolic events occurred in 8.4% of the patients in each group, and acute respiratory distress syndrome occurred in 3.3% and 0.8% of patients in the liberal-strategy and restrictive-strategy groups, respectively. Conclusions: In critically ill patients with traumatic brain injury and anemia, a liberal transfusion strategy did not reduce the risk of an unfavorable neurologic outcome at 6 months. (Funded by the Canadian Institutes of Health Research and others; HEMOTION ClinicalTrials.gov number, NCT03260478.).
Article
We compared methods to project absolute risk, the probability of experiencing the outcome of interest in a given projection interval accommodating competing risks, for a person from the target population with missing predictors. Without missing data, a perfectly calibrated model gives unbiased absolute risk estimates in a new target population, even if the predictor distribution differs from the training data. However, if predictors are missing in target population members, a reference dataset with complete data is needed to impute them and to estimate absolute risk, conditional only on the observed predictors. If the predictor distributions of the reference data and the target population differ, this approach yields biased estimates. We compared the bias and mean squared error of absolute risk predictions for seven methods that assume predictors are missing at random (MAR). Some methods imputed individual missing predictors, others imputed linear predictor combinations (risk scores). Simulations were based on real breast cancer predictor distributions and outcome data. We also analyzed a real breast cancer dataset. The largest bias for all methods resulted from different predictor distributions of the reference and target populations. No method was unbiased in this situation. Surprisingly, violating the MAR assumption did not induce severe biases. Most multiple imputation methods performed similarly and were less biased (but more variable) than a method that used a single expected risk score. Our work shows the importance of selecting predictor reference datasets similar to the target population to reduce bias of absolute risk predictions with missing risk factors.
Article
Full-text available
Missing data present challenges for development and real‐world application of clinical prediction models. While these challenges have received considerable attention in the development setting, there is only sparse research on the handling of missing data in applied settings. The main unique feature of handling missing data in these settings is that missing data methods have to be performed for a single new individual, precluding direct application of mainstay methods used during model development. Correspondingly, we propose that it is desirable to perform model validation using missing data methods that transfer to practice in single new patients. This article compares existing and new methods to account for missing data for a new individual in the context of prediction. These methods are based on (i) submodels based on observed data only, (ii) marginalization over the missing variables, or (iii) imputation based on fully conditional specification (also known as chained equations). They were compared in an internal validation setting to highlight the use of missing data methods that transfer to practice while validating a model. As a reference, they were compared to the use of multiple imputation by chained equations in a set of test patients, because this has been used in validation studies in the past. The methods were evaluated in a simulation study where performance was measured by means of optimism corrected C‐statistic and mean squared prediction error. Furthermore, they were applied in data from a large Dutch cohort of prophylactic implantable cardioverter defibrillator patients.
Article
Full-text available
Background: We determined the impact of data volume and diversity and training conditions on recurrent neural network methods compared with traditional machine learning methods. Methods and results: Using longitudinal electronic health record data, we assessed the relative performance of machine learning models trained to detect a future diagnosis of heart failure in primary care patients. Model performance was assessed in relation to data parameters defined by the combination of different data domains (data diversity), the number of patient records in the training data set (data quantity), the number of encounters per patient (data density), the prediction window length, and the observation window length (ie, the time period before the prediction window that is the source of features for prediction). Data on 4370 incident heart failure cases and 30 132 group-matched controls were used. Recurrent neural network model performance was superior under a variety of conditions that included (1) when data were less diverse (eg, a single data domain like medication or vital signs) given the same training size; (2) as data quantity increased; (3) as density increased; (4) as the observation window length increased; and (5) as the prediction window length decreased. When all data domains were used, the performance of recurrent neural network models increased in relation to the quantity of data used (ie, up to 100% of the data). When data are sparse (ie, fewer features or low dimension), model performance is lower, but a much smaller training set size is required to achieve optimal performance compared with conditions where data are more diverse and includes more features. Conclusions: Recurrent neural networks are effective for predicting a future diagnosis of heart failure given sufficient training set size. Model performance appears to continue to improve in direct relation to training set size.
Article
Full-text available
PURPOSE: We set out to develop a real-time computerised decision support system (CDSS) embedded in the electronic health record (EHR) with information on risk factors, estimated risk, and guideline-based advice on treatment strategy in order to improve adherence to cardiovascular risk management (CVRM) guidelines with the ultimate aim of improving patient healthcare. METHODS: We defined a project plan including the scope and requirements, infrastructure and interface, data quality and study population, validation and evaluation of the CDSS. RESULTS: In collaboration with clinicians, data scientists, epidemiologists, ICT architects, and user experience and interface designers we developed a CDSS that provides ‘live’ information on CVRM within the environment of the EHR. The CDSS provides information on cardiovascular risk factors (age, sex, medical and family history, smoking, blood pressure, lipids, kidney function, and glucose intolerance measurements), estimated 10-year cardiovascular risk, guideline-compliant suggestions for both pharmacological and non-pharmacological treatment to optimise risk factors, and an estimate on the change in 10-year risk of cardiovascular disease if treatment goals are adhered to. Our pilot study identified a number of issues that needed to be addressed, such as missing data, rules and regulations, privacy, and patient participation. CONCLUSION: Development of a CDSS is complex and requires a multidisciplinary approach. We identified opportunities and challenges in our project developing a CDSS aimed at improving adherence to CVRM guidelines. The regulatory environment, including guidance on scientific evaluation, legislation, and privacy issues needs to evolve within this emerging field of eHealth.
Article
Full-text available
Purpose We set out to develop a real-time computerised decision support system (CDSS) embedded in the electronic health record (EHR) with information on risk factors, estimated risk, and guideline-based advice on treatment strategy in order to improve adherence to cardiovascular risk management (CVRM) guidelines with the ultimate aim of improving patient healthcare. Methods We defined a project plan including the scope and requirements, infrastructure and interface, data quality and study population, validation and evaluation of the CDSS. Results In collaboration with clinicians, data scientists, epidemiologists, ICT architects, and user experience and interface designers we developed a CDSS that provides ‘live’ information on CVRM within the environment of the EHR. The CDSS provides information on cardiovascular risk factors (age, sex, medical and family history, smoking, blood pressure, lipids, kidney function, and glucose intolerance measurements), estimated 10-year cardiovascular risk, guideline-compliant suggestions for both pharmacological and non-pharmacological treatment to optimise risk factors, and an estimate on the change in 10-year risk of cardiovascular disease if treatment goals are adhered to. Our pilot study identified a number of issues that needed to be addressed, such as missing data, rules and regulations, privacy, and patient participation. Conclusion Development of a CDSS is complex and requires a multidisciplinary approach. We identified opportunities and challenges in our project developing a CDSS aimed at improving adherence to CVRM guidelines. The regulatory environment, including guidance on scientific evaluation, legislation, and privacy issues needs to evolve within this emerging field of eHealth.
Article
Full-text available
Background: Cardiovascular risk management (CVRM) is notoriously difficult because of multi-morbidity and the different phenotypes and severities of cardiovascular disease. Computerized decision support systems (CDSS) enable the clinician to integrate the latest scientific evidence and patient information into tailored strategies. The effect on cardiovascular risk factor management is yet to be confirmed. Methods: We performed a systematic review and meta-analysis evaluating the effects of CDSS on CVRM, defined as the change in absolute values and attainment of treatment goals of systolic blood pressure (SBP), low density lipoprotein cholesterol (LDL-c) and HbA1c. Also, CDSS characteristics related to more effective CVRM were identified. Eligible articles were methodologically appraised using the Cochrane risk of bias tool. We calculated mean differences, relative risks, and if appropriate (I2 < 70%), pooled the results using a random-effects model. Results: Of the 14,335 studies identified, 22 were included. Four studies reported on SBP, 3 on LDL-c, 10 on CVRM in patients with type II diabetes and 5 on guideline adherence. The CDSSs varied considerably in technical performance and content. Heterogeneity of results was such that quantitative pooling was often not appropriate. Among CVRM patients, the results tended towards a beneficial effect of CDSS, but only LDL-c target attainment in diabetes patients reached statistical significance. Prompting, integration into the electronical health record, patient empowerment, and medication support were related to more effective CVRM. Conclusion: We did not find a clear clinical benefit from CDSS in cardiovascular risk factor levels and target attainment. Some features of CDSS seem more promising than others. However, the variability in CDSS characteristics and heterogeneity of the results - emphasizing the immaturity of this research area - limit stronger conclusions. Clinical relevance of CDSS in CVRM might additionally be sought in the improvement of shared decision making and patient empowerment.
Article
Full-text available
The overwhelming amount, production speed, multidimensionality, and potential value of data currently available-often simplified and referred to as big data -exceed the limits of understanding of the human brain. At the same time, developments in data analytics and computational power provide the opportunity to obtain new insights and transfer data-provided added value to clinical practice in real time. What is the role of the health care professional in collaboration with the data scientist in the changing landscape of modern care? We discuss how health care professionals should provide expert knowledge in each of the stages of clinical decision support design: data level, algorithm level, and decision support level. Including various ethical considerations, we advocate for health care professionals to responsibly initiate and guide interprofessional teams, including patients, and embrace novel analytic technologies to translate big data into patient benefit driven by human(e) values. ©Tim Bezemer, Mark CH de Groot, Enja Blasse, Maarten J ten Berg, Teus H Kappen, Annelien L Bredenoord, Wouter W van Solinge, Imo E Hoefer, Saskia Haitjema. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 19.03.2019.
Article
Objective: Roux-en-Y gastric bypass (RYGB) induces type 2 diabetes remission (DR) in 60% of patients at 1 year, yet long-term relapse occurs in half of these patients. Scoring methods to predict DR outcomes 1 year after surgery that include only baseline parameters cannot accurately predict 5-year DR (5y-DR). We aimed to develop a new score to better predict 5y-DR. Research design and methods: We retrospectively included 175 RYGB patients with type 2 diabetes with 5-year follow-up. Using machine learning algorithms, we developed a scoring method, 5-year Advanced-Diabetes Remission (DiaRem) (5y-Ad-DiaRem), predicting longer-term DR postsurgery by integrating medical history, bioclinical data, and antidiabetic treatments. The scoring method was based on odds ratios and variables significantly different between groups. This score was further validated in three independent RYGB cohorts from three European countries. Results: Compared with 5y-DR patients, 5y-Relapse patients exhibited more severe type 2 diabetes at baseline, lost significantly less weight during the 1st year after RYGB, and regained more weight afterward. The 5y-Ad-DiaRem includes baseline (diabetes duration, number of antidiabetic treatments, and HbA1c) and 1-year follow-up parameters (glycemia, number of antidiabetic treatments, remission status, 1st-year weight loss). The 5y-Ad-DiaRem was accurate (area under the receiver operating characteristic curve [AUROC], 90%; accuracy, 85%) at predicting 5y-DR, performed better than the DiaRem and Ad-DiaRem (AUROC, 81% and 84%; accuracy, 79% and 78%, respectively), and correctly reclassified 13 of 39 patients misclassified with the DiaRem. The 5y-Ad-DiaRem robustness was confirmed in the independent cohorts. Conclusions: The 5y-Ad-DiaRem accurately predicts 5y-DR and appears relevant to identify patients at risk for relapse. Using this score could help personalize patient care after the 1st year post-RYGB to maximize weight loss, limit weight regains, and prevent relapse.
Article
Background: Chained equations imputation is widely used in medical research. It uses a set of conditional models, so is more flexible than joint modelling imputation for the imputation of different types of variables (e.g. binary, ordinal or unordered categorical). However, chained equations imputation does not correspond to drawing from a joint distribution when the conditional models are incompatible. Concurrently with our work, other authors have shown the equivalence of the two imputation methods in finite samples. Methods: Taking a different approach, we prove, in finite samples, sufficient conditions for chained equations and joint modelling to yield imputations from the same predictive distribution. Further, we apply this proof in four specific cases and conduct a simulation study which explores the consequences when the conditional models are compatible but the conditions otherwise are not satisfied. Results: We provide an additional “non-informative margins” condition which, together with compatibility, is sufficient. We show that the non-informative margins condition is not satisfied, despite compatible conditional models, in a situation as simple as two continuous variables and one binary variable. Our simulation study demonstrates that as a consequence of this violation order effects can occur; that is, systematic differences depending upon the ordering of the variables in the chained equations algorithm. However, the order effects appear to be small, especially when associations between variables are weak. Conclusions: Since chained equations is typically used in medical research for datasets with different types of variables, researchers must be aware that order effects are likely to be ubiquitous, but our results suggest they may be small enough to be negligible.
Article
Among several features used for clinical binary classification, behavioral performance, questionnaire scores, test results and physical exam reports can be counted. Attempts to include neuroimaging findings to support clinical diagnosis are scarce due to difficulties in collecting such data, as well as problems in integration of neuroimaging findings with other features. The binary classification method proposed here aims to merge small samples from multiple sites so that a large cohort which better describes the features of the disease can be built. We implemented a simple and robust framework for detection of fibromyalgia, using likelihood during decision level fusion. This framework supports sharing of classifier applications across clinical sites and arrives at a decision by fusing results from multiple classifiers. If there are missing opinions from some classifiers due to inability to collect their input features, such degradation in information is tolerated. We implemented this method using fNIRS data collected from fibromyalgia patients across three different tasks. Functional connectivity maps are derived from these tasks as features. In addition, self-reported clinical features are also used. Five classifiers are trained using kNN, LDA and SVM. Fusion of classification opinions from multiple classifiers based on log likelihood outperformed other fusion methods reported in the literature for detection of FM. When 2, 3, 4 and 5 classifers are fused, sensitivity and specificity figures of 100% could be obtained based on the choice of the classifier set.