ArticlePDF Available

Abstract and Figures

In this study, we adapt a Cox-based model for recurrent events; the Prentice, Williams and Peterson Total-Time (PWP-TT) that has largely, been used under the assumption of non-informative censoring and evaluate it under an informative censoring setting. Empirical evaluation was undertaken with the aid of the semi-parametric framework for recurrent events suggested by Huang [1] and implemented in R Studio software. For validation we used data from a typical HIV care setting in Kenya. Of the three models under consideration; the standard Cox Model had gender hazard ratio (HR) of 0.66 (p-value=0.165), Andersen-Gill had HR 0.46 (with borderline p-value=0.054) and extended PWP TT had HR 0.22 (p-value=0.006). The PWP-TT model performed better as compared to other models under informative setting. In terms of risk factors under informative setting, LTFU due to stigma; gender [base=Male] had HR 0.544 (p-value =0.002), age [base is < 37] had HR 0.772 (p-value=0.008), ART regimen [base= First line] had HR 0.518 (p-value= 0.233) and differentiated care model (Base=not on DCM) had HR 0.77(p-value=0.036). In conclusion, in spite of the multiple interventions designed to address incidences of LTFU among HIV patients, within-person cases of LTFU are usually common and recurrent in nature, with the present likelihood of a person getting LTFU influenced by previous occurrences and therefore informative censoring should be checked.
Content may be subject to copyright.
International Journal of Statistics in Medical Research, 2020, 9, 20-27 20
E-ISSN: 1929-6029/20 © 2020 Lifescience Global
Analysis of Recurrent Events with Associated Informative
Censoring: Application to HIV Data
Jonathan Ejoku1, Collins Odhiambo1,* and Linda Chaba1
1Strathmore Institute of Mathematical Sciences, Strathmore University, Madaraka Estate, Ole Sangale Road,
P.O. Box 59857, 00200, City Square, Nairobi, Kenya
Abstract: In this study, we adapt a Cox-based model for recurrent events; the Prentice, Williams and Peterson Total
-Time (PWP-TT) that has largely, been used under the assumption of non-informative censoring and evaluate it under
an informative censoring setting. Empirical evaluation was undertaken with the aid of the semi-parametric framework for
recurrent events suggested by Huang [1] and implemented in R Studio software. For validation we used data from a
typical HIV care setting in Kenya. Of the three models under consideration; the standard Cox Model had gender hazard
ratio (HR) of 0.66 (p-value=0.165), Andersen-Gill had HR 0.46 (with borderline p-value=0.054) and extended PWP TT
had HR 0.22 (p-value=0.006). The PWP-TT model performed better as compared to other models under informative
setting. In terms of risk factors under informative setting, LTFU due to stigma; gender [base=Male] had HR 0.544
(p-value =0.002), age [base is < 37] had HR 0.772 (p-value=0.008), ART regimen [base= First line] had HR 0.518
(p-value= 0.233) and differentiated care model (Base=not on DCM) had HR 0.77(p-value=0.036). In conclusion, in spite
of the multiple interventions designed to address incidences of LTFU among HIV patients, within-person cases of LTFU
are usually common and recurrent in nature, with the present likelihood of a person getting LTFU influenced by previous
occurrences and therefore informative censoring should be checked.
Keywords: Recurrent events, Loss to follow-up, HIV, Prentice, Williams and Peterson Gap-Time, Informative
censoring.
1. BACKGROUND
Recurrent events occur in a variety of
disciplines/areas of life such as recurrent opportunistic
infections in HIV patients, episodes of asthmatic
attacks. Underlying processes that generate data from
these events are called “recurrent event processes and
the data they provide are called “recurrent event data”
[3]. There is a wide range of research on the analysis of
this data including the analysis of recurrence of
sports injuries, multiple episodes of childhood
diseases [4] and hospitalizations for chronic kidney
kidney disease [5].
A key characteristic of recurrent events is that
observations per individual are usually not independent
and in many cases are correlated, with current event
incidences influenced by previous incidences. A
number of approaches that have been proposed to
analyze recurrent events data, including both
parametric and non-parametric methods such as the
use of the Poisson and Negative binomial models [3].
Both of these however accommodate only
time-independent covariates, with the Poisson further
pre-supposing a constant events rate per individual.
While these and a number of other models have
been suggested, the models most predominant in
recurrent event literature for ordered outcomes are the
Anderson and Gill [6], Prentice, Williams, and Peterson
total time (PWP-TT) and gap time (PWP-GT) models
*Address correspondence to this author at the Strathmore Institute of
Mathematical Sciences, Strathmore University, Madaraka Estate, Ole Sangale
Road. P.O. Box 59857, 00200, City Square, Nairobi, Kenya; Mob: +254 703
034 000/200/300 or [+254] [0] 730-734000/200/300; Fax: +254 20 6007498;
E-mail: codhiambo@strathmore.edu
[7], Wei, Lin and Weissfeld (WLW) [8]; and Lee, Wei
and Amato (LWA). These models are essentially
extensions of the extensively used semi-parametric
Cox Proportional Hazards Model, predominantly
applied under the assumption of non-informative
censoring.
To the best of our knowledge, none of these models
have been applied to researches with an underlying
assumption of informative censoring. Subsequently, we
hope this paper will significantly bridge the gap by
evaluating the common recurrent event based models
in situations where informative censoring exists.
Informative censoring in HIV related-studies exist in
situations, where there are informed drop-outs; such as
sicker persons or those in either WHO staging III or IV
being withdrawn from the study or persons getting lost
to the study due to stigma-related factors. Ghosh et al.,
[9] note that informative censoring occurs in situations
where the censoring times “depend on the observed or
non-observed recurrent times”. An illustration of this in
the HIV setting, is the informative drop-outs such as the
withdrawal of sicker patients from a study way before
the study ends. Ghosh et al., [9] specify two possible
sources of informative censoring in practical settings:-
the voluntary withdrawal of subjects for reasons related
to the event process as well death.
The exclusion of censoring assumption from
different studies may generally lead to biased
estimates. Castelli et al. [10] in adapting the Inverse
Probability of Censoring Weighted [IPCW] to study the
survival times of asthma patients while including
informative censoring [patients felt they were okay and
Analysis of Recurrent Events with Associated Informative Censoring International Journal of Statistics in Medical Research, 2020, Vol. 9 21
did not need to consult a doctor] show that information
coming from censoring process improves the survival
estimate. Another approach is that by Ghosh et al., [9]
where they introduce a semi-parametric approach for
recurrent events in the presence of dependent
censoring and apply it ALIVE cohort study. A
comparison with the accelerated failure times model as
proposed by Lin et al. [11], which assumes
non-informative censoring found that their proposed
method yielded “a much larger estimate for the effect of
baseline HIV status on hospitalizations than the
method proposed by Lin”.
There is a body of literature to show that drop-outs
in HIV programs are informative. Berheto et al., [12] in
assessing the predictors of LTFU in Patients Living with
HIV/AIDS after Initiation of Antiretroviral Therapy found
that sicker patients [baseline CD4 < 200 cells/mm3
HR 1.7, 95% Cis 1.3-2.2, regimen substitution HR
5.2; 95% Cis 3.6-7.3] and receiving non-isoniazid (INH)
prophylaxis [HR 3.7; 95% CIs 2.3-6.2] as accelerators
for LTFU. According to Assemie et al., [13], being on
WHO clinical stage IV as well as receiving isoniazid
preventive therapy were significant predictors of LTFU.
On the premise of this, informative censoring is thus
incorporated in this study.
2. METHOD AND SETTING
Let the rate of occurrence of recurrent events in a
given time interval i.e. (0, ΓO), where ΓO > 0 is set with
the information of possible epochs of recurrences
observable up to ΓO. Suppose ℕ" (t) is the number of
recurrent events that occur on or before t, t > 0. The
functional rate of a recurrent event at !,!!![0,!!
!], is
expressed as
!!=lim
!!
Pr[[t!+!]!![t]!>!0]!
.
The rate is considered theoretically different from
the intensity function. We define the functional rate as
the occurrence rate of recurrent events and is not
conditional to the event history. The intensity function
on the other hand is the occurrence rate and is
conditional on the event history.
Let the cumulative rate function be described as
Ʌ!=![!]
!
!
!".
Also, suppose ϒ be the censoring time for
observation of the recurrent event process at
termination. The interest remains the occurrence rate
in the time interval (0, ΓO), and therefore, recurrent
event data beyond!!
!, is not important. We define τ =
min [ϒ, !!
!] as the new censoring time used in the
proposed models.
Consider PWP-TT defined as
!!" !=!!!!exp{!!"!}!with the following features
events are ordered and handled by stratification,
everyone is at risk for the first stratum, but only
those who
had an event in the previous stratum are at
risk for the successive one. The model estimates both
overall and event-specific effects and uses robust
standard errors to account for correlation.
2.1. Review of the Prentice, Williams, and Peterson
[PWP] Model
The PWP model as proposed by Prentice, Williams,
and Peterson [7] is a ‘conditional model’ where the
individuals are at risk for an event if and only if they
were at risk for a previous event. To achieve this, each
event occurrence is put into a different stratum with all
participants at risk in the first stratum. Under this model,
only participants that experienced the previous event
would then be at risk for the next event [14]. When time
since entry is of interest, this model condenses to total
time model (PWP-TT), while when time since last event
is of interest the model becomes a gap time model
Figure 1: Representation of the PWP Models.
22 International Journal of Statistics in Medical Research, 2020, Vol. 9 Ejoku et al.
commonly abbreviated as PWP-GT. See Figure 1. The
key difference between this model and the
Andersen-Gill (AG) is in terms of the effects of the
covariates in different strata. In the AG model, effects
of covariates are constant across all strata, while this
varies in different strata for the PWP [14].
2.2. Mathematical Representation of PWP Model
To include informative censoring, we built on
generic the approach proposed by Huang et al. [1]
where they jointly model the recurrent event process
and failure times. The key to this approach is to model
this relationship via a subject specific latent variable,!!,
that models the association between the intensity of the
recurrent event and the hazard of the failure time. This
approach is able to account for only time- independent
predictors while also allowing for informative censoring.
As relayed by Huang et al., the intensity function is
!!!=!!!!!exp x!.
And Hazard of the failure time, D, is given by:
!=!!!!exp !" .
Where:
!! and ! are baseline intensity and hazard functions
respectively.
! and ! are coefficients.
!![!] is the p-dimensional set of time-dependent
covariates
Based on the proposed intensity function by Huang,
the PWP models are modified as below. A
mathematical representation of each of the models is
presented below, as well as the individual model
specification with informative censoring incorporated
(Table 1). Essentially, each model is multiplied by the
unobserved frailty, !!.
Where:-
PWP-GT is the Prentice, Williams, and Peterson
Gap-Time Model,
PWP-TT is the Prentice, Williams, and Peterson
Total-Time Model
And: -
!!! is the at risk indicator for the j-th event and ith
person at time t. This is 1 when at risk for event j, and
zero when not at risk for event j
!! is the baseline intensity function
!![!] is the p-dimensional set of time-dependent
covariates
! is a q x 1 dimensional parameter.
2.3. Evaluation of the PWP-GT model under
informative censoring
2.3.1. Parameter Estimation
The key to this approach is to use a subject specific
latent variable to model the association between the
recurrent event and hazard of the failure times.
Additionally, no assumptions are made on the
distributions of censoring times and latent variables.
Specifically, we estimate the cumulative hazard and
intensity functions. For estimation, we use the
approach relayed by Huang et al. as it remains
consistent with the PWP model with the at risk indicator
treated as a nuisance parameter.
The estimation by Huang is briefly provided below.
Starting with notation:-
Let N[t] which represents events occurring at or
before some time To, D be failure time, say LTFU due
to stigma for this study, C failure time due to other
reasons than D, x is a vector of 1xp covariates.
Let Y be the point at which the observation of
recurrent events ceases, such that, Y=min[C, D, To].
We introduce a non-negative latent variable z, such
that given X=x, and Z=z, the intensity function is
given by
!!!=!!!!!exp x!,
with E[Z|x] = E[z]
And Hazard function by
!=!!!!exp !" .
Implicitly, a large/small z increases or decreases
the intensity and hazard respectively. The rate function
is defined as !!!!!exp![!" ]
Table 1: Mathematical Representation of the Intensity Functions of PWP Models with Informative Censoring
Name of The Model
Recurrent Model Mathematical presentation
Form under Informative/Non-informative censoring
PWP-GT
!!" !!!"[!!
!!!]!exp![x′!]
!!!!" !!!"[!!
!1]!exp![x′!]
PWP-TT
!!" !!!"[!]!exp![x′!]
!!!!" !!!"[!]!exp![x′!]
Analysis of Recurrent Events with Associated Informative Censoring International Journal of Statistics in Medical Research, 2020, Vol. 9 23
Assuming that for individual i, the collection
{[!!,!!,!!!!,0,!!"! ,!
!} are iid, then the density
function is given by
!!!!!!!exp![!!!]
!!!!Λ!!exp![!!!].
However we assume that the density does not
depend on Zi, Mi nor x, reducing the density to
!!!
Λ!!
.
Since we assume iid, the conditional likelihood is
generated for n subjects, assuming mi events per
individual as
.
!
!!!
.
!
!!!
!![!!"]
Λ!!!
.
Huang notes that the rate estimator for Λ!!! can
be given by [1!
!] [1]
where d is the number of patients experiencing event at
time t, and R is total persons at risk. To estimate the
hazard !, a class of estimators given below is solved.
!!!!!!!
![!!Λ!!]!!exp !" =0.
Huang proposes the estimation of ! by replacing
Λ![!]!! by 1.
Asymptotic properties of these estimators are
extensively studied by Huang and bear no repeating in
this article.
2.3.2. Empirical Evaluation of Data
Patient-level data from four facilities in central
Kenya, collected between 2013 and 2018 was used.
The recurrent event of interest were the incidences of
loss to follow-up. Censoring for a drop-out was
considered informative, based on the reason
provided for drop-out, a person cited ‘stigma’ or
adverse drug reaction. Additionally, time independent
covariates were incorporated in the analysis. Data
preparation was prepared to reflect the PWP-GT
outline format as laid out by Therneau [15]. Primarily,
we sought to establish if the incorporation of
informative censoring improves estimation of the
hazard and rate estimates. While this study
acknowledges the possibility of competing risks, they
were not incorporated in this study.
2.3.3. Computing Environment
The models were assessed under the Rstudio
computing environment. To account for informative
censoring, the ‘method’ function within ‘reReg function’
in the reReg was set to "cox.HW". Fitting the models
was achieved by using the cluster and strata functions
in the base survival package required by the reReg
package. The Hmisc package was used to provide
basic descriptive statistics.
3. APPLICATION TO HIV DATA
3.1. Data Description
Determination of the incidence of due to LTFU was
computed using the cumulative baseline hazard
technique with start of ART as time 0 and LTFU at the
time a particular patient failed to return to CCC clinic for
38 weeks since the scheduled appointment date.
Patients who did not experience the event [LTFU],
were right censored at the last clinic visit. Other
patients exit such as mortality and transfer out were
considered as non-informative censoring as they relate
to LTFU, as mortality and transfer out that were known
to specific clinics and had happened within 38 days of
the last clinical appointment were managed in a
competing risk framework. After determination of LTFU,
several consulted efforts to retrace clients who are lost
are implemented. The efforts include calling back
clients and visiting them at their homes with the aim of
returning them back to clinic. Patients who return back
to care are exposed to recurrent of the event [LTFU].
Among the sample of LTFU clients, we also collected
other parameters include Patient ID: This is the patient
ID which may be repeated due to recurrence of LTFU,
time to LTFU: the event of interest is LTFU in this
setting which is recurrent, status; for episodes of LTFU
vs no episode; a particular individual can experience
several LTFU episodes, event (Informative Censoring);
the patient is no longer observed-Informative
censoring was defined by the patient being stopped
from highly active antiretroviral therapy (HAART) due
to either drug reaction/ stigma, Differentiated HIV care
for patients who are either under differentiated care
model or not, age of the patient, and gender of the
patient ARV regimen line as defined by WHO. All
PLHIV adults on HAART who enrolled at four facilities
in central Kenya in January 2013 to December 2018
were considered for analysis. Those who had at-least
one follow-up visit were eligible to be included. Children
below 15 years and confirmed pregnant mothers were
excluded. We also excluded PLHIV who had unknown
ART start date, unknown outcome, and transferred in
with incomplete base-line information. Data was pulled
from point of care electronic medical database,
consolidated in MS excel, and exported to R Studio for
further analysis. The main event variable in the
analysis was time to LTFU (in months) with other exits
24 International Journal of Statistics in Medical Research, 2020, Vol. 9 Ejoku et al.
treated as a competing event. LTFU was defined as
patients not taking ART refill for a period of 38 days or
missed clinical appointment over the same duration.
4. DISCUSSION
4.1. Recurrent Events Model
In this study, we extended PWP-TT model to cover
informative censoring when making inferences on
recurrent events under HIV retention setting. The
proposed methodology intrinsically extends the existing
method in literature on recurrent events under typical
HIV resource limited setting. A well-defined truncation
time T, with Zi overall i.e. !!!!" !!!"(!!
!
1)!exp![!!!!!] has been assumed. Here, we have
concentrated on the LTFU as recurrent events and
fitted real data. It can also be developed further to other
regression/Cox methods. The concept can be
extended to other approaches by modeling !!"[!]!and
!!![!!
!1]!with proportional hazards models. The
main purpose of this work was to provide an overview,
applicable statistical techniques when analyzing
recurrent event data under informative censoring in
HIV retention setting. The typical real data used in this
work is from a routine well established HIV care clinic.
The longitudinal approach to analyze recurrent-event
data applied here can also be applicable to other
observational cohort studies. Because the technical
approach employed here are extensions of Cox
proportional hazards regression, explicit issues that
affect model modification can also be handled in the
same manner as the classic applied techniques.
4.2. Results
Majority of the patients were female 37 (64%). The
average age was 36.4 years (SD=6.28). There were
256 incidences of patients lost to follow up, albeit there
were no noted incidences among seven patients. Table
2 shows a breakdown of incidence of LTFU as well as
those resulting from the informative terminal event,
stigmatization. These are disaggregated by gender and
age [binary covariate set at population mean]. Overall,
6 in 10 patients [61%] experienced 1 to 6 LTFU
incidences. The median time to LTFU was 8.6 months
[range: 3.0-82.9 months].
Patients were either censored at the end of the
study or informative drop-out due to stigma or drug
reaction. Sixteen instances of informative censoring
were reported. Comparison of crude proportions of
LTFU incidence due to stigma revealed a higher
incidence among patients above 37 years. However,
there was no significant difference by gender. The
timing of the LTFU incidences and corresponding
informative censoring arising from either stigma or drug
reaction are shown in Figure 2.
Additionally, cumulative sample mean plots are
provided to provide the cumulative number of LTFU
occurrences during the study period. The number of
incidences is shown to increase consistently across the
monitoring window.
To assess the performance of the proposed
PWP-TT model under joint modelling of both the failure
and event time, we apply it to the patient level data
described in section 3.1. For this analysis, sex [1 for
male, 2 for female], age [1 for 37, 2 for >37 years],
Regimen [1 for first line, 2 for second line], and
differentiated care [1 for on differentiated care, 0 for not
on differentiated care] were used. The event was LTFU
(1 for LTFU occurrence, 0 for no occurrence), while a
composite censoring variable was defined (1 if due to
stigma, 0 for other exits). Specifically, we investigate
the effect of these covariates on rate of LTFU and risk
of LTFU arising from stigma. Parameter estimates,
standard errors (SE), and corresponding p-values are
shown in Table 3. Standard errors were estimated by
resampling 100 times from patient data. Additionally,
Table 2: A distribution of the Total Number of LTFU Incidences Occurring among 58 Patients
Total
Patients
LTFU due to stigma
Number of LTFU incidences since ART start
0
1
2
3
4
5
6
7
8
9
21
8
2
1
0
5
3
0
7
3
-
0
36%
50.0%
10%
5%
0%
24%
14%
0%
33%
14%
0%
37
8
5
1
0
8
5
1
6
10
1
64%
50.0%
14%
3%
0%
22%
14%
3%
16%
27%
3%
30
6
3
2
0
6
3
0
5
10
1
52%
37.5%
10%
7%
0%
20%
10%
0%
17%
33%
3%
28
10
4
0
0
7
5
1
8
3
0
48%
62.5%
14%
0%
0%
25%
18%
4%
29%
11%
0%
Analysis of Recurrent Events with Associated Informative Censoring International Journal of Statistics in Medical Research, 2020, Vol. 9 25
results from the AG and Cox model are also computed
(output not included) and compared.
From Table 3, it can be noted that patients on
differentiated care had lower likelihoods for both LTFU
recurrence (30% lower) as well as for the incidence of
LTFU arising from stigma (80% less likely). These
findings are unsurprising given that one well known
benefit of differentiated care models is to fight stigma.
Gender and age are also significant predictors of LTFU
due to stigma. The instantaneous risk for LTFU from
stigma is lower for females [82% less risk], but higher
for persons over the age of 37 years.
For the Cox model, none of the covariates was
significant for risk of LTFU from stigma. On the other
hand, the AG model also indicated a 73% risk of LTFU
from stigma for persons on a differentiated care
program, as well as a 23% reduced recurrence of
LTFU. The effect and direction of the other covariates
for the AG are also comparable with those for the
PWP-TT. These results suggest Cox regression for
recurrent events may not be suitable in observational
studies.
As an alternative to cox regression models, one
may consider using the Generalized Estimating
Equations (GEE), which may be employed in instances
of longitudinal and correlated data, especially if the
responses are binary. One key advantage of this
approach is its robustness in giving consistent results
even with a mis-specification of the correlation
structure. However one key limitation is that responses
need to be correlated, which is not always the case
with recurrent survival data.
On the other hand, the analysis of recurrent-events
based only on the first event time is not ideal when
examining the effect of risk factors. This is underlined
by Ullah et al., [16] who wrote several of articles on
Figure 2: Recurrent history of LTFU per patient. Red triangles represent informative censoring due to stigma/drug reaction. The
green circles are incidences of LTFU.
Table 2: Estimates for Risk Factors with Standard Errors, and p-Values
Risk factor
Estimate
SE
z value
p-value
LTFU Recurrence
Gender [base=Male]
0.043
0.131
0.323
0.746
Age [base is < 37]
-0.014
0.087
-0.166
0.868
Regimen [base= First line]
0.067
0.162
0.412
0.681
Differentiated care [Base=not on Differentiated care]
-0.356
0.164
-2.166
0.03
*
LTFU due to stigma
Gender [base=Male]
-1.694
0.544
-3.113
0.002
**
Age [base is < 37]
2.034
0.772
2.634
0.008
**
Regimen [base= First line]
0.618
0.518
1.192
0.233
Differentiated care [Base=not on Differentiated care]
-1.615
0.77
-2.097
0.036
*
Codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1.
26 International Journal of Statistics in Medical Research, 2020, Vol. 9 Ejoku et al.
recurrent injuries with a use of statistical methods,
which only account for the first event thus excluding
key information from subsequent injuries. Odhiambo et
al., [17] applies a Block-Borges-Savits (BBS) minimal
repair model to HIV retention data to recurrent event
data, with subsequent use of smooth tests to assess fit.
Several approaches have been proposed in
literature to account for censoring that rises in survival
analysis setting.
The major assumption in AG model is that the
inter-event time increments are conditionally
uncorrelated with given covariates. It also assumes the
same baseline hazard for all persons, which may not
be the case for LTFU in an HIV setting, given that
intensive adherence counseling may alter the
subsequent likelihood of an individual getting LTFU. On
the other hand, it is best suited to cases of independent
increments across observation units, and is also the
easiest extension to the Cox model to replicate. On the
other hand, either of the PWP models (GT or TT)
adjusts for varying baseline hazards across
observation units which may be efficient in the case of
LTFU, where due to adherence counseling, the
baseline hazard may change.
Generally, the choice of recurrent event data
analysis technique is determined by several factors, i.e.
events; relationship between events; varying effects
across recurrences; the medical/biological process;
and independence/dependence structure. Usually the
stratified models, as PWP [total or gap times] or
multi-state models, are useful whenever there are
relatively few recurrent-events per individual and the
risk of recurrences.
A recurrent events model will ideally help to provide
insights into the program/disease structure and
process. Hence, it is critical to consider the censoring
mechanism and perform analysis that enhances
comprehension of the risk factors.
LTFU
In a typical HIV care clinic, the risk posed by
instances of LTFU is undesired and has the potential of
undoing antiretroviral treatment benefits. Specifically,
patients’ retention in HIV care is critical to ensuring
better health outcomes especially in reduced viral load
suppression, mitigating mortality and averting possible
drug resistance caused by non-adherence.
In spite of the multiple interventions in place to
address incidences of LTFU among HIV patients like
enhanced adherence counselling, within-person cases
of LTFU are usually common and recurrent in nature,
with the present likelihood of a person getting lost to
follow-up influenced by previous occurrences.
ABBREVIATION
AD = Anderson and Gill
AIDS = Acquired Immunodeficiency Syndrome
ART = Antiretroviral Therapy
BBS = Block-Borges-Savits
GEE = Generalized Estimating Equations
HAART = Highly Active Antiretroviral Therapy
CCC = Comprehensive Care Center
DCM = Differentiated Care Model
HIV = Human Immunodeficiency Virus
HR = Hazard Ratio
LTFU = Lost to follow-up
LWA = Lee, Wei and Amato
PWP-GT = Prentice, Williams, and Peterson gap time
PLHIV = Persons Living with HIV
PWP-TT = Prentice, Williams, and Peterson total
time
SD = Standard Deviation
SE = Standard Error
WLW = Wei, Lin and Weissfeld and
INH = Isonicotinylhydrazide
WHO = World Health Organization
REFERENCES
[1] Huang C-Y, Wang M-C. Joint Modeling and Estimation for
Recurrent Event Processes and Failure Time Data. J Am
Stat Assoc 2004; 99(468): 1153-65.
https://doi.org/10.1198/016214504000001033
[2] Chiou SH [Steven], Huang C-Y. reReg: Recurrent Event
Regression [Internet]. 2018 [cited 2019 Jul 23]. Available
from: https://CRAN.R-project.org/package=reReg
[3] Lawless JF. Regression Methods for Poisson Process Data.
J Am Stat Assoc 1987; 82(399): 808-15.
https://doi.org/10.1080/01621459.1987.10478502
[4] Kelly PJ, Lim LL-Y. Survival analysis for recurrent event data:
an application to childhood infectious diseases. Stat Med
2000; 19(1): 13-33.
https://doi.org/10.1002/(SICI)1097-0258(20000115)19:1<13:
:AID-SIM279>3.0.CO;2-5
[5] Yang W, Jepson C, Xie D, Roy JA, Shou H, Hsu JY, et al.
Statistical Methods for Recurrent Event Analysis in Cohort
Studies of CKD. Clin J Am Soc Nephrol 2017; 12(12):
2066-73.
https://doi.org/10.2215/CJN.12841216
Analysis of Recurrent Events with Associated Informative Censoring International Journal of Statistics in Medical Research, 2020, Vol. 9 27
[6] Andersen PK, Gill RD. Cox’s Regression Model for Counting
Processes: A Large Sample Study. Ann Stat 1982; 10(4):
1100-20.
https://doi.org/10.1214/aos/1176345976
[7] Prentice RL, Williams BJ, Peterson AV. On the regression
analysis of multivariate failure time data. Biometrika 1981;
68(2): 373-9.
https://doi.org/10.1093/biomet/68.2.373
[8] Wei LJ, Lin DY, Weissfeld L. Regression Analysis of
Multivariate Incomplete Failure Time Data by Modeling
Marginal Distributions. J Am Stat Assoc 1989; 84(408):
1065-73.
https://doi.org/10.1080/01621459.1989.10478873
[9] Ghosh D, Lin DY. Semiparametric Analysis of Recurrent
Events Data in the Presence of Dependent Censoring.
Biometrics 2003; 59(4): 877-85.
https://doi.org/10.1111/j.0006-341X.2003.00102.x
[10] Christel Castelli, Philippe Saint-Pierre, Jean-Pierre Daures.
Informative censoring in survival analysis and application to
asthma | Christel Castelli | Request PDF [Internet].
ResearchGate. 2006 [cited 2019 Jul 22]. Available from:
https://www.researchgate.net/publication/262413091_Inform
ative_censoring_in_survival_analysis_and_application_to_a
sthma/citations
[11] Lin DY, Wei LJ, Ying Z. Accelerated Failure Time Models for
Counting Processes. Biometrika 1998; 85(3): 605-18.
https://doi.org/10.1093/biomet/85.3.605
[12] Berheto TM, Haile DB, Mohammed S. Predictors of Loss to
follow-up in Patients Living with HIV/AIDS after Initiation of
Antiretroviral Therapy. North Am J Med Sci 2014; 6(9):
453-9.
https://doi.org/10.4103/1947-2714.141636
[13] Assemie MA, Muchie KF, Ayele TA. Incidence and predictors
of loss to follow up among HIV-infected adults at Pawi
General Hospital, northwest Ethiopia: competing risk
regression model. BMC Res Notes [Internet]. 2018 May 10
[cited 2019 Jul 4]; 11.
https://doi.org/10.1186/s13104-018-3407-5
[14] Amorim LD, Cai J. Modelling recurrent events: a tutorial for
analysis in epidemiology. Int J Epidemiol 2015; 44(1):
324-33.
https://doi.org/10.1093/ije/dyu222
[15] Therneau TM, Grambsch PM. Modeling Survival Data:
Extending the Cox Model. Springer Science & Business
Media 2000; p. 372.
https://doi.org/10.1007/978-1-4757-3294-8
[16] Ullah S, Gabbett TJ, Finch CF. Statistical modelling for
recurrent events: an application to sports injuries. Br J Sports
Med 2014; 48(17): 1287-93.
https://doi.org/10.1136/bjsports-2011-090803
[17] Odhiambo C, Odhiambo J, Omolo B. A Smooth Test of
Goodness-of-Fit for the Baseline Hazard Function for
Time-to-First Occurrence in Recurrent Events: An
Application to HIV Retention Data. Int J Stat Med Res 2017;
6(3): 104-113.
https://doi.org/10.6000/1929-6029.2017.06.03.2
Received on 26-02-2020 Accepted on 19-03-2020 Published on 29-03-2020
https://doi.org/10.6000/1929-6029.2020.09.03
© 2020 Ejoku et al.; Licensee Lifescience Global.
This is an open access article licensed under the terms of the Creative Commons Attribution Non-Commercial License
(http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted, non-commercial use, distribution and reproduction
in any medium, provided the work is properly cited.
... According to the Prentice, Williams, and Peterson (PWP) model (Prentice, Ross L and Williams, Benjamin J and Peterson, Arthur V, 1981), a person is not at risk for a future event until the preceding event happens. Despite the fact that there is a substantial body of literature on modeling recurrent events using the PWP model in the fields of medicine (Ejoku, Jonatha, 2020), consumer behavior (Bijwaard, Govert E and Franses, Philip Hans and Paap, Richard, 2006), and (Zhou, Fanyin and Fu, Lijun and Li, Zhiyong and Xu, Jiawei, 2022). ...
Article
Full-text available
p class="MsoNormal" style="margin-top: 12pt; text-align: justify;"> This research evaluates the efficacy of survival models in forecasting startup failures and investigates their economic implications. Several machine learning survival models, including Kernel SVM, DeepSurv, Survival Random Forest, and MTLR, are assessed using the concordance index (C-index) as a measure of prediction accuracy. The findings reveal that more sophisticated models, such as Multi-Task Logical Regression (MTLR) and Random Forest, outperform the standard Cox and Kaplan Meier (K-M) models in terms of predicted accuracy. </p
... The Prentice, Williams, and Peterson (PWP) model (Prentice et al., 1981) assumes that a person is not at risk for a future occurrence until the preceding event occurs. Even though there is considerable literature on modeling recurrent events using the PWP model in the fields of medical (e.g., Ejoku, Odhiambo, & Chaba, 2020;Moulton & Dibley, 1997;Pea, Slate, & Gonza lez, 2007;Pfennig et al., 2010), consumer behavior (Bijwaard, Franses, & Paap, 2006), and product or equipment reliability (e.g., 1983;Jiang, Landers, & Reed Rhoads, 2006). ...
Article
Full-text available
This study investigates the likelihood of time to bank failures in the US between 2001 and April 2023, based on data collected from the Federal Deposit Insurance Corporation's report on "Bank Failures in Brief - Summary 2001 through 2023". The dataset includes 564 instances of bank failures and several variables that may be related to the likelihood of such events, such as asset amount, deposit amount, ADR, deposit level, asset level, inflation rate, short-term interest rates, bank reserves, and GDP growth rate. We explore the efficacy of machine learning survival models in predicting bank failures and compare the performance of different models. Our findings shed light on the factors that may influence the probability of bank failures with a time perspective and provide insights for improving risk management practices in the banking industry.
... Using the borrowing-strength approach in the estimating procedure, our model allows users to specify any combination of the submodels between the recurrent event process and the terminal events when they are fitted jointly. Since the reReg package's debut on CRAN, it has been applied in many medical studies (e.g., Richter et al. 2020;Ejoku, Odhiambo, and Chaba 2020;Deo et al. 2021). ...
Article
Full-text available
Recurrent event analyses have found a wide range of applications in biomedicine, public health, and engineering, among others, where study subjects may experience a sequence of event of interest during follow-up. The R package reReg offers a comprehensive collection of practical and easy-to-use tools for regression analysis of recurrent events, possibly with the presence of an informative terminal event. The regression framework is a general scalechange model which encompasses the popular Cox-type model, the accelerated rate model, and the accelerated mean model as special cases. Informative censoring is accommodated through a subject-specific frailty without any need for parametric specification. Different regression models are allowed for the recurrent event process and the terminal event. Also included are visualization and simulation tools.
... However, in the Prentice, Williams and Peterson (PWP) model (Prentice et al., 1981), an individual is assumed not at risk for a subsequent event until the previous event has occurred. Even though there is considerable literature on modelling of recurrent events using the PWP model in the fields of medical (e.g., Ejoku, Odhiambo, & Chaba, 2020;Moulton & Dibley, 1997;Peña, Slate, & González, 2007;Pfennig et al., 2010), consumer behaviour (Bijwaard, Franses, & Paap, 2006) and product or equipment reliability (e.g., Ascher, 1983;Jiang, Landers, & Reed Rhoads, 2006). Only a few studies can be found in corporate finance. ...
Article
Full-text available
Companies often suffer periods of financial distress before filing for bankruptcy. Unlike one-off bankruptcies, financial distress can occur repeatedly within the same individual firm. This paper is focused on the recurrence of financial distress and studies the Chinese stock market, where Special Treatment – an official indicator of financial distress – can be repeatedly applied to a listed company. We employ a stratified hazard model to predict the probability of subsequent distress with variables, including duration dependency, event-based factors, institutional variables, financial ratios, market-based variables and macroeconomic conditions. Our empirical results show that accounting and market-based variables have limited power in predicting the recurrence of distress, whereas the duration of recovery, restructuring events and their interaction terms with the accounting and macroeconomic factors affect the recurrent risk significantly. Tested on out-of-time samples, our proposed hazard models show a robust performance in the prediction of recurrent risk over time.
... Since the reReg package's debut on CRAN, it has been applied in many medical studies (e.g. Richter, Sierocinski, Singer, Bülow, Hackmann, Chenot, and Schmidt 2020;Ejoku, Odhiambo, and Chaba 2020;Deo, Sundaram, Sahadevan, Selvaganesan, Mohan, Rubelowsky, Josephson, Elgudin, Kilic, and Cmolik 2021). ...
Preprint
Full-text available
Recurrent event analyses have found a wide range of applications in biomedicine, public health, and engineering, among others, where study subjects may experience a sequence of event of interest during follow-up. The R package reReg (Chiou and Huang 2021) offers a comprehensive collection of practical and easy-to-use tools for regression analysis of recurrent events, possibly with the presence of an informative terminal event. The regression framework is a general scale-change model which encompasses the popular Cox-type model, the accelerated rate model, and the accelerated mean model as special cases. Informative censoring is accommodated through a subject-specific frailty without no need for parametric specification. Different regression models are allowed for the recurrent event process and the terminal event. Also included are visualization and simulation tools.
Article
Full-text available
Objective: This study was aimed at assessing the incidence of lost-to-follow-up and its predictors among HIV-positive adults after initiation into antiretroviral therapy at Pawi General Hospital, northwest Ethiopia. Results: The overall cumulative incidence of lost-to-follow-up after ART initiation was high, 11.6 (95% CI 9.8-13.7) per 100 adult-years follow-up time. Independent significant predictors of lost to follow up were being aged 15-28 years (aSHR = 0.44; 95% CI 0.24-0.83), being on WHO clinical stage IV (aSHR = 2.09; 95% CI 1.02-3.13); and receiving isoniazid preventive therapy (aSHR = 0.11; 95% CI 0.06-0.18).
Article
Full-text available
Motivated by HIV retention, we present an application of the smooth test of goodness-of-fit under right-censoring to time to first occurrence of a recurrent event. The smooth test applied here is an extension of Neyman’s smooth test to a class of hazard functions for the initial distribution of a recurrent failure-time event. We estimate the baseline hazard function of time-to-first loss to follow-up, using a Block, Borges and Savits (BBS) minimal repair model of the data (n= 2,987, 72% censored). Simulations were conducted at various percentages of censoring to assess the performance of the smooth test. Results show that the smooth test performed well under right-censoring.
Article
Full-text available
Background: Long-term regular follow up of ART is an important component of HIV care. Patients who are lost to follow-up (LTFU) while on treatment compromise their own health and the long-term success of ART programs. Aim: This study was aimed at determining the incidence and risk factors for LTFU in HIV patients on ART at ART clinic of Mizan-Aman General Hospital, Ethiopia. Materials and Methods: A retrospective cohort study of 2133 people living with HIV/AIDS and attending an ART clinic between 2005 and 2013 was undertaken. LTFU was defined as not taking an ART refill for a period of 3 months or longer from the last attendance for refill and not yet classified as ‘dead’ or ‘transferred-out’. The log-rank test was used to measure differences in time to LTFU between groups and Cox proportional hazards modeling was used to measure predictors of LTFU. Results: Of 2133 patients, 53.9% were female. The mean (SD) age of the cohort was 31.5 (8.0), 16 (2.2), and 3.8 (3.0) years for adults, adolescents, and children, respectively. Around 574 (26.7%) patients were defined as LTFU. The cumulative incidence of LTFU was 8.8 (95% CIs 8.1-9.6) per 1000 person months. Patients with regimen substitution (HR 5.2; 95% CIs 3.6-7.3), non-isoniazid (INH) prophylaxis (HR 3.7; 95% CIs 2.3-6.2), adolescent (HR 2.1; 95% CIs 1.3-3.4), and had a baseline CD4 count < 200 cells/mm3 (HR 1.7, 95% CIs 1.3-2.2) were at higher risk of LTFU. WHO clinical stage III (HR 0.6; 95% CIs 0.4-0.9) and IV (HR 0.8; 95% CIs 0.6-1.0) patients at entry were less likely to be LTFU than clinical stage I patients. There was no significant difference in risk of LTFU in males and females. Conclusion: Overall, these data suggested that LTFU in this study was high. Patients phase of life, drug related factors, and clinical stages were associated with LTFU in this study. Effective control measures in the at-risk population need to be implemented to improve retention.
Article
Full-text available
Many survival studies record the times to two or more distinct failures on each subject. The failures may be events of different natures or may be repetitions of the same kind of event. In this article, we consider the regression analysis of such multivariate failure time observations. Each marginal distribution of the failure times is formulated by a Cox proportional hazards model. No specific structure of dependence among the distinct failure times on each subject is imposed. The regression parameters in the Cox models are estimated by maximizing the failure-specific partial likelihoods. The resulting estimators are shown to be asymptotically jointly normal with a covariance matrix that can be consistently estimated. Simultaneous inferential procedures are then proposed. Extensive Monte Carlo studies indicate that the normal approximation is adequate for practical use. The new methods allow time-dependent covariates, missing observations, and arbitrary patterns of censorship. They are illustrated with two real-life examples. For recurrent failure time data, various regression methods have been proposed in the literature. These methods, however, generally assume stringent structures of dependence among the recurrences of each subject. Moreover, as shown in the present article, they are rather sensitive to model misspecification.
Article
Cardiovascular events, such as hospitalizations because of congestive heart failure, often occur repeatedly in patients with CKD. Many studies focus on analyses of the first occurrence of these events, and discard subsequent information. In this article, we review a number of statistical methods for analyzing ordered recurrent events of the same type, including Poisson regression and three commonly used survival models that are extensions of Cox proportional hazards regression. We illustrate the models by analyzing data from the Chronic Renal Insufficiency Cohort Study to identify risk factors for congestive heart failure hospitalizations in patients with CKD. We show that recurrent event analyses provide additional insights about the data compared with a standard survival analysis of time to the first event.
Article
In many biomedical studies, the event of interest can occur more than once in a participant. These events are termed recurrent events. However, the majority of analyses focus only on time to the first event, ignoring the subsequent events. Several statistical models have been proposed for analysing multiple events. In this paper we explore and illustrate several modelling techniques for analysis of recurrent time-to-event data, including conditional models for multivariate survival data (AG, PWP-TT and PWP-GT), marginal means/ rates models, frailty and multi-state models. We also provide a tutorial for analysing such type of data, with three widely used statistical software programmes. Different approaches and software are illustrated using data from a bladder cancer project and from a study on lower respiratory tract infection in children in Brazil. Finally, we make recommendations for modelling strategy selection for analysis of recurrent event data.
Book
Introduction.- Estimating the Survival and Hazard Functions.- The Cox Model.- Residuals.- Functional Form.- Testing Proportional Hazards.- Influence.- Multiple Events per Subject.- Frailty Models.- Expected Survival.
Article
In epidemiological studies, survival analyses are often carried out in order to better understand the onset of an event. The data have the particularity of being incomplete due to the different censoring phenomena. Traditional methods make the hypothesis of censoring being independent from the event, which may be a source of bias in certain pathologies. The Inverse Probability of Censoring Weighted (IPCW) method adapts the Kaplan-Meier estimators and the Cox partial likelihood method to cases, with non-independent censoring. This method uses the information resulting from censoring to modify the contribution of individuals in the estimators. This method is applied to asthma, a case in which therapists believe that patients lost to follow-up are patients who are in otherwise good health, and do not feel the necessity to consult a doctor (informative censoring).
Article
Recurrent event data are commonly encountered in longitudinal follow-up studies related to biomedical science, econometrics, reliability, and demography. In many studies, recurrent events serve as important measurements for evaluating disease progression, health deterioration, or insurance risk. When analyzing recurrent event data, an independent censoring condition is typically required for the construction of statistical methods. In some situations, however, the terminating time for observing recurrent events could be correlated with the recurrent event process, thus violating the assumption of independent censoring. In this article, we consider joint modeling of a recurrent event process and a failure time in which a common subject-specific latent variable is used to model the association between the intensity of the recurrent event process and the hazard of the failure time. The proposed joint model is flexible in that no parametric assumptions on the distributions of censoring times and latent variables are made, and under the model, informative censoring is allowed for observing both the recurrent events and failure times. We propose a "borrow-strength estimation procedure" by first estimating the value of the latent variable from recurrent event data, then using the estimated value in the failure time model. Some interesting implications and trajectories of the proposed model are presented. Properties of the regression parameter estimates and the estimated baseline cumulative hazard functions are also studied.
Article
This article is directed toward situations where individuals can experience repeated events, and data on an individual consist of the number and occurrence times of events, along with concomitant variables. Methods of regression analysis are presented, based on Poisson process and proportional intensity assumptions. These include parametric and semi-parametric approaches to model fitting, model assessment, and the treatment of random effects. In addition, insight is gained as to the central role of Poisson and mixed Poisson regression analysis of counts in these methods, and as to the effects of unobserved heterogeneity on semi-parametric analyses.The methods in the article are based on the proportional intensity Poisson process model, for which an individual with given fixed covariate vector x has repeated events occur according to a nonhomogeneous Poisson process with intensity function λx(t) = λ0(t)exp(x′β). Estimation of β and the baseline intensity λ0(t) are considered when λ0(t) is specified up to a parameter θ, and close connections with the semiparametric methods due to Cox (e.g., Andersen and Gill 1982) are noted. Random effects are incorporated by representing λxi(t) as λ0(t)exp(αi, + xi′β), where the αi's are iid random variables. Details are examined and an example is given for the case in which exp(αi,) has a gamma distribution.