Introduction to reda through Examples

In this vignette, we introduce how to explore recurrent event data by mean cumulative function, and modeling the event counts of interest by gamma frailty model with the reda package through examples. Most functions in the package are S4 methods that produce S4 class objects. The details of function syntax and the produced objects are available in the package manual, which will thus not be covered in this vignette.

Introduction

Simulated recurrent event data

library(reda)
packageVersion("reda")
## [1] '0.5.4'

First of all, the sample recurrent event data we are going to use in the following examples is called simuDat, which contains totally 500 observations of 6 variables.

head(simuDat)
##   ID time event group    x1 gender
## 1  1    1     1 Contr -1.93 female
## 2  1   22     1 Contr -1.93 female
## 3  1   23     1 Contr -1.93 female
## 4  1   57     1 Contr -1.93 female
## 5  1  112     0 Contr -1.93 female
## 6  2  140     0 Treat -0.11 female
str(simuDat)
## 'data.frame':    500 obs. of  6 variables:
##  $ ID    : num  1 1 1 1 1 2 3 3 4 4 ...
##  $ time  : num  1 22 23 57 112 140 40 168 14 112 ...
##  $ event : int  1 1 1 1 0 0 1 0 1 0 ...
##  $ group : Factor w/ 2 levels "Contr","Treat": 1 1 1 1 1 2 1 1 1 1 ...
##  $ x1    : num  -1.93 -1.93 -1.93 -1.93 -1.93 -0.11 0.2 0.2 -0.43 -0.43 ...
##  $ gender: Factor w/ 2 levels "female","male": 1 1 1 1 1 1 1 1 1 1 ...

where

  • ID: Subjects identification (ID).
  • time: Event or censoring time.
  • event: Event indicator, 1 = event; 0 = censored.
  • group: Treatment group indicator.
  • x1: Continuous variable.
  • gender: Gender of subjects.

The dataset was simulated by thinning method (Lewis and Shedler 1979) and further processed for a better demonstration purpose. (Note that reda also provides functions for simulating survival data, and recurrent event data. See vignette("reda-simulate") for details.)

Data checking

The process’s ID, event times, event indicators or costs, time origins, and possible terminal events of the follow-up is specified in the function Recur(), which serves as the formula response and contains considerate data checking procedures for recurrent event data. See vignette("reda-Recur") for details.

Exploratory analysis

Nonparametric Mean Cumulative Function Estimates

The nonparametric mean cumulative function (MCF) estimates are widely utilized in exploring the trend of recurrent event data. MCF is also called cumulative mean function (CMF) in literature (see e.g., Lawless and Nadeau 1995). Let Ni(t) denote the number of events that occurred up to time t of process i. The MCF of Ni(t) denoted by Mi(t), is defined as follows: Mi(t) = 𝔼{Ni(t)}. For k independent processes having the same MCF, the Nelson-Aalen Estimator (Nelson 2003) is often used, which is defined as follows: $$ \hat{M}(t) = \int_0^t \frac{dN(s)}{\delta(s)}, $$ where $dN(s)=\sum_{i=1}^k dN_i(s)$, dNi(s) is the jump size of process i at time s, $\delta(s) = \sum_{i=1}^k \delta_i(s)$, δi(s) is the at-risk indicator of process i at time s. One variant is called the cumulative sample mean (CSM) function introduced by Cook and Lawless (2007), which assumes that δi(s) = 1 for s ≥ 0.

The nonparametric estimate of MCF at each time point does not assume any particular underlying model. The variance estimates at each time point can be computed by the Lawless and Nadeau method (Lawless and Nadeau 1995), Poisson process method, the bootstrap method (Efron 1979) with subjects as resampling units. For CSM, the cumulative sample variance (CSV) method can be used instead. The approximate confidence intervals are provided as well, which are constructed based on the asymptotic normality of the MCF estimates itself or logarithm of the MCF estimates.

The function mcf() is a generic function for the MCF estimates from a sample data or a fitted gamma frailty model (as demonstrated later). If a formula with Recur() as formula response is specified in function mcf(), the formula method for estimating the sample MCF will be called. The covariate specified at the right hand side of the formula should be either 1 or any “linear” combination of factor variables in the data. The former computes the overall sample MCF. The latter computes the sample MCF for each level of the combination of the factor variable(s) specified, respectively.

The valve-seat dataset in Nelson (1995) and the simulated sample data are used for demonstration as follows:

## Example 1. valve-seat data
valveMcf0 <- mcf(Recur(Days, ID, No.) ~ 1, data = valveSeats)

## Example 2. the simulated data
simuMcf <- mcf(Recur(time, ID, event) ~ group + gender,
               data = simuDat, subset = ID %in% seq_len(50))

After estimation, we may plot the sample MCF by function plot(), which returns a ggplot object so that the plot produced can be easily further customized by ggplot2. The legendname and legendLevels can be specified to easily customize the legend in the plot. Two examples are given as follows:

## overall sample MCF for valve-seat data in Nelson (1995)
plot(valveMcf0, conf.int = TRUE, mark.time = TRUE, addOrigin = TRUE, col = 2) +
    ggplot2::xlab("Days") + ggplot2::theme_bw()

## sample MCF for different groups (the default theme)
plot(simuMcf, conf.int = TRUE, lty = 1:4, legendName = "Treatment & Gender")

Note that in the first plot, the censoring times was marked on the step curve by specifying mark.time = TRUE and the time origins was included in the curve by specifying addOrigin = TRUE. In addition, the type and color of the line can be specified through lty and col, respectively.

As for the variance estimates, the Poisson process method assumes that the underlying counting process is a Poisson process and may underestimate the variance if the assumption cannot be justified. While, the Lawless and Nadeau method is more robust to departures from the Poisson process assumption. The nonparametric bootstrap method can be considered as well if an extra computational burden is not of concern. We may perform a quick comparison among the standard error estimates and the confidence intervals from these methods for the valve seats data as follows:

## Poisson process method
valveMcf1 <- mcf(Recur(Days, ID, No.) ~ 1, valveSeats, variance = "Poisson")

## bootstrap method (with 1,000 bootstrap samples)
set.seed(123)
valveMcf2 <- mcf(Recur(Days, ID, No.) ~ 1, valveSeats,
                 variance = "bootstrap", control = list(B = 1e3))

## comparing the standard error estimates
library(ggplot2)
ciDat <- rbind(cbind(valveMcf0@MCF, Method = "Lawless & Nadeau"),
               cbind(valveMcf1@MCF, Method = "Poisson"),
               cbind(valveMcf2@MCF, Method = "Bootstrap"))
ggplot(ciDat, aes(x = time, y = se)) +
    geom_step(aes(color = Method, linetype = Method)) +
    xlab("Days") + ylab("SE estimates") + theme_bw()

## comparing the confidence intervals
ggplot(ciDat, aes(x = time)) +
    geom_step(aes(y = MCF), color = "grey") +
    geom_step(aes(y = lower, color = Method, linetype = Method)) +
    geom_step(aes(y = upper, color = Method, linetype = Method)) +
    xlab("Days") + ylab("Confidence intervals") + theme_bw()

From the comparison, we may find that the SE estimates and the confidence intervals from the Lawless and Nadaeu method and the bootstrap method have a good agreement, while the Poisson process method gives slightly smaller SE estimates and a narrower confidence band. In practice, the Lawless and Nadeau method is suggested if it is hard to justify the Poisson process assumption.

Comparing two-sample MCFs

The function mcfDiff.test() is an implementation of the pseudo-score tests for comparing two-sample MCFs proposed by Cook, Lawless, and Nadeau (1996), while the function mcfDiff() gives the difference estimates and wraps the pseudo-score testing results from mcfDiff.test() (by default).

Suppose we are interested in comparing the two-sample MCFs between the treatment and control group in the simulated data. We may simply feed the mcf.formula object returned from the function mcf() to function mcfDiff() as follows:

## one sample MCF object of two groups
mcf0 <- mcf(Recur(time, ID, event) ~ group, data = simuDat)
(mcf_diff0 <- mcfDiff(mcf0))
## Call: 
## mcfDiff(mcf1 = mcf0)
## 
## Two-Sample Pseudo-Score Tests:
##                 Statistic Variance    Chisq DF Pr(>Chisq)   
## Constant Weight   52.5855 670.1014   4.1266  1   0.042214 * 
## Linear Weight     35.7687 158.9869   8.0472  1   0.004557 **
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
## 
## Variance Estimator: robust

Then what if the covariate group contains more than two levels? In that case, we may compute the sample MCF for each group, respectively, and feed both of the generated mcf.formula objects to mcfDiff as the first two arguments. Alternatively, an intuitive - method is available for comparing the difference between two mcf.formual objects, mcf1 and mcf2, returned from the mcf formula method, which means that the function call mcf1 - mcf2 is equivalent to mcfDiff(mcf1, mcf2). A simple example is given below.

## explicitly ask for the difference of two sample MCF
mcf1 <- mcf(Recur(time, ID, event) ~ 1, simuDat, group %in% "Contr")
mcf2 <- mcf(Recur(time, ID, event) ~ 1, simuDat, group %in% "Treat")
mcf1 - mcf2
## Call: 
## mcfDiff(mcf1 = mcf1, mcf2 = mcf2)
## 
## Two-Sample Pseudo-Score Tests:
##                 Statistic Variance    Chisq DF Pr(>Chisq)   
## Constant Weight   52.5855 670.1014   4.1266  1   0.042214 * 
## Linear Weight     35.7687 158.9869   8.0472  1   0.004557 **
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
## 
## Variance Estimator: robust

Similarly, a plot() method based on ggplot2 is available for visual comparison.

plot(mcf_diff0)

Gamma frailty model

Model with constant rate function

The default model when argument df, knots, and degree are not specified is gamma frailty model with (one piece) constant rate function, which is equivalent to negative binomial regression with the same shape and rate parameter in the gamma prior.

(constFit <- rateReg(Recur(time, ID, event) ~ group + x1, data = simuDat))
## Call: 
## rateReg(formula = Recur(time, ID, event) ~ group + x1, data = simuDat)
## 
## Coefficients of covariates: 
## groupTreat         x1 
## -0.6133832  0.3248600 
## 
## Frailty parameter:  0.5878151 
## 
## Boundary knots: 
## 0, 168
## 
## Coefficients of pieces:
## M-spline1 
##  5.124268

The function rateReg() returns rateReg object, which can be printed out by calling the object. (Internally, show() method for rateReg object is called.)

Model with piecewise constant rate function

When argument df or knots (at least one internal knot) is specified, the model becomes gamma frailty model with piecewise constant rate function or so-called HEART model (Fu, Luo, and Qu 2016) if argument degree is specified to be zero as default.

We may specify df and leave knots and degree as default. Then piecewise constant rate function will be applied and the number of pieces will equal df. The internal knots will be automatically specified at suitable quantiles of the covariate representing event and censoring time.

For example, two pieces’ constant rate function can be simply specified by setting df = 2. The internal knot will be the median time of all the event and censoring time. Also, we can fit the models on the first 50 subjects by specifying argument subset.

# two pieces' constant rate function
(twoPiecesFit <- rateReg(Recur(time, ID, event) ~ group + x1, df = 2,
                         data = simuDat, subset = ID %in% 1:50))
## Call: 
## rateReg(formula = Recur(time, ID, event) ~ group + x1, data = simuDat, 
##     subset = ID %in% 1:50, df = 2)
## 
## Coefficients of covariates: 
## groupTreat         x1 
## -0.7884707  0.3388577 
## 
## Frailty parameter:  0.6917509 
## 
## Internal knots: 
## 102
## 
## Boundary knots: 
## 0, 168
## 
## Coefficients of pieces:
## M-spline1 M-spline2 
##  3.464584  3.226911

In the example shown above, the internal knots is set automatically to be 102 and the baseline rate function is two pieces’ constant.

If internal knots are specified explicitly, the df will be neglected even if it is specified. An example of model with six pieces’ constant rate function is given as follows:

(piecesFit <- rateReg(Recur(time, ID, event) ~ group + x1, data = simuDat,
                      knots = seq(from = 28, to = 140, by = 28)))
## Call: 
## rateReg(formula = Recur(time, ID, event) ~ group + x1, data = simuDat, 
##     knots = seq(from = 28, to = 140, by = 28))
## 
## Coefficients of covariates: 
## groupTreat         x1 
## -0.6371381  0.3062645 
## 
## Frailty parameter:  0.5868205 
## 
## Internal knots: 
## 28, 56, 84, 112, 140
## 
## Boundary knots: 
## 0, 168
## 
## Coefficients of pieces:
## M-spline1 M-spline2 M-spline3 M-spline4 M-spline5 M-spline6 
## 0.7212985 0.8535333 0.5770422 0.8558860 1.0556631 1.3405290

Model with spline rate function

When argument degree is specified to be a positive integer, the baseline rate function is fitted by splines. The type or flavor of the splines can be specified by argument spline. The available option for spline are bSplines for B-splines and mSplines for M-splines. (See R package spline2 for details about the spline functions used internally.) A partial matching on names is allowed.

For example, one may want to fit the baseline rate function by a cubic spline with two internal knots. Then we may explicitly specify degree = 3 and knots to be a length-two numeric vector. Or we may simply specify degree = 3 and df = 6 Then the internal knots will be automatically specified at suitable quantiles of the covariate representing event and censoring time. Generally speaking, the degree of freedom of spline (or the number of spline bases) equals the summation of the number of internal knots and the degree of each spline base, plus one if intercept is included in spline bases.

## internal knots are set as 33% and 67% quantiles of time variable
(splineFit <- rateReg(Recur(time, ID, event) ~ group + x1, data = simuDat,
                      df = 6, degree = 3, spline = "mSplines"))
## Call: 
## rateReg(formula = Recur(time, ID, event) ~ group + x1, data = simuDat, 
##     df = 6, degree = 3, spline = "mSplines")
## 
## Coefficients of covariates: 
## groupTreat         x1 
## -0.6360235  0.3057244 
## 
## Frailty parameter:  0.5882684 
## 
## Internal knots: 
## 73.33333, 138
## 
## Boundary knots: 
## 0, 168
## 
## Coefficients of spline bases:
## M-spline1 M-spline2 M-spline3 M-spline4 M-spline5 M-spline6 
## 0.3431193 1.5542060 0.1478426 1.7829521 1.3690238 0.1828846
## or internal knots are expicitly specified
(splineFit <- rateReg(Recur(time, ID, event) ~ group + x1, data = simuDat,
                      spline = "bSp", degree = 3L, knots = c(56, 112)))
## Call: 
## rateReg(formula = Recur(time, ID, event) ~ group + x1, data = simuDat, 
##     spline = "bSp", degree = 3L, knots = c(56, 112))
## 
## Coefficients of covariates: 
## groupTreat         x1 
## -0.6353585  0.3061942 
## 
## Frailty parameter:  0.5884794 
## 
## Internal knots: 
## 56, 112
## 
## Boundary knots: 
## 0, 168
## 
## Coefficients of spline bases:
## M-spline1 M-spline2 M-spline3 M-spline4 M-spline5 M-spline6 
## 0.2532251 1.1505032 0.7425349 1.0091695 1.7681150 0.4446083

Summary of model fits

A brief summary of the fitted model is given by show() method as shown in the previous examples. Further, summary() method for rateReg object provides a more specific summary of the model fitted. For instance, the summary of the models fitted in section of model fitting can be called as follows:

summary(constFit)
## Call: 
## rateReg(formula = Recur(time, ID, event) ~ group + x1, data = simuDat)
## 
## Coefficients of covariates: 
##                coef exp(coef) se(coef)       z Pr(>|z|)  
## groupTreat -0.61338   0.54152  0.28527 -2.1502  0.03154 *
## x1          0.32486   1.38384  0.16641  1.9522  0.05092 .
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
## 
## Parameter of frailty: 
##         parameter        se
## Frailty 0.5878151 0.1102496
## 
## Boundary knots:
## 0, 168
## 
## Degree of spline bases: 0 
## 
## Coefficients of spline bases:
##             coef se(coef)
## M-spline1 5.1243   0.9632
## 
## Loglikelihood:  -1676.421
summary(piecesFit, showCall = FALSE)
## 
## Coefficients of covariates: 
##                coef exp(coef) se(coef)       z Pr(>|z|)  
## groupTreat -0.63714   0.52880  0.28590 -2.2285  0.02585 *
## x1          0.30626   1.35834  0.16692  1.8348  0.06654 .
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
## 
## Parameter of frailty: 
##         parameter        se
## Frailty 0.5868205 0.1099042
## 
## Internal knots: 
## 28, 56, 84, 112, 140
## 
## Boundary knots:
## 0, 168
## 
## Degree of spline bases: 0 
## 
## Coefficients of spline bases:
##              coef se(coef)
## M-spline1 0.72130   0.1601
## M-spline2 0.85353   0.1844
## M-spline3 0.57704   0.1334
## M-spline4 0.85589   0.1869
## M-spline5 1.05566   0.2271
## M-spline6 1.34053   0.2948
## 
## Loglikelihood:  -1663.792
summary(splineFit, showCall = FALSE, showKnots = FALSE)
## 
## Coefficients of covariates: 
##                coef exp(coef) se(coef)       z Pr(>|z|)  
## groupTreat -0.63536   0.52975  0.28550 -2.2254  0.02605 *
## x1          0.30619   1.35825  0.16651  1.8389  0.06593 .
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
## 
## Parameter of frailty: 
##         parameter        se
## Frailty 0.5884794 0.1103127
## 
## Degree of spline bases: 3 
## 
## Coefficients of spline bases:
##              coef se(coef)
## M-spline1 0.25323   0.1056
## M-spline2 1.15050   0.3261
## M-spline3 0.74253   0.4156
## M-spline4 1.00917   0.4539
## M-spline5 1.76811   0.4657
## M-spline6 0.44461   0.1825
## 
## Loglikelihood:  -1663.285

The summary includes the function call, estimated covariate coefficients, estimated parameter of frailty variable, internal knots (if exist), boundary knots, degree of spline bases if splines are applied, coefficients of rate function bases (pieces), and log-likelihood of the model fitted. Outputs of function call or knots, may be suppressed by specifying argument showCall or showKnots to be FALSE, respectively, in summary() method, which would be especially useful for a relatively concise summary in a reproducible report using Rmarkdown, etc.

What’s more, the corresponding coef() and confint() method for point estimates and confidence interval for covariate coefficients are provided as well. Let’s take the fitted model with spline rate function as an example.

## point estimates of covariate coefficients
coef(splineFit)
## groupTreat         x1 
## -0.6353585  0.3061942
## confidence interval for covariate coefficients
confint(splineFit, level = 0.95)
##                   2.5%       97.5%
## groupTreat -1.19492720 -0.07578985
## x1         -0.02015825  0.63254670

Model selection

Two handy functions are provided for model selection. We may compare and select the models with different baseline rate function based on Akaike Information Criterion (AIC) by function AIC() or Bayesian Information Criterion (BIC) by function BIC(). A friendly warning will be thrown out if the numbers of observation were different in the model comparison by AIC.

AIC(constFit, piecesFit, splineFit)
##           df      AIC
## constFit   4 3360.843
## piecesFit  9 3345.585
## splineFit  9 3344.570
BIC(constFit, piecesFit, splineFit)
##           df      BIC
## constFit   4 3377.701
## piecesFit  9 3383.516
## splineFit  9 3382.501

Baseline rate function

Function baseRate() produces baseRate.rateReg object representing the estimated baseline rate function for a fitter model. An associated plot() method is available. For example, the baseline rate function and its confidence band estimated by cubic splines can be plotted as follows:

baseRateObj <- baseRate(splineFit)
plot(baseRateObj, conf.int = TRUE)

Estimated MCF from the fitted model

If rateReg object is supplied to function mcf(), the method for rateReg is called, which returns the estimated baseline MCF from the fitted model if newdata is not specified in the function. The example estimating and plotting the baseline MCF from the fitted model with piecewise constant rate function is shown as follows:

piecesMcf <- mcf(piecesFit)
plot(piecesMcf, conf.int = TRUE, col = "blueviolet") + xlab("Days")

The argument newdata allows one to estimate the MCF for a given dataset instead of the baseline MCF. If newdata is specified, the data frame should have the same column names as the covariate names appearing in the formula of original fitting. The MCF will be estimated for each unique row in the data frame and its confidence intervals are constructed based on Delta-method.

In addition, we may specify the name for grouping each unique row and the levels of each group through groupName and groupLevels, respectively. For example, we may specify groupName = "Gender" and groupLevels = c("Male", "Female") for estimation of different gender groups.

As the last two examples in this vignette, we estimate the MCF from fitted model with spline rate function for the different treatment groups and plot the estimated MCFs and their confidence intervals correspondingly.

newDat <- data.frame(x1 = c(0, 0), group = c("Treat", "Contr"))
estmcf <- mcf(splineFit, newdata = newDat, groupName = "Group",
              groupLevels = c("Treatment", "Control"))
plot(estmcf, conf.int = TRUE, col = c("royalblue", "red"), lty = c(1, 5)) +
    ggtitle("Control vs. Treatment") + xlab("Days")

The data frame containing the MCF estimates is stored in the slot named MCF. So it is not hard to make further customization to the MCF plot.

plot(estmcf) +
    geom_ribbon(data = estmcf@MCF, alpha = 0.2,
                aes(x = time, ymin = lower, ymax = upper, fill = Group)) +
    ggtitle("Control vs. Treatment") + xlab("Days")

Reference

Cook, Richard J, and Jerald Lawless. 2007. The Statistical Analysis of Recurrent Events. Springer Science & Business Media.
Cook, Richard J, Jerald F Lawless, and Claude Nadeau. 1996. “Robust Tests for Treatment Comparisons Based on Recurrent Event Responses.” Biometrics 52 (2): 557–71.
Efron, B. 1979. Bootstrap Methods: Another Look at the Jackknife.” The Annals of Statistics 7 (1): 1–26.
Fu, Haoda, Junxiang Luo, and Yongming Qu. 2016. “Hypoglycemic Events Analysis via Recurrent Time-to-Event (HEART) Models.” Journal of Biopharmaceutical Statistics 26 (2): 280–98.
Lawless, Jerald F, and Claude Nadeau. 1995. “Some Simple Robust Methods for the Analysis of Recurrent Events.” Technometrics 37 (2): 158–68.
Lewis, P A, and G S Shedler. 1979. “Simulation of Nonhomogeneous Poisson Processes by Thinning.” Naval Research Logistics Quarterly 26 (3): 403–13.
Nelson, Wayne B. 1995. “Confidence Limits for Recurrence Data-Applied to Cost or Number of Product Repairs.” Technometrics 37 (2): 147–57.
———. 2003. Recurrent Events Data Analysis for Product Repairs, Disease Recurrences, and Other Applications. Vol. 10. SIAM.