# Cox regression model

A regression model introduced by D.R. Cox [a4] and subsequently proved to be one of the most useful and versatile statistical models, in particular with regards to applications in survival analysis (cf. also Regression analysis).

Let $X _ { 1 } , \ldots , X _ { n }$ be stochastically independent, strictly positive random variables (cf. also Random variable), to be thought of as the failure times of $n$ different items, such that $X _ { k }$ has hazard function (i.e.

\begin{equation*} \mathsf{P} ( X _ { k } > t ) = \operatorname { exp } \left( - \int _ { 0 } ^ { t } u _ { k } ( s ) d s \right) \end{equation*}

for $t \geq 0$) of the form

\begin{equation*} u _ { k } ( t ) = \alpha ( t ) e ^ { z _ { k } ^ { T } ( t ) \beta }. \end{equation*}

Here, $\alpha$ is an unknown hazard function, the baseline hazard obtained if $\beta = 0$, and $\beta ^ { T } = ( \beta _ { 1 } , \dots , \beta _ { p } )$ is a vector of $p$ unknown regression parameters. The $z _ { k } ^ { T } ( t ) = ( z _ { k , 1 } ( t ) , \dots , z _ { k , p } ( t ) )$ denote known non-random vectors of possibly time-dependent covariates, e.g. individual characteristics of a patient referring to age, sex, method of treatment as well as physiological and other measurements.

The parameter vector $\beta$ is estimated by maximizing the partial likelihood [a5]

\begin{equation} \tag{a1} C ( \beta ) = \prod _ { j = 1 } ^ { n } \frac { \operatorname { exp } ( z _ { j } ^ { T } ( T _ { j } ) \beta ) } { \sum _ { k \in R _ { j } } \operatorname { exp } ( z _ { k } ^ { T } ( T _ { j } ) \beta ) }, \end{equation}

where $T _ { 1 } < \ldots < T _ { n }$ are the $X _ { k }$ ordered according to size, $Y _ { j } = i$ if it is item $i$ that fails at time $T _ { j }$, and $R _ { j } = \{ k : X _ { k } \geq T _ { j } \}$ denotes the set of items $k$ still at risk, i.e. not yet failed, immediately before $T _ { j }$. With this setup, the $j$th factor in $C ( \beta )$ describes the conditional distribution of $Y_{j}$ given $T _ { 1 } , \dots , T _ { j }$ and $Y _ { 1 } , \dots , Y _ { j - 1 }$.

For many applications it is natural to allow for, e.g., censorings (cf. also Errors, theory of) or truncations (the removal of an item from observation through other causes than failure) as well as random covariate processes $Z _ { k } ( t )$. Formally this may be done by introducing the counting processes $N _ { k } ( t ) = 1 _ { ( X _ { k } \leq t ,\, I _ { k } ( X _ { k } ) = 1 ) }$ registering the failures if they are observed, where is a $0 - 1$-valued stochastic process with $I _ { k } ( t ) = 1$ if item $k$ is at risk (under observation) just before time $t$. If $\mathcal{F} _ { t }$ denotes the $\sigma$-algebra for everything observed (failures, censorings, covariate values, etc.) on the time interval $[ 0 , t ]$, it is then required that $N _ { k }$ have $\mathcal{F} _ { t }$-intensity process

\begin{equation} \tag{a2} \lambda _ { k } ( t ) = \alpha ( t ) e ^ { Z _ { k } ^ { T } ( t ) \beta } I _ { k } ( t ), \end{equation}

i.e. $N _ { k } ( t ) - \int _ { 0 } ^ { t } \lambda _ { k } ( s ) d s$ defines a $\mathcal{F} _ { t }$-martingale (cf. also Martingale), while intuitively, for small $h > 0$, the conditional probability given the past that item $k$ will fail during the interval $] t , t + h ]$ is approximately $h \alpha ( t ) e ^ { Z _ { k } ^ { T } ( t ) \beta }$, provided $k$ is at risk at time $t$. For $\beta$ known, (a2) is then an example of Aalen's multiplicative intensity model [a1] with the integrated baseline hazard $A ( t ) = \int _ { 0 } ^ { t } \alpha ( s ) d s$ estimated by, for any $t$,

\begin{equation} \tag{a3} \hat { A } ( t | \beta ) = \int _ { ]0 , t] } \frac { 1 } { \sum _ { k = 1 } ^ { n } I _ { k } ( s - ) e ^ { Z _ { k } ^ { T } ( s - ) \beta } } d \overline { N } ( s ), \end{equation}

writing $\overline { N } = \sum _ { k } N _ { k }$ and where $s-$ signifies that it is the values of and $Z _ { k }$ just before the observed failure times that should be used. Since in practice $\beta$ is unknown, in (a3) one of course has to replace $\beta$ by the estimator $\widehat { \beta }$, still obtained maximizing the partial likelihood (a1), replacing $n$ by the random number of observed failures, replacing by $Z _ { k }$, and using $R _ { j } = \{ k : I _ { k } ( T _ { j } - ) = 1 \}$ with $T _ { j }$ now the $j$th observed failure. (Note that in contrast to the situation with non-random covariates described above, there is no longer an interpretation of the factors in $C ( \beta )$ as conditional distributions.)

Using central limit theorems for martingales (cf. also Central limit theorem: Martingale), conditions may be given for consistency and asymptotic normality of the estimators $\widehat { \beta }$ and $\widehat { A } ( t | \widehat { \beta } )$, see [a3].

It is of particular interest to be able to test for the effect of one or more covariates, i.e. to test hypothesis of the form $\beta _ { \text{l} } = 0$ for one or more given values of $\operatorname{l}$, $1 \leq 1 \leq p$. Such tests include likelihood-ratio tests derived from the partial likelihood (cf. also Likelihood-ratio test), or Wald test statistics based on the asymptotic normality of $\widehat { \beta }$. A thorough discussion of the tests in particular and of the Cox regression model in general is contained in [a2], Sect. VII.2; [a2], Sect. VII.3, presents methods for checking the proportional hazards structure assumed in (a2).

Refinements of the model (a2) include models for handling e.g. stratified data, Markov chains with regression structures for the transition intensities, etc. It should be emphasized that these models, including (a2), are only partially specified in the sense that with (a2) alone nothing much is said about the distributions of the $Z _ { k }$ or . This, in particular, makes it extremely difficult to use the models for, e.g., the prediction of survival times.

How to Cite This Entry:
Cox regression model. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Cox_regression_model&oldid=50588
This article was adapted from an original article by Martin Jacobsen (originator), which appeared in Encyclopedia of Mathematics - ISBN 1402006098. See original article