*This post is part of our Q&A series.*

A question from graduate students in our Fall 2019 offering of “Biostatistical Methods: Survival Analysis and Causality” at UC Berkeley:

**Question:**

Hi Mark,

First of all, I have doubts regarding the simultaneous confidence interval for Kaplan- Meier, since I am not necessarily interested in inference for a parameter. I would like > to know if the 95% confidence band for my KM estimator will hold using the same formula > we did in our

`R`

lab without covariates (taken from lectures).Secondly, I am interested in testing a hypothesis. I remember discussing in classes the > difference between testing the null hypothesis when each of my parameters is 0, is different than the null hypothesis that all of the parameters are 0. I understand that the first is trying to address each parameter individually, while the second one is testing whether all of the parameters, simultaneously, are significant.

For instance, for three covariates.

Case 1:

`$H_01: w1 = 0$`

`$H_02: w2 = 0$`

`$H_03: w3 = 0$`

Case 2:

`$H_0: w1 = w2 = w3 = 0$`

Therefore, is the simultaneous confidence interval calculating the second case only? For a KM estimator, can it be understood that the simultaneous confidence interval can be computed for one parameter across time, and for all of the parameters across time?

Thanks, F.M.

**Answer:**

Hi F.M.,

I believe your question concerns understanding of the construction and utility of a simultaneous confidence interval.

Let’s consider the case that we are interested in a survival function `$S(t)$`

.
Suppose that we have an estimator `$S_n(t)$`

which is asymptotically linear with
known influence curve `$IC_t(O)$`

, such as a Kaplan-Meier estimator for the simple
marginal right-censored data structure or a TMLE for longitudinal data structures
involving baseline covariates, treatment and potentially time-dependent covariates.

Then, `$S_n(t)-S(t)$`

across a collection of points `$t$`

will have influence curve the
stacked `$IC=(IC_t: t)$`

across these time points `$t$`

. This allows us to estimate the
correlation matrix of `$(\sqrt{n}(S_n(t)-S(t)): t)$`

by the correlation matrix `$\rho$`

of stacked influence curve `$IC=(IC_t: t)$`

. The `$0.95$`

quantile `$q(0.95)$`

of the
max of absolute value of a random draw from `$N(0,\rho)$`

consistently estimates the
`$0.95$`

quantile of `$\text{max}_t | S_n-S|(t)/SE(S_n(t))$`

, where `$SE(S_n(t))$`

is
an estimator of the standard error (SE) of `$S_n(t)$`

such as the square root of the
sample variance of `$IC_t$`

, divided by `$n$`

. In particular, ```
$S_n(t)\pm q(0.95)
SE(S_n(t))$
```

is now a simultaneous confidence band.

The multivariate normal distribution can be used as a null distribution in
multiple testing of `$H_0(t): S(t)=S_{H_0}(t)$`

for some hypothesized survivor function
`$S_{H_0}$`

, across collection of time points `$t$`

, thereby obtaining a powerful
multiple testing procedure taking into account the strong correlations between
`$S_n(t)$`

across `$t$`

. This improves strongly on Bonferroni adjustments or, in
general, methods that only utilize marginal `$p$`

-values (thus ignore correlations
between the test statistics).

Similarly, we could do all this for a contrast `$S_1-S_0(t)$`

given an estimator of
`$S_1-S_0$`

, so that it would provides inference for a difference of survival
functions w.r.t two populations or difference of treatment specific survival
functions. In general, the above story applies to any functional valued or vector
valued pathwise differentiable parameter `$\mu=(\mu(t); t)$`

for which an
asymptotically linear estimator `$\mu_n$`

is available (when doing functions we would
establish `$\sqrt{n}(\mu_n-\mu)$`

as random function converges in distribution to a
Gaussian process in function space, but these results generally hold under bounded
influence curves.

The multivariate normal distribution `$N(0,\rho)$`

can also be used to obtain a test
for an overall test `$H_0: S=S_{H_0}$`

across all `$t$`

. To start with the simultaneous
confidence interval could be used as a test by checking if `$S_{H_0}$`

is contained in
the simultaneous band. Alternatively, we could define as our test statistic the
Euclidean norm of `$\rho^{-1/2} (S_n-S_{H_0})/SE(S_n)$`

, where `$\rho^{1/2}$`

is the
square root of the correlation matrix `$\rho$`

, so that`$\rho{-1/2}$`

is its inverse.
Under `$H_0:S=S_{H_0}$`

, this Euclidean norm converges to the Euclidean norm of a
`$N(0,\mathbb{1}_{t \times t})$`

, and thus has a limit `${\chi}^2$`

distribution with as
many degrees of freedom as number of time points `$t$`

.

The main point is that we have an estimate of the sampling distribution of
`$(S_n-S)/ SE(S_n) \sim N(0,\Sigma)$`

, or better, we have its sample mean
approximation as the sample mean of the stacked influence curve, and that allows
us to understand the distribution of any functional of this (Functional delta
method), or any test statistic.

So the influence curve based inference can be used for single overall test about the whole curve $S$, or for testing multiple hypotheses under any type of Type I error control (see the book by Dudoit and van der Laan for the use of this multivariate normal distribution to obtain general multiple testing procedures, even based on step-down, step-up, and any kind of Type I error).

Best Wishes, Mark

**P.S.**, remember to write in to our blog at ```
vanderlaan (DOT) blog [AT]
berkeley (DOT) edu
```

. Interesting questions will be answered on our blog!