Abstract
In real life, more often experimental units are susceptible to more than one risk factor. Moreover, some experimental units may not fail even if they are observed over a long period of time. In statistical analysis, competing risks models handle the first scenario while cure rate models have been introduced to analyze the long-term survivors in the population. In this paper, we consider a cure rate model when the failure of a unit can be due to either of the two competing causes. To analyze the competing risk data in presence of long-term survivors, we consider the latent failure times approach introduced by Cox (J R Stat Soc Ser B (Methodol) 21(2):411–421, 1959). The latent failure times are assumed to follow exponential distributions, and they are independently distributed. Under this setup, a random censoring scheme is applied and the observed data consist of either censored times or actual failure times along with the cause of failures. We derive the maximum likelihood estimators (MLEs) using the expectation-maximization (EM) algorithm based on the missing value principle. As the overall survival function is not a proper survival function, the asymptotic behavior of the MLEs is not immediate. We provide the sufficient conditions for the existence, uniqueness, consistency and the asymptotic normality of the MLEs. Monte Carlo simulations are performed to support the theoretical validation numerically. For illustrative purposes, we have analyzed one real dataset, and the results are quite satisfactory.
This is a preview of subscription content, access via your institution.
References
- 1.
Balakrishnan N, Pal S (2013) COM Poisson cure rate models and associated likelihood based inference with exponential and Weibull lifetimes. In: Frenkel IB, Karagrigoriou A, Lisnianski A et al (eds) Probabilistic models and statistical inference, applied reliability engineering and risk analysis. Wiley, Chichester, pp 308–348
- 2.
Berkson J, Gage RP (1952) Survival curve for cancer patients following treatment. J Am Stat Assoc 47(259):501–515
- 3.
Boag JW (1949) Maximum likelihood estimates of the proportion of patients cured by cancer therapy. J R Stat Soc Ser B (Methodol) 11(1):15–53
- 4.
Crowder M (2001) Classical competing risks model. Chapman & Hall/CRC, New York
- 5.
Cox DR (1959) The analysis of exponentially distributed lifetimes with two types of failures. J R Stat Soc Ser B (Methodol) 21(2):411–421
- 6.
Dunsmuir W, Tweedie R, Flack L, Mengersen K (1989) Modelling of transitions between employment states for young Australians. Aust N Z J Stat 31(1):165–196
- 7.
Farewell VT (1982) The use of mixture models for the analysis of survival data with long-term survivors. Biometrics 38:1041–1046
- 8.
Ghitany ME, Maller RA (1992) Asymptotic results for exponential mixture models with long-term survivors. Stat A J Theor Appl Stat 23(4):321–336
- 9.
Ghitany ME, Maller RA, Zhou S (1994) Exponential mixture models with long-term survivors and covariates. J Multivar Anal 49(2):218–241
- 10.
Gallardo DI, Gómez YM, de Castro M (2018) A flexible cure rate model based on the polylogarithm distribution. J Stat Comput Simul 88(11):2137–2149
- 11.
Goldman AI (1984) Survivorship analysis when cure is a possibility: a Monte Carlo study. Stat Med 3(2):153–163
- 12.
Greenhouse JB, Wolfe RA (1984) A competing risks derivation of a mixture model for the analysis of survival data. Commun Stat Theory Methods 13(25):3133–3154
- 13.
Gulland JA (1955) On the estimation of population parameters from marked members. Biometrika 42(1–2):269–270
- 14.
Kannan N, Kundu D, Nair P, Tripathi RC (2010) The generalized exponential cure rate model with covariates. J Appl Stat 37(10):1625–1636
- 15.
Kundu D (2004) Parameter estimation for partially complete time and type of failure data. Biom J J Math Methods Biosci 46(2):165–179
- 16.
Maltz DM, McCleary R (1977) The mathematics of behavioral change: recidivism and construct validity. Eval Q 1:421–38
- 17.
Maltz DM (1984) Recidivism. Academic Press, Florida
- 18.
Prentice RL, Kalbfleish JD, Peterson JAV, Flurnoy N, Farewell VT, Breslow NE (1978) The analysis of failure times in presence of competing risks. Biometrics 34:541–554
- 19.
Self SG, Liang KL (1987) Asymptotic properties of maximum likelihood estimators and likelihood ratio tests under nonstandard conditions. J Am Stat Assoc 82:605–610
- 20.
Wiangnak P, Pal S (2018) Gamma lifetimes and associated inference for interval-censored cure rate model with COM Poisson competing cause. Commun Stat Theory Methods 47(6):1491–1509
Acknowledgements
The authors would like to thank the unknown reviewers for their constructive comments which have helped to improve the manuscript significantly.
Author information
Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
In the Appendix, we provide the proofs of Theorems 4.1, 4.2 and 4.3. To prove these Theorems, we need a series of lemmas and some preliminaries. We provide a flowchart to show how these Lemmas have been used to prove the Theorems.
Result | Prerequisite results |
---|---|
Fisher information matrix | Lemmas 6.1 and 6.2. |
Lemma 6.3 | Fisher information matrix |
Lemma 6.6 | Lemmas 6.4 and 6.5 |
Lemma 6.7 | Lemma 6.6 |
Lemma 6.8 | Lemma 6.7 |
Lemma 6.10 | Lemmas 6.6 and 6.9 |
Theorem 4.1 | Lemmas 6.3 and 6.7 |
Theorem 4.2 | Lemmas 6.3 and 6.8 |
Theorem 4.3 | Lemmas 6.3 and 6.10 |
The following notations have been used throughout this section.
Let \(\{ X_n \}\) be a sequence of random variables.
-
\(X_n=o_p(1)\) implies \(\lim \limits _{n \rightarrow \infty } P(|X_n|>\epsilon )=0\) \(\forall \epsilon >0.\)
-
\(X_n=O_p(1)\) implies for any \(\epsilon > 0,\) there exists a finite \(M > 0\) and a finite \(N > 0\) such that \(P ( | X_n |> M ) < \epsilon , \ \forall n > N\).
-
\(\lambda _{\mathrm{{min}}}\{\mathbf {A}\}\) is the minimum eigen value of the matrix \(\mathbf {A}.\)
-
The observed Fisher information matrix is
$$\begin{aligned} {\mathbf {F}}_n(\varvec{\theta })=-\dfrac{\partial {\mathbf {S}}_n (\varvec{\theta })}{\partial \varvec{\theta } }, \end{aligned}$$(11)where \({\mathbf {S}}_n (\varvec{\theta })\) is defined in (2).
-
The expected Fisher information matrix about \(\varvec{\theta _0}\) is
$$\begin{aligned} {\mathbf {D}}_n(\varvec{\theta _0}) =E[{\mathbf {F}}_n(\varvec{\theta _0})]=E[{\mathbf {S}}_n(\varvec{\theta _0}){\mathbf {S}}_n(\varvec{\theta _0})^{\top }]=Cov({\mathbf {S}}_n(\varvec{\theta _0})). \end{aligned}$$(12) -
We define the region \(\mathbf {N}_n(\varvec{\theta _0})\) for fixed \(A \ge 1,\) as
$$\begin{aligned} \mathbf {N}_n(\varvec{\theta _0})=\{ \varvec{\theta } :\varvec{ \theta } \in \Theta ,\,\, (\varvec{\theta } - \varvec{\theta _0} )^{\top } {\mathbf {D}}_n(\varvec{\theta }_o) (\varvec{\theta } - \varvec{\theta _0} ) \le A^2 \}. \end{aligned}$$(13)
Elements of Observed Information Matrix
Note that the observed information matrix as defined in (11) can be written as
where
Here,
Fisher Information Matrix
Note that we can write the Fisher information matrix as defined in (12), as
where \( \mathbf {{\mathscr {D}}}_{i}(\varvec{\theta }_o) = \begin{bmatrix} d_{i}^{11}(\varvec{\theta }_o) &{} d_{i}^{12}(\varvec{\theta }_o)&{} d_{i}^{13}(\varvec{\theta }_o)\\ d_{i}^{21}(\varvec{\theta }_o) &{} d_{i}^{22}(\varvec{\theta }_o) &{} d_{i}^{23}(\varvec{\theta }_o)\\ d_{i}^{31}(\varvec{\theta }_o) &{} d_{i}^{32}(\varvec{\theta }_o) &{} d_{i}^{33}(\varvec{\theta }_o)\\ \end{bmatrix} \) and \( d_i^{kl}(\varvec{\theta }_o)=E[f_i^{kl}(\varvec{\theta _0})],\, k,l=1,2,3.\) The expressions for the elements of \(\mathbf {{\mathscr {D}}}_{i}(\varvec{\theta _0}) \) can be obtained using Lemma 6.1.
Lemma 6.1
Let R(.) be a non-negative measurable function on \(\big [0, \infty \big )\). Then, for all \(1\le i\le n,\, \lambda _{10}>0, \, \lambda _{20}>0\) and \(p_0 \in (0,1 \big ],\)
provided the expectation on the right exists.
Proof
For any \(u>0\),
Since \(Y_{i}\) is independent of \(C_{i},\) we have
Hence,
By Fubini’s theorem, we have
provided the expectation on the right-side exists. \(\square \)
The following Lemma 6.2 summarizes the expressions for the elements of \(\mathbf {{\mathscr {D}}}_{i}(\varvec{\theta _0}) \).
Lemma 6.2
For all \(1\le i\le n,\, \lambda _{10}>0, \, \lambda _{20}>0\) and \(p_0 \in (0,1 ],\) the following are finite:
Proof
The proof follows easily applying Lemma 6.1. \(\square \)
Remarks
Type-I censoring (termination of the experiment at time \(\tau >0\)) turns out to be a particular case when \(P(C_i=\tau )=1 ,\, i=1,2,\ldots ,n.\) Then, all the expectations will exist since
will always exist.
Lemma 6.3
If the censored times \(C_i\)’s for \(i=1,\ldots ,n,\) are iid random variables whose common distribution does not degenerate at origin, then
will hold, i.e., there exists some \(\eta >0\) such that \(\lambda _{\mathrm{min}}\{{\mathbf {D}}_n(\varvec{\theta _0})\} \ge n\eta \) for n large enough.
Proof
When \(C_i\)’s are iid random variables, \(\lambda _{\mathrm{min}}\{ {\mathbf {D}}_n (\varvec{\theta _0})\} = n \lambda _{\mathrm{min}} \{ \mathbf {{\mathscr {D}}}_{1}(\varvec{\theta _0}) \}.\) Therefore, it is enough to show that \( \mathbf {{\mathscr {D}}}_{1}(\varvec{\theta _0}) \) is positive definite.
From (21), (22), (23), it is clear that \(d^{11}_1(\varvec{\theta _0})\), \(d^{22}_1(\varvec{\theta _0})\) and \(d^{33}_1(\varvec{\theta _0})\) all are 0 if and only if \(C_1=0\). This is true, since, for \(\lambda _{10}>0\), \(\lambda _{20}>0,\) \(p_0 \in (0,1),\) and \(C\ge 0\), the following functions of C defined as
are non-negative, increasing, see for example Ghitany and Maller [8]) and at \(C=0,\)
\(g_{i}(C)=0,\,i=1,2,3\). Thus, from (12), we have
where \(s_{ij}(\varvec{\theta _0})'s\) are defined in (3), (4) and (5). We will prove the result by contradiction. Let us first assume that the determinant of \( \mathbf {{\mathscr {D}}}_1(\varvec{\theta _0})\) is 0. Then, \(s_{11}(\varvec{\theta _0}),\) \(s_{12}(\varvec{\theta _0})\) and \(s_{13}(\varvec{\theta _0})\) are almost surely linearly related. We can write
for some \(a_1,a_2, a_3.\) When \(C_1\ne 0\) almost surely, then \(P(Z_1=1 )>0\) and \(P(\Delta _1=1) >0.\) Now plugging in \(z_i=1\) and \(\delta _i=1\) in (30), we obtain \(t^{*}_{11}(a_1+a_2)=\frac{a_1}{\lambda _{10}}+\frac{a_3}{p_0},\) almost surely on a set, say \({\mathcal {B}}\), with positive probability. This implies that on \({\mathcal {B}}\), \(t^{*}_{11}\) is degenerate which is impossible, since on \({\mathcal {B}}\), \(t^{*}_{11}\) is exponential. Therefore, \(\mathbf {{\mathscr {D}}}_1 \) is non-singular, hence positive definite. \(\square \)
Lemma 6.4
For \(\lambda _{10}, \lambda _{20} >0\) and \(p_0 \in (0,1)\), \( | f^{rs}_i( \varvec{\theta _0} ) | <M(\varvec{\theta _0} )\) for all \(i=1, \ldots ,n\) and \(r,s=1,2,3\) where \(f^{rs}_i( \varvec{\theta } )'s\) are defined in (15), (16), (17), (18), and (19) and \(M(\varvec{\theta _0} )\) is a positive finite quantity depending on \(\varvec{\theta _0}\).
Proof
At \(\varvec{\theta }= \varvec{\theta _0}\), \(f^{rs}_i(\varvec{\theta _0})\)’s are continuous function of \(t_i\) for \(r,s=1,2,3.\) Again for all \(i=1, \ldots ,n\),
\(z_i\) and \(\delta _i\) being binary variables all the \(f^{rs}_i(\varvec{\theta _0})\)’s have finite limit at 0 and \(\infty \). It implies that \( |f^{rs}_i(\varvec{\theta _0} )| < M^{rs}(\varvec{\theta _0} )\) where \( M^{rs}(\varvec{\theta _0})\) is a positive quantity for \(r,s=1,2,3.\)
Let \(M(\varvec{\theta _0}) =\max _{\begin{array}{c} r,s \end{array}} M^{rs}(\varvec{\theta _0})\). Therefore, \( |f^{rs}_i(\varvec{\theta _0})| < M( \varvec{\theta _0} )\) for all \(r,s=1,2,3\) and \(i = 1, \ldots ,n.\) \(\square \)
Lemma 6.5
For any \(\varvec{\theta } \in \mathbf {N}_n(\varvec{\theta _0})\) and (26) holds, \( | w^{rs}_{ij}(\varvec{\theta }) |<M_1(\varvec{\theta }) <M_2(\varvec{\theta _0}) \) for all \(i=1, \ldots ,n\) and \(j, r,s=1,2,3,\) where, \(w_i^{rs}(\varvec{\theta })= \frac{\partial f_i^{rs}(\varvec{\theta } )}{\partial \varvec{\theta } }=( w_{i1}^{rs}( \varvec{\theta }), w_{i2}^{rs}( \varvec{\theta }), w_{i3}^{rs}( \varvec{\theta }) ).\) Here, \(M_1(\varvec{\theta })\) is a positive finite quantity depending on \(\varvec{\theta }\) and \(M_2( \varvec{\theta _0} )\) is a positive finite quantity depending on \(\varvec{\theta _0} \).
Proof
For fixed \(\varvec{\theta }\), and for all \(i=1, \ldots ,n\), \(j=1,2,3\), each of \(w^{rs}_{ij}(\varvec{\theta })\) is a continuous function of \(t_i\) with finite limit at 0 and \(\infty .\) Hence, for each \(j,r,s=1,2,3\), there exists a positive quantity \(M^{rs}_{j} (\varvec{\theta })\) such that \( | w^{rs}_{ij}(\varvec{\theta } )| \le M^{rs}_{j} (\varvec{\theta })\). For \(\lambda _{10}, \lambda _{20} >0\) and \(p_0 \in (0,1)\), each \(M^{rs}_{j} (\varvec{\theta })\) is continuous function of \(\varvec{\theta }\). Again when \(\varvec{\theta } \in \mathbf {N}_n(\varvec{\theta _0})\), using (26) we obtain
where \(\eta >0\) is a suitable constant and this implies that for each \(j,r,s=1,2,3\) there exist a quantity \(N^{rs}_{j} ( \varvec{\theta _0})\) such that \(M^{rs}_{j} (\varvec{\theta }) \le N^{rs}_{j} (\varvec{\theta _0} )\). Let \(\max _{\begin{array}{c} j,r,s \end{array}} M^{rs}_{j}( \varvec{\theta } )=M_1( \varvec{\theta } )\) and \(\max _{\begin{array}{c} j,r,s \end{array}} N^{rs}_{j}(\varvec{\theta _0} )=M_2( \varvec{\theta _0} )\). Therefore, for each \(i=1,\ldots ,n\) and for each \(j,r,s=1,2,3,\) \(| w^{rs}_{ij}( \varvec{\theta } ) |< M_1(\varvec{\theta } ) < M_2( \varvec{\theta _0})\). \(\square \)
Lemma 6.6
For \(\lambda _{10}, \lambda _{20} >0\) and \(p_0 \in (0,1),\) if the condition (26) holds, \({\mathbf {D}}_n(\varvec{\theta _0})\) is a positive definite matrix and
will hold along with \(\lambda _{\mathrm{min}} \{{\mathbf {D}}_n(\varvec{\theta _0})\} \rightarrow \infty \) as \(n \rightarrow \infty \).
Proof
Condition (26) implies that there exist a \(\eta >0\) such that \(\lambda _{\mathrm{min}}\{{\mathbf {D}}_n(\varvec{\theta _0})\} \ge n\eta \) for n large enough. Therefore, all the eigenvalues of \({\mathbf {D}}_n(\varvec{\theta _0})\) are positive for large n which implies \({\mathbf {D}}_n(\varvec{\theta _0})\) is a positive definite matrix for large n. Now, we can define \({\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta _0})\), the symmetric square root of \({\mathbf {D}}^{-1}_n(\varvec{\theta _0}).\)
Now to show that (31) will hold, we decompose \({\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta _0}) {\mathbf {F}}_n( \varvec{\theta }) {\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta _0})\) as
Hence, it is sufficient to prove that for any unit vector \({\mathbf {u}}\) and for any \(\varvec{\theta } \in \mathbf {N}_n(\varvec{\theta _0})\) ,
To show \(e_n^{(1)}( \varvec{\theta _0}) = o_{p}(1)\), we need the results of Lemma 6.4. Now, \(f^{rs}_i(\varvec{\theta _0}) \) is a sequence of independent random variables and based on Lemma 6.4, we have \( | f^{rs}_i(\varvec{\theta _0}) | <M(\varvec{\theta _0}),\) \(\,r,s=1,2,3.\) Hence,
Therefore, by the weak law of large numbers, for all \(r,s=1,2,3,\)
For any unit vector \({\mathbf {u}}\), let \({\mathbf {u}}_n = \frac{ {\mathbf {D}}^{-\frac{1}{2}}_n( \varvec{\theta _0}) {\mathbf {u}} }{\sqrt{{\mathbf {u}}^{\top }{\mathbf {D}}^{-1}_n( \varvec{\theta _0}) {\mathbf {u}} } }\). Then, \({\mathbf {u}}_n \) is also a unit vector and hence
for any matrix \(M_n=(m_{n}^{rs}), \, r,s=1,2,3\). Thus, we have
Using the result in (26), we have for \(\eta >0\), \({\mathbf {u}}^{\top } {\mathbf {D}}^{-1}_n(\varvec{\theta _0}) {\mathbf {u}} \le \lambda _{max}\{{\mathbf {D}}^{-1}_n(\varvec{\theta _0}) \} \le \frac{1}{n\eta }.\) Then,
To show \(e_n^{(2)}(\varvec{\theta }) \overset{P}{\rightarrow } 0 \), we need the following Taylor series expansion of \(f_i^{rs}(\varvec{\theta } )\) about \(\varvec{\theta _0}. \)
where \(\varvec{{\tilde{\theta }}}\) is any point between \(\varvec{\theta }\) and \(\varvec{\theta _0}\), \(w_i^{rs}(\varvec{\theta })= \frac{\partial f_i^{rs}(\varvec{\theta } )}{\partial \varvec{\theta } }=( w_{i1}^{rs}( \varvec{\theta }), w_{i2}^{rs}( \varvec{\theta }), w_{i3}^{rs}( \varvec{\theta }) ),\ \,i=1,2,\ldots ,n,\,r,s=1,2,3.\) For fixed (r, s), \(w_i^{rs}(\varvec{\theta })\) is a sequence of independent random vectors. The expressions of the elements of \(w_i^{rs}(\varvec{\theta })\) are as follows.
For further development we need the result of Lemma 6.5. As a result of Lemma 6.5,
where \(\Vert .\Vert \) denotes the norm of a vector.
Also, when \( \varvec{\theta } \in \mathbf {N}_n( \varvec{\theta _0} )\), \({| \varvec{\theta } - \varvec{\theta _0 } |}^2 \le \frac{A^2}{\lambda _{\mathrm{min}}\{{\mathbf {D}}_n(\varvec{\theta _0 })\}} \le \frac{A^2}{n\eta } \rightarrow 0\) as \(n \rightarrow \infty \).
Since \(\varvec{ {\tilde{\theta }}} \in \mathbf {N}_n( \varvec{\theta _0} )\), it follows that
Therefore,
This concludes the proof of Lemma (6.6). \(\square \)
Lemma 6.7
For \(\lambda _{10}, \lambda _{20} >0\) and \(p_0 \in (0,1),\) if condition (26) holds, a unique MLE of \(\varvec{\theta }_o\), say \(\widehat{\varvec{\theta _n}}\), will exist in \(\mathbf {N}_n(\varvec{\theta }_o)\) with probability 1 as \(n \rightarrow \infty \).
Proof
If the condition (26) holds, from Lemma 6.6, \({\mathbf {D}}_n(\varvec{\theta _0})\) is positive definite matrix and a symmetric square root of \({\mathbf {D}}^{-1}_n(\varvec{\theta _0})\) will always exist. Let \({\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta _0})\) be the symmetric square root of \({\mathbf {D}}^{-1}_n(\varvec{\theta _0})\). Further consider the Taylor-series expansion of the log-likelihood function \({\mathbf{l}}_n(\varvec{\theta })\) about the true parameter value \(\varvec{\theta _0}\) as follows:
where \(\varvec{{\bar{\theta }}}\) is any point on the line segment joining \(\varvec{\theta _0}\) and \(\varvec{\theta }\).
Let \(\varvec{\theta _n}\) be on the boundary, \(\partial \mathbf {N}_n(\varvec{\theta _0}) \) of \(\mathbf {N}_n(\varvec{\theta _0})\). We define \(v_n= \dfrac{ {\mathbf {D}}^{\frac{1}{2}}_n(\varvec{\theta _0}) (\varvec{\theta _n} - \varvec{\theta _0}) }{A}\). Then, for \(\varvec{\theta _n} \in \partial \mathbf {N}_n(\varvec{\theta _0})\), \(v_n\) is a unit vector. For any arbitrary \(\epsilon \in (0,1)\),
Again if the condition (26) holds, from Lemma 6.6, we get the eigenvalues of \({\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta _0}) {\mathbf {F}}_n( \varvec{{\bar{\theta }}}) {\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta _0})\) converging to 1 in probability. Therefore, \(P(\frac{1}{2} ( \varvec{\theta _n} - \varvec{\theta _0} )^{\top } {\mathbf {F}}_n( \varvec{{\bar{\theta }}}) (\varvec{\theta _n} - \varvec{\theta _0} \le \frac{\epsilon A^2}{2}) = o(1)\).
As \( ( \varvec{\theta _n} - \varvec{\theta _0})^{\top } {\mathbf {S}}_n(\varvec{\theta _0})\) has mean 0, applying Chebyshev’s inequality we obtain,
Thus,
Therefore, for A and n large enough, \( {\mathbf{l}}_n(\varvec{\theta )} < {\mathbf{l}}_n( \varvec{\theta _0})\) for any \( \varvec{\theta } \) lying on \(\partial {\mathbf {N}_n( \varvec{\theta _0)} }\). As \({\mathbf {D}}_n(\varvec{\theta _0})\) is positive definite, by (31), \({\mathbf {F}}_n( \varvec{\theta } )\) is positive definite. Therefore, \({\mathbf{l}}_n( \varvec{\theta } )\) is a concave function in \(\mathbf {N}_n(\varvec{\theta _0} )\) which is an ellipsoid. \({\mathbf{l}}_n( \varvec{\theta } )\) does not have a maximum on the boundary of \(\mathbf {N}_n( \varvec{\theta _0} )\) and hence it has a local maximum inside \(\mathbf {N}_n( \varvec{\theta _0} )\) which is a global maximum because of its concavity. Therefore, an MLE \(\varvec{\widehat{\theta }_n}\) of \(\varvec{\theta _0} \) exists inside the ellipsoid \(\mathbf {N}_n(\varvec{\theta _0})\) with probability 1 as \(n \rightarrow \infty \) and it is unique. \(\square \)
Proof of Theorem 4.1
Based on Lemmas 6.3 and 6.7, the proof immediately follows. \(\square \)
Lemma 6.8
For \(\lambda _{10}, \lambda _{20} >0\) and \(p_0 \in (0,1),\) if the condition (26) holds, \(\lambda _{\mathrm{min}} \{{\mathbf {D}}_n\} \rightarrow \infty \) as \(n \rightarrow \infty \), and \(\varvec{\widehat{\theta }_n}\) is a consistent estimator of \(\varvec{{\theta }_o}.\)
Proof
Based on Lemma 6.7, and as the MLE \(\varvec{\widehat{\theta }_n} \in \mathbf {N}_n(\varvec{\theta _0})\), \(( \varvec{\widehat{\theta }_n} - \varvec{\theta _0} )^{\top } {\mathbf {D}}_n(\varvec{\theta }_o) \big ( \varvec{\widehat{\theta }_n} - \varvec{\theta _0} \big ) \le A^2\). Again \(( \varvec{\widehat{\theta }_n} - \varvec{\theta _0} )^{2} \lambda _{\mathrm{min}}\{ {\mathbf {D}}_n(\varvec{\theta }_o)\} \le ( \varvec{\widehat{\theta }_n} - \varvec{\theta _0})^{'} {\mathbf {D}}_n (\varvec{\theta }_o)\big ( \varvec{\widehat{\theta }_n} - \varvec{\theta _0} \big ).\) If in addition, \(\lambda _{\mathrm{min}} \{{\mathbf {D}}_n\} \rightarrow \infty \) as \(n \rightarrow \infty \), \(\varvec{\widehat{\theta }_n} \rightarrow \varvec{\theta _0}\) in probability. \(\square \)
Proof of Theorem 4.2
Based on Lemmas 6.3 and 6.8, the proof immediately follows. \(\square \)
Lemma 6.9
For \(\lambda _{10}, \lambda _{20} >0\) and \(p_0 \in (0,1)\) there exists a finite positive quantity \(G(\varvec{\theta _0})\) such that \(E{|s_i(\varvec{\theta _0})|}^4 \le G(\varvec{\theta _0}).\)
Proof of Theorem 4.2
Along the same lines as the previous two lemmas, Lemmas 6.4 and 6.5, we can show that \(s_i(\varvec{\theta _0})\) is also bounded, and hence, the moments are bounded. \(\square \)
Lemma 6.10
For \(\lambda _{10}, \lambda _{20} >0\) and \(p_0 \in (0,1),\) if the condition (26) holds, \( {\mathbf {D}}^{1/2}_n(\varvec{\theta }_o)\big (\widehat{\varvec{\theta _n}}-\varvec{\theta }_o\big ) \) is asymptotically normally distributed with mean vector \( {\varvec{0}}\) and covariance matrix \(\mathbf {I}.\)
Proof
Considering the element wise Taylor series expansion of \( {\mathbf {S}}_n(\varvec{\theta _0} )\) about \(\varvec{\widehat{\theta }_n}\), we obtain
where \(\varvec{{\widetilde{\theta }}_n} = ( \varvec{\theta ^{*}_n} , \varvec{\theta ^{**}_n}, \varvec{\theta ^{***}_n})\), \(\varvec{\theta ^{*}_n}\), \(\varvec{\theta ^{**}_n}\) and \(\varvec{\theta ^{***}_n}\) are three points on the line segment joining \(\varvec{\theta _0}\) and \(\varvec{\widehat{\theta }_n}\), and
As the likelihood function is maximized at MLE \(\varvec{\widehat{\theta }_n}\), \({\mathbf {S}}_n(\varvec{\widehat{\theta }}_n)={\varvec{0}},\) and we can write
where \( A_n(\varvec{{\widetilde{\theta }}_n}) = {\mathbf {D}}^{-\frac{1}{2}}_n (\varvec{\theta }_o) \mathbf {K}_n(\varvec{{\widetilde{\theta }}_n}){\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta }_o)\). To prove the asymptotic normality of \( {\mathbf {D}}^{\frac{1}{2}}_n(\varvec{\theta }_o) ( \varvec{\widehat{\theta }_n} - \varvec{\theta _0} )\), it is enough to show that for any \(\varvec{\theta ^{*}_n}\), \(\varvec{\theta ^{**}_n}\) and \(\varvec{\theta ^{***}_n} \) \(\in \mathbf {N}_n(\varvec{\theta _0} )\), \(A_n(\varvec{{\widetilde{\theta }}_n}) \overset{P}{\rightarrow } \mathbf {I}\) and \({\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta }_o) {\mathbf {S}}_n(\varvec{\theta _0})\) converges in distribution to a standard normal random variable. Now
Applying (32), \( {\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta }_o)\big \{ {\mathbf {F}}_n(\varvec{\theta _0}) - {\mathbf {D}}_n(\varvec{\theta }_o)\big \} {\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta }_o) \) converges in probability to the zero matrix. Also for any unit vector \({\mathbf {u}}\) and \(\varvec{\theta ^{*}_n}, \varvec{\theta ^{**}_n}, \varvec{\theta ^{***}_n} \in \mathbf {N}_n(\varvec{\theta _0} )\), we can prove that
\({\mathbf {u}}^{'} {\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta }_o) \big \{ \mathbf {K}_n(\varvec{{\widetilde{\theta }}_n}) - {\mathbf {F}}_n(\varvec{\theta _0} \big \} ) {\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta }_o) {\mathbf {u}} = o_{p}(1)\). The proof follows exactly along the same lines as the proof of (33). Hence, we have \(A_n(\varvec{{\widetilde{\theta }}_n}) \overset{P}{\rightarrow } \mathbf {I}\). \({\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta }_o) {\mathbf {S}}_n(\varvec{\theta _0})\) can be written as
Let \(X_{in}(\varvec{\theta _0})= {\mathbf {u}}^{\top } {\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta }_o){\mathbf {s}}_{i}(\varvec{\theta _0}),\,i=1,2,\ldots ,n,\) where \({\mathbf {u}}\) is a unit vector. Then, \(E(X_{in}(\varvec{\theta _0}))=0\) and \(\sum _{i=1}^{n} {\mathrm{Var}}(X_{in}(\varvec{\theta _0}) ) =1.\) To prove that \( {\mathbf {D}}^{-\frac{1}{2}}_n(\varvec{\theta }_o) {\mathbf {S}}_n(\varvec{\theta _0})\) converges to the standard normal distribution, it is enough to show that the sequence \(\{ X_{in}(\varvec{\theta _0}) \}\) follows Lindberg condition, i.e., for every \(\xi >0\),
We need Lemma 6.9 to prove the Lindberg condition. Based on Lemma 6.9,
Let \(I({\mathscr {A}})\) denote the indicator variable on a set \({\mathscr {A}}.\) By (35), Cauchy Schwartz’s inequality and Chebyshev’s inequality, we obtain
\(\square \)
Proof of Theorem 4.3
Based on Lemmas 6.3 and 6.10, the proof immediately follows. \(\square \)
Rights and permissions
About this article
Cite this article
Pal, A., Mondal, S. & Kundu, D. A Cure Rate Model for Exponentially Distributed Lifetimes with Competing Risks. J Stat Theory Pract 15, 21 (2021). https://doi.org/10.1007/s42519-020-00145-0
Accepted:
Published:
Keywords
- Cure rate model
- Long-term survivors
- Competing risk
- EM algorithm
- Asymptotic normality
- Consistency
- Maximum likelihood estimator
Mathematics Subject Classification
- Primary 62F12
- secondary 62N05