Abstract
In this paper, the best linear unbiased estimator of regression coefficients in the singular linear model was considered. Under the weighted balanced loss function the minimum risk properties of linear estimators of regression coefficients in the class of linear unbiased estimators are derived. Some kinds of relative efficiencies of the best linear unbiased estimator are given, and the lower bounds or upper bounds of these relative efficiencies are also presented.
Similar content being viewed by others
1 Introduction
Let us consider the following linear model
where y presents an \(n\times 1\) vector of observation, X presents an \(n\times p\) known matrix of rank p, \(\beta \) shows a \(p\times 1\) vector of unknown parameters, \(\epsilon \) shows an \(n\times 1\) vector of disturbances with expectation \(E(\epsilon )=0\) and variance-covariance matrix \(Cov(\epsilon )=\sigma ^{2}I_{n}\).
Suppose that \(\tilde{\beta }\) presents any estimator of \(\beta \), then the quadratic loss function which presents the goodness of fit of the model is defined as follows
while the authors usually use the following squared loss function or the weighted squared error loss function to define the precision of estimation
or weighted squared error loss function
As we all know, both the three criteria are important and it may be desirable to employ both the criteria simultaneously in practice. Zellner (1994) considered both the criteria of goodness of fit and precision of estimation together, and proposed the following balanced loss function:
with w shows a scalar between 0 and 1. When \(w=0\), the loss function (1.5) presents the precision of estimation and when \(w=1\), the loss function (1.5) shows the goodness of fitted model.
Furthermore, Shalabh (1995) use the conception of simultaneous prediction of actual and average values of study variable, introduced the following loss function
with w presents a scalar between 0 and 1. It is easy to know that (1.6) is an extension of (1.5) and it is also take care of the covariability between the goodness of fit and precision of estimation.
The balanced loss function has been received considerable attention in the literature. Rodrigues and Zellner (1994) have applied the balanced loss function in the estimation of mean time to failure. Gruber (2004) have considered the empirical Bayes and approximate minimum mean square error estimator under a general balanced loss function. Zhu et al. (2010) got the best linear unbiased estimator under the balanced loss function. Hu et al. (2010) got the optimal estimator under the balanced loss function and also considered the relative efficiency of the optimal estimator with the ordinary least squares estimator.
As the weighted balanced loss function is very popular and important, and there is no author to study the best linear unbiased estimator under the weighted balanced loss function in the singular linear model. We also want to know what form of the optimal estimator under the weighted balanced loss function. Is the optimal estimator in singular linear model under weighted balanced loss function is same as the usual best linear unbiased estimator which is derived under the squared loss function or balanced loss function? In this we will do this work. We will derive the optimal estimator under the weighted balanced loss function. Some relative efficiencies are also presented in this paper.
The rest of this paper is organized as follows. In Sect. 2, we present the singular linear model and the weighted balanced loss function. In Sect. 3 we consider the best linear unbiased estimator and discuss some relative efficiencies. A numerical example is presented to show the lower and upper bounds of these relative efficiencies in Sect. 4 and some conclusion remarks are presented in Sect. 5.
2 Linear model and loss function
In this paper we discuss the following singular linear model
where y presented an \(n\times 1\) observable random vector, \(\epsilon \) shows an \(n\times 1\) random error vector, X shows an \(n\times p\) known matrix with \(rank(X)=p\), \(\sigma ^2\) is known constant, V presents an \(n\times n\) known nonnegative definite matrix where \(\beta \) are \(p\times 1\) unknown parameters. Plenty of statisticians have presented many ways to get the best linear estimator of regression coefficient of this model. Rao (1985) give a simple way to obtain the best linear unbiased estimator . Fan and Wu (2014) discussed the best linear unbiased estimator under the weighted balanced loss function when V is a positive definite matrix.
With the unified theory least squares and Shalabh (1995) thought of weighted balanced loss, we present the following weighted balanced loss function:
for w is a scalar between 0 and 1, \(T=V+XUX'\), U is a positive definite matrix, S is a positive definite matrix and \(\tilde{\beta }\) presents an any estimator of \(\beta \). Then the corresponding risk function is denoted by
3 Main results
In this section we will list the main results of this paper. Firstly, we discuss the best linear unbiased estimator under the weighted balanced loss function in the singular linear model.
3.1 The best linear unbiased estimator
Write \(\mathfrak {I}_1=\{Ly; \ L \in R^{p\times n}, \ L \ is \ a \ constant\ matrix\ and \ LX=I_p\}\), the optimal estimator in \(\mathfrak {I}_1\) is got when V is an nonnegative definite matrix.
Firstly, we list five Lemmas which are needed in the following discussions.
Definition 3.1
If Ly in \(\mathfrak {I}_1\) let
reach the minimum, then Ly is the best linear unbiased estimator.
Lemma 3.1
Let X be an \(n\times p\) matrix, L is a \(p\times n\) matrix, then \(\frac{\partial trL'X'XL}{\partial L}=2X'XL\).
Proof
See Wang (1987). \(\square \)
Lemma 3.2
Let A be an \(n\times n\) positive definite matrix, and \(\lambda _1\ge \cdots \ge \lambda _n>0\) be the ordered eigenvalues of A, P is an \(n\times k\) matrix with \(P'P=I_k\), \(n>k\), denote \(l=min(k,n-k)\), then
Proof
See Yang (1988). \(\square \)
Lemma 3.3
Let A be an \(n\times n\) positive definite matrix, and \(\lambda _1\ge \cdots \ge \lambda _n>0\) be the ordered eigenvalues of A, P is an \(n\times k\) matrix with \(P'P=I_k\), \(n>k\), denote \(l=min(k,n-k)\), then
Proof
See Rao (1985). \(\square \)
Lemma 3.4
Let \(T=V+XUX'\) and \(rank(T)=rank(V\vdots X)\), then
-
(1)
\(\mathfrak {R}(V\vdots X)=\mathfrak {R}(T)\)
-
(2)
\((y-X\beta )'T^-(y-X\beta )\), \(X'T^-X\) and \(X'T^-y\) are independent of the choice of \(T^-\).
-
(3)
\(X'T^-T=X'\), \(X(X'T^-X)^-XT^-X=X\).
Proof
See Wang (1987) and Wang et al. (2006). \(\square \)
Lemma 3.5
Let \(rank(X)=p\), \(T=V+XUX'\) and \(rank(T)=rank(V\vdots X)\), then \(X'T^-X\) is a reversible matrix.
Proof
See Hu et al. (2010). \(\square \)
In the following theorem we will derive the optimal estimator of \(\beta \) in \(\mathfrak {I}_1\) under the weighted balanced loss function.
Theorem 3.1
For the model (2.1), when V is an nonnegative definite matrix, if \(L=(X'T^{-}X)^{-1}X'T^{-}\), then Ly is the best linear unbiased estimator of \(\beta \) under the weighted balanced loss function.
Proof
Suppose that \(Ly\in \mathfrak {I}_1\) and according to the weighted balanced loss function (2.2), then its risk function is presented as follows (Lemmas 3.4, 3.5)
where \(M=w(2-w)X'T^-X+(1-w)^2S\). If we want Ly to be the best linear unbiased estimator of \(\beta \) that is equal to
Using the Lagrange way. Let
for \(\lambda \) shows a \(p\times p\) matrix of Lagrangian Multipliers. Differentiate with respect to L and \(\lambda \) (Lemma 3.1):
Then we solve this problem and get
with G is an any \(p\times p\) matrix.
Now we prove that the risk of \((X'T^{-}X)^{-1}X'T^{-}y\) is equal to the risk of \(Ly=[(X'T^{-}X)^{-1}X'T^{-}+G(I_n-TT^-)]y\).
Since \((I_n-TT^-)V=(I_n-TT^-)(T-XUX')=O\) and
So we have \(R\left( [(X'T^{-}X)^{-1}X'T^{-}+G(I_n-TT^-)]y,\beta \right) =R((X'T^{-}X)^{-1}X'T^{-}y, \beta )\).
Next we prove that \((X'T^{-}X)^{-1}X'T^{-}y\) reaches the minimum risk in \(\mathfrak {I}_1\).
Suppose that \(\tilde{L}y\) be an any estimator of \(\beta \) in \(\mathfrak {I}_1\). By \(LX=I_p\), we obtain \(\tilde{L}=(X'T^{-}X)^{-1}X'T^{-}+\mu N\), where \(\mu \) is an any \(p\times n\) matrix and \(N=I_n-X(X'T^{-}X)^{-1}X'T^{-}\), then we obtain
and the equality holds if and only if \(L=(X'T^{-}X)^{-1}X'T^{-}\). \(\square \)
Corollary 3.1
For the model (2.1), if \(\mathfrak {R}(X)\subset \mathfrak {R}(V)\) and \(L=(X'T^{-}X)^{-1}X'V^{-}\), then Ly is the best linear unbiased estimator for \(\beta \) under the weighted loss function.
Remark 3.1
Under the weighted balanced loss function, the best linear unbiased estimator of \(\beta \) in \(\mathfrak {I}_1\) is \(\hat{\beta }=(X'T^{-}X)^{-1}X'T^{-}y\), which is same as the estimator \(\hat{\beta }=(X'T^{-}X)^{-1}X'T^{-}y\) got by Zhu et al. (2010) and Hu et al. (2010) under the balanced loss function.
Remark 3.2
Under the weighted balanced loss function, if V is a positive definite matrix then the best linear unbiased estimator of \(\beta \) in \(\mathfrak {I}_1\) is \(\hat{\beta }=(X'V^{-1}X)^{-1}X'V^{-1}y\), which is same as the estimator obtained by Fan and Wu (2014) under the weighted balanced loss function.
3.2 Relative efficiencies of the best linear unbiased estimator
In Sect. (3.1), we derive the best linear unbiased estimator \(\hat{\beta }_T=(X'T^{-}X)^{-1}X'T^{-}y\) under the weighted balanced loss function in the singular linear regression model. However, as we all know, the covariance matrix is usually unknown, at this time, we use the ordinary least squares estimator \(\hat{\beta }_{OLS}=(X'X)^{-1}X'y\) to replace the best linear unbiased estimator. We know this will lead to some loss. We know this will lead to some loss. There are many papers have discussed this loss, such as: Rao (1985), Yang and Wang (2009), Yang and Wu (2011), Liu (2000),Liu et al. (2009), Wang and Yang (2012) and Wu (2014). In this subsection, we definite two relative efficiencies to measure the loss under the weighted balanced loss risk function.
Now we define two relative efficiencies:
where \(R(\tilde{\beta },\beta )\) is defined in (2.3), \(\hat{\beta }_T=(X'T^{-}X)^{-1}X'T^{-}y\) and \(\hat{\beta }_{OLS}=(X'X)^{-1}X'y\).
Based on the definition of weighted balanced loss risk function, we obtain
and
Now we present the bounds of the two relative efficiencies.
Theorem 3.2
For the model (2.1), when V is an nonnegative definite matrix, let \(T=Q'\Lambda Q\), Q is an \(k\times p\) orthogonal matrix, \(\Lambda =diag(\lambda _1,\ldots ,\lambda _k)\), where \(\lambda _1\ge \cdots \ge \lambda _k>0\). \(c_1\ge \cdots \ge c_p>0=c_{p+1}= \cdots =c_k\) are the ordered eigenvalues of \(QXM^{-1}X'Q'\), then we have
where \(rank(X)=p\), \(rank(T)=k\) and \(k>p\), \(m=min(p,k-p)\).
Proof
Since \(\mathfrak {R}(X)\subset \mathfrak {R}(T)\), then QX is \(k\times p\) matrix with \(rank(QX)=p\). Denote \(QX=\gamma \), then we have \(rank(\gamma )=p\) and
where \(\rho =\gamma M^{-1/2}\), \(rank(\rho )=k\). Then do singular value decomposition for \(\rho \), we obtain \(\rho =Q_1\Gamma ^{1/2} Q_2'\), where \(Q_1\) is an \(n\times p\) orthogonal matrix and \(Q_2\) is a \(p\times p\) orthogonal matrix. \(\Gamma =diag(c_1,\ldots ,c_p)\), then by Lemma 3.3 and equation (3.14), we obtain
The proof of Theorem 3.2 is completed. Now, we present another theorem of this paper. \(\square \)
Theorem 3.3
For the model (2.1), when V is an nonnegative definite matrix, let \(T=Q'\Lambda Q\), Q is an \(k\times p\) orthogonal matrix, \(\Lambda =diag(\lambda _1, \ldots ,\lambda _k)\), where \(\lambda _1\ge \cdots \ge \lambda _k>0\). \(c_1\ge \cdots \ge c_p>0=c_{p+1}= \cdots =c_k\) are the eigenvalues of \(QXM^{-1}X'Q'\). If \(w^2tr(T^-T)-2pw-(1-w)^2trSU=0\), then we have
where \(rank(X)=p\), \(rank(T)=k\), \(m=min(p,k-p)\) and \(k>p\).
Proof
By the proof of Theorem 3.2 and Lemma 3.2 we have
The proof of Theorem 3.4 is completed. \(\square \)
Theorem 3.4
For the model (2.1), when V is an nonnegative definite matrix, let \(T=Q'\Lambda Q\), Q is an \(k\times p\) orthogonal matrix, \(\Lambda =diag(\lambda _1,\ldots ,\lambda _k)\), where \(\lambda _1\ge \cdots \ge \lambda _k>0\). \(c_1\ge \cdots \ge c_p>0=c_{p+1}=\cdots =c_k\) are the eigenvalues of \(QXM^{-1}X'Q'\). If \(w^2tr(T^-T)-2pw-(1-w)^2trSU\ne 0\) and \(X'T^-X\le U^{-1}\), then
where \(rank(X)=p\), \(rank(T)=k\) and \(k>p\), \(m=min(p,k-p)\).
Proof
By Theorem 3.2, we have
\(\square \)
4 Numerical example
In this section we will use a numerical example to explain the theoretical results which presented in Sect. 4. The data is given as follows, and the data was studied by many authors such as Chang and Yang (2012).
In this paper we consider \(\sigma ^2=0.05\), \(U_1=I\) and \(U_2=2I\) and
It is easy to know that V is an nonnegative definite matrix. Then we can compute the estimated \(e_1\) and \(e_2\), which is showed in Figs. 1 and 2.
By the figure s we find that the \(e_1\) and \(e_2\) are all smaller than 1.
5 Conclusion
In this paper, we derive the best linear unbiased estimator when the covariance matrix is an nonnegative matrix, and by the results we find that the best linear unbiased estimator under weighted balanced loss function in singular linear model is same as the best linear unbiased estimator got by Rao (1985) under the squared loss function and Hu et al. (2010) under the balanced loss function. Furthermore, some relative efficiencies based on the weighted balanced loss function are presented and the lower and upper bounds of these relative efficiencies are presented.
References
Chang XF, Yang H (2012) Performance of the preliminary test two-parameter estimators based on the conflicting test-statistics in a regression model with student-t error. Statistics 43(3):291–303
Fan RY, Wu JB (2014) The best linear unbiased estimation of regression coefficient under weighted balanced loss function. Chin J Appl Probab Stat 30(3):303–312
Gruber MHJ (2004) The efficiency of shrinkage estimators with respect to Zellner’s balanced loss function. Commun Stat Theory Methods 33(2):235–249
Hu GK, Peng P, Luo H (2010) The relative efficiencies of best linear unbiased estimator of regression coefficient under balanced loss function. J Math 30:1059–1064 (In Chinese)
Liu SZ (2000) Efficiency comparisons between the OLSE and the BLUE in a singular linear model. J Stat Plan Inference 84(1–2):191–200
Liu SZ, Lu CY, Puntanen S (2009) Matrix trace Wielandt inequalities with statistical applications. J Stat Plan Inference 139(7):2254–2260
Rao CR (1985) The inefficiency of least squares: extension of the Kantorovich inequality. Linear Algebra Appl 70:249–255
Rodrigues J, Zellner A (1994) Weighted balanced loss function and estimation of the mean time to failure. Commun Stat Theory Methods 23(12):3609–3616
Shalabh (1995) Performance of Stein-rule procedure for simultaneous prediction of actual and average values of study variable in linear regression model. Bull Int Stat Inst 56:1375–1390
Wang SG (1987) Linear models and its applications. The Education of Anhui Press, Hefei (In Chinese)
Wang LT, Yang H (2012) Matrix Euclidean norm Wielandt inequalities and their applications to statistics. Stat Pap 53(3):521–530
Wang SG, Wu MX, Jia ZZ (2006) The inequalities of matrices. The Science Press, Beijing (In Chinese)
Wu JB (2014) Some properties of relative efficiency of estimators in a two linear regression equations system with identical parameter vectors. J Inequal Appl 2014:279
Yang H (1988) Extensions of the Kantorovich inequality and the error ratio efficiency of the mean square. Math Appl 4:85–90 (In Chinese)
Yang H, Wang LT (2009) An Alternative form of the Watson Efficiency. J Stat Plan Inference 139(8):2767–2774
Yang H, Wu JB (2011) Some Matrix norm Kantorovich inequalities and their applications. Commun Stat Theory Methods 40(22):4078–4085
Zellner A (1994) Bayesian and non-Bayesian estimation using balanced loss functions. In: Gupta SS, Berger JO (eds) Statistical decision theory and related topics V. Springer, Verlag, pp 377–390
Zhu H, Rao ZY, Hu GK (2010) Best linear unbiased estimation of regression coefficient under balanced loss function. College Math 26(1):96–99 (In Chinese)
Acknowledgments
The authors are highly obliged to the editor and the reviewer for the comments and suggestions which improved the paper in its present form. This work was supported by the National Natural Science Foundation of China (No. 11501072, 11401419), and the Project of Humanities and Social Sciences of MOE (Ministry of Education in China) (No. 16YJC910005).
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Wu, J., Liu, C. The best linear unbiased estimator in a singular linear regression model. Stat Papers 59, 1193–1204 (2018). https://doi.org/10.1007/s00362-016-0811-6
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00362-016-0811-6