This article compares a predictor with the best linear unbiased predictor (BLUP) for a unified form of all unknown parameters under a stochastically restricted linear mixed model (SRLMM) in terms of the mean squared error matrix (MSEM) criterion. The methodology of block matrix inertias and ranks is employed to compare the MSEMs of these predictors. The comparison results are also demonstrated for a linear mixed model with and without an exact restriction, as well as special cases of the unified form of all unknown parameters in the SRLMM.
Citation: Nesrin Güler, Melek Eriş Büyükkaya. Statistical inference of a stochastically restricted linear mixed model[J]. AIMS Mathematics, 2023, 8(10): 24401-24417. doi: 10.3934/math.20231244
[1] | Ruixia Yuan, Bo Jiang, Yongge Tian . A study of the equivalence of inference results in the contexts of true and misspecified multivariate general linear models. AIMS Mathematics, 2023, 8(9): 21001-21021. doi: 10.3934/math.20231069 |
[2] | Yongge Tian . An effective treatment of adding-up restrictions in the inference of a general linear model. AIMS Mathematics, 2023, 8(7): 15189-15200. doi: 10.3934/math.2023775 |
[3] | Jairo A. Angel, Francisco M.M. Rocha, Jorge I. Vélez, Julio M. Singer . A new test for detecting specification errors in Gaussian linear mixed-effects models. AIMS Mathematics, 2024, 9(11): 30710-30727. doi: 10.3934/math.20241483 |
[4] | Qi Shao, Yongkun Li . Almost periodic solutions for Clifford-valued stochastic shunting inhibitory cellular neural networks with mixed delays. AIMS Mathematics, 2024, 9(5): 13439-13461. doi: 10.3934/math.2024655 |
[5] | Yi Zhang, Yuanpeng Zhao, Na Li, Yingying Wang . Observer-based sliding mode controller design for singular bio-economic system with stochastic disturbance. AIMS Mathematics, 2024, 9(1): 1472-1493. doi: 10.3934/math.2024072 |
[6] | Mohamed Said Mohamed, Haroon M. Barakat, Aned Al Mutairi, Manahil SidAhmed Mustafa . Further properties of Tsallis extropy and some of its related measures. AIMS Mathematics, 2023, 8(12): 28219-28245. doi: 10.3934/math.20231445 |
[7] | Hiroshi Takahashi, Ken-ichi Yoshihara . Approximation of solutions of multi-dimensional linear stochastic differential equations defined by weakly dependent random variables. AIMS Mathematics, 2017, 2(3): 377-384. doi: 10.3934/Math.2017.3.377 |
[8] | Mahvish Samar, Xinzhong Zhu . Structured conditioning theory for the total least squares problem with linear equality constraint and their estimation. AIMS Mathematics, 2023, 8(5): 11350-11372. doi: 10.3934/math.2023575 |
[9] | Chantapish Zamart, Thongchai Botmart, Wajaree Weera, Prem Junsawang . Finite-time decentralized event-triggered feedback control for generalized neural networks with mixed interval time-varying delays and cyber-attacks. AIMS Mathematics, 2023, 8(9): 22274-22300. doi: 10.3934/math.20231136 |
[10] | Dayang Dai, Dabuxilatu Wang . A generalized Liu-type estimator for logistic partial linear regression model with multicollinearity. AIMS Mathematics, 2023, 8(5): 11851-11874. doi: 10.3934/math.2023600 |
This article compares a predictor with the best linear unbiased predictor (BLUP) for a unified form of all unknown parameters under a stochastically restricted linear mixed model (SRLMM) in terms of the mean squared error matrix (MSEM) criterion. The methodology of block matrix inertias and ranks is employed to compare the MSEMs of these predictors. The comparison results are also demonstrated for a linear mixed model with and without an exact restriction, as well as special cases of the unified form of all unknown parameters in the SRLMM.
We will begin by introducing some notation before proceeding. We will write Λ∈Rk,l if Λ is a k×l real matrix, Λ∈Rsk if Λ∈Rk,k and is symmetric, Λ∈R≽k if Λ∈Rsk and is positive semi-definite. We will use r(Λ), C(Λ), Λ′ and Λ+ as symbols to present the rank, the column space, the transpose and the Moore-Penrose generalized inverse of Λ∈Rk,l, respectively. Λ⊥=Ik−ΛΛ+ stands for the orthogonal projector, where Ik∈Rsk is the identity matrix of order k. The symbols i+(Λ) and i−(Λ) represent positive and negative inertias of Λ∈Rsk, respectively, while i±(Λ) and i∓(Λ) are used to denote both the positive and the negative inertias of Λ∈Rsk jointly. The inequality Λ1−Λ2≽0 or Λ1≽Λ2 means Λ1−Λ2∈R≽k in the L¨owner partial ordering for Λ1,Λ2∈Rsk. Similarly, when stating the other inequalities between Λ1,Λ2∈Rsk, Λ1−Λ2≻0 (Λ1≻Λ2), Λ1−Λ2≺0 (Λ1≺Λ2), Λ1−Λ2≼0 (Λ1≼Λ2) are used if the difference Λ1−Λ2 is positive definite, negative definite and negative semi-definite matrices, respectively. We will write E(λ) for the expectation vector and D(λ)=cov(λ,λ) for the dispersion matrix of a random vector λ∈Rk,1, where cov(λ,λ) denotes the covariance matrix of λ.
A linear mixed model (LMM) is a variant of linear regression model that combines fixed and random effects in the same analysis, allowing for more flexibility in model fitting. In some statistical problems, besides the sample information of LMMs, there may exist additional information, usually produced as prior information for the models. This prior information can be expressed as either a linear stochastic restriction or a linear exact restriction on unknown parameters, and it is usually added to the assumptions of models. Linear exact restrictions occur when there is a specific linear hypothesis being tested or when exact knowledge is available among certain parameters. On the other hand, linear stochastic restrictions arise when prior information is derived from previous investigations or established long-term relationships with relevant studies; see, e.g., [1,2].
An LMM with a linear stochastic restriction, also known as a stochastically restricted LMM (SRLMM), can be given as follows:
y=Xα+Zγ+ε with Sα+e=s, | (1.1) |
E[γε]=0,D[γε]=σ2Λ1,E(e)=0,D(e)=σ2Λ3,cov{[γε],e}=σ2Λ2, i.e., D[γεe]=σ2[Λ1Λ2Λ′2Λ3]:=σ2Λ, | (1.2) |
where y∈Rn,1 is a vector of responses, X∈Rn,k, Z∈Rn,p, S∈Rm,k are known matrices of arbitrary ranks and s∈Rm,1 is a known vector, α∈Rk,1 is a parameter vector of fixed effects, γ∈Rp,1 is a vector of random effects, ε∈Rn,1 and e∈Rm,1 are vectors of random errors, Λ∈R≽n+p+m of arbitrary ranks and its entries Λi are known, i=1,2,3 and σ2 is a positive unknown parameter.
In the present study, we consider the SRLMM in (1.1). We derive the best linear unbiased predictors (BLUPs) of a unified form of all unknown parameters in the SRLMM utilizing some quadratic matrix optimization methods, which include block matrix inertias and ranks. Furthermore, we discuss some of the BLUPs' basic characteristics. We specifically address the comparison problem between any predictor and the BLUP for all unknown parameters in a unified form under the SRLMM in the sense of the mean squared error matrix (MSEM) criterion. It is well-known that the MSEM of any predictor or estimator ˜λ of λ in a linear regression model is defined as the matrix
MSEM(˜λ)=E(˜λ−λ)(˜λ−λ)′. | (1.3) |
To compare two given predictors or estimators ˜λ1 and ˜λ2, the difference
Δ(˜λ1,˜λ2)=MSEM(˜λ1)−MSEM(˜λ2) | (1.4) |
is written. In this case, ˜λ2 is said to be superior to ˜λ1 with respect to the MSEM criterion iff Δ(˜λ1,˜λ2)≽0 holds. Using the methodology of block matrix inertias and ranks, a number of equalities and inequalities for the comparisons of MSEMs of two predictors, with one being the BLUP, of a unified form of all unknown parameters are established by considering the difference in (1.4). Such comparisons allow us to evaluate whether two predictors, one of which is BLUP, of the same unknown parameter vector are more effective than each other. We also derive comparison results for special cases of a unified form of all unknown parameters under the SRLMM. Furthermore, the results are reduced to both a constrained linear mixed model (CLMM), i.e., an LMM with a linear exact restriction, and an unconstrained LMM (ULMM). We use a numerical example to explain the theoretical results. There have been numerous previous and recent studies in the literature that focus on an exact restriction on unknown parameters under LMMs and linear models from different perspectives; see, e.g., [3,4,5,6,7,8,9,10,11,12,13,14] and the references therein. There has also been considerable attention on stochastic restriction on unknown parameters in a linear regression model from different approaches; see, e.g., [1,2,15,16,17,18], among others. We may also refer to [19,20,21,22,23,24,25,26,27,28], among others, for the corresponding studies in which similar approaches were conducted for the comparisons of predictors or estimators by utilizing the methodology of block matrix inertias and ranks. It is worth noting that studies on stochastic or exact restrictions are not limited to the context of linear models and in the field of statistics. The stochastic or exact restrictions on linear or nonlinear equations have been widely investigated from various aspects in a variety of fields, including biostatistics, educational measurement, sociology, finance, biology, chemistry, mechanics and economics, see, e.g., [29,30,31,32,33,34].
The development of various potential predictors (or, also potential estimators) of unknown parameter vectors in linear regression models is one of the main goals of theoretical and practical studies. In order to define predictors of unknown parameter vectors in the model, various types of optimality criteria have been utilized in mathematics. The unbiasedness of predictors of unknown parameter vectors according to the parameter spaces in the model is one of the crucial characteristics among others. When there are numerous unbiased predictors for the same parameter space, finding the one with the smallest dispersion matrix becomes significant. The BLUP, by its definition, corresponds to these two requirements. Thus, due to the minimizing property of the BLUP's dispersion matrix, the BLUP has a significant role to play in statistical inference theory and is frequently used as the foundation for evaluating the effectiveness of various predictors. As stated in the next section, the dispersion matrix of the BLUP is equal to its MSEM. In this situation, in the context of the SRLMM, comparing any predictor with the BLUP for a unified form of all unknown parameters using the MSEM criterion allows us to determine the optimality of the compared predictor. As it is known, one of the main criteria for determining the superiority of two different predictors of the unknown parameter vectors in the model is the MSEM criterion, which is widely used to measure predictors accurately.
The expressions, which are obtained from the comparison issues of the predictors, include various complicated matrix operations, including the use of Moore-Penrose generalized inverses. Some of the methods to simplify expressions composed by matrices and their Moore-Penrose generalized inverses are based on conventional operations on matrices, as well as some known equalities for inverses or Moore-Penrose generalized inverses of matrices. These methods' efficacy is quite limited, and the related processes are quite tedious. However, the methodology of block matrix inertias and ranks offers an effective approach to simplify such complex matrix expressions. It allows for the representation of results as simple inertia and rank equalities. The theory of matrix inertias and ranks has been developed as an efficient methodology for the simplification of such complex matrix expressions. In linear algebra, the inertia and rank of a matrix are fundamental quantities that are straightforward to understand and calculate. Using the theory of inertias and ranks of the matrix, various inequalities and equalities between the MSEMs of predictors correspond to the comparison of quantities derived from inertias and ranks of matrices.
The paper presents novel theoretical techniques and innovative approaches for comparing any predictor and the BLUP of the same general vector of unknown parameters under the SRLMM. The comparisons are based on the MSEM criterion, and the methodology of block matrix inertias and ranks is employed to establish exact formulas for these comparisons. It is important to note that we do not make any distributional assumptions for the random vectors in the models, except for assuming the existence of first and second moments. Additionally, no restrictions are imposed on the ranks of given matrices. Furthermore, there are no specific requirements concerning the patterns of the submatrices of Λ in (1.2). In statistical practice, if we meet the cases where Λ is unknown, then the estimators of submatrices of Λ can be used. Alternatively, if Λ is given with some parametric forms and some specific patterns, then the estimators obtained from the observed data can be substituted. These are other types of inference work, which we do not consider in this article.
Lastly, we list a few lemmas that are essential for the subsequent sections of the paper, contributing to the overall theoretical framework developed for comparing predictors under the SRLMM.
Lemma 1.1 ([35]). Let Λ1, Λ2 ∈Rk,l, or, let Λ1, Λ2 ∈Rsk. Then,
r(Λ1−Λ2)=0⇔Λ1=Λ2. |
i−(Λ1−Λ2)=k⇔Λ1≺Λ2andi+(Λ1−Λ2)=k⇔Λ1≻Λ2. |
i−(Λ1−Λ2)=0⇔Λ1≽Λ2andi+(Λ1−Λ2)=0⇔Λ1≼Λ2. |
Lemma 1.2 ([35]). Let Λ1∈Rsk, Λ2∈Rk,l, Λ3∈Rsl and a∈R. Then,
r(Λ1)=i+(Λ1)+i−(Λ1). | (1.5) |
i±(aΛ1)=i±(Λ1)ifa>0andi±(aΛ1)=i∓(Λ1)ifa<0. | (1.6) |
i±[Λ1Λ2Λ′2Λ3]=i±[Λ1−Λ2−Λ′2Λ3]=i∓[−Λ1Λ2Λ′2−Λ3]. | (1.7) |
i±[Λ100Λ3]=i±(Λ1)+i±(Λ3)andi+[0Λ2Λ′20]=i−[0Λ2Λ′20]=r(Λ2). | (1.8) |
i±[Λ1Λ2Λ′20]=r(Λ2)+i±(Λ⊥2Λ1Λ⊥2). | (1.9) |
i+[Λ1Λ2Λ′20]=r[Λ1,Λ2]andi−[Λ1Λ2Λ′20]=r(Λ2)ifΛ1≽0. | (1.10) |
i±[Λ1Λ2Λ′2Λ3]=i±(Λ1)+i±(Λ3−Λ′2Λ+1Λ2)ifC(Λ2)⊆C(Λ1). | (1.11) |
Lemma 1.3 ([36]). Let Λ1∈Rk,l and Λ2∈Rm,l be given matrices, and let H∈R≽k. Suppose that there exists X∈Rm,k such that XΛ1=Λ2. Then the maximal positive inertia of XHX′−YHY′ s.t. all solutions of YΛ1=Λ2 is
maxYΛ1=Λ2i+(XHX′−YHY′)=r[XHΛ′1]−r(Λ1)=r(XHΛ⊥1). |
Hence, a solution X of XΛ1=Λ2 exists such that XHX′≼YHY′ holds for all solutions of YΛ1=Λ2 ⇔ both the equations XΛ1=Λ2 and XHΛ⊥1=0 are satisfied by X.
Let us consider the SRLMM in (1.1). By unifying two given equation parts in (1.1), we obtain
S:ys=Xsα+Zsγ+εs=Xsα+[Zs,In+m][γεs],ys=[ys],Xs=[XS],Zs=[Z0],εs=[εe]. | (2.1) |
Through the unifying operation that we use in (2.1), which is a common procedure for approaching two or more equations, the SRLMM in (1.1) is converted to the implicitly linear stochastically restricted LMM in (2.1). We can take into consideration the following general vector to produce conclusions on predictors of all unknown vectors under S:
λ=Kα+Lγ+Mεs=Kα+[L,M][γεs]=Kα+[L,M1,M2][γεe], | (2.2) |
where K∈Rt,k, L∈Rt,p, M1∈Rt,n and M2∈Rt,m with M=[M1,M2] are given matrices of arbitrary ranks. Then, from (1.2), (2.1) and (2.2),
E(ys)=Xsα,D(ys)=σ2[Zs,In+m]Λ[Zs,In+m]′:=σ2ΦΛΦ′,E(λ)=Kα,D(λ)=σ2[L,M]Λ[L,M]′:=σ2ΥΛΥ′,cov(λ,ys)=σ2[L,M]Λ[Zs,In+m]′:=σ2ΥΛΦ′, | (2.3) |
where Φ=[Zs,In+m] and Υ=[L,M]. Further, in this study, the model S is assumed to be consistent, i.e., ys∈C[Xs,ΦΛΦ′] holds with probability 1; see, e.g., [37].
We introduce the following definition of the predictability of λ and its subcases; see, e.g., [38,39]. Later, the definition of BLUP of λ under S is given; see, e.g., [40,41].
Definition 2.1. Let consider S in (2.1) and λ in (2.2). Then,
1) If a matrix F∈Rt,(n+m) exists with E(Fys−λ)=0, then λ is defined to be predictable, that is, λ is predictable ⇔ C(K′)⊆C(X′s) ⇔ Kα is estimable.
2) Xsα+Zsγ+εs is always predictable and Xsα is always estimable.
3) α is estimable ⇔ r(Xs)=k.
4) γ, ε and e are always predictable.
Definition 2.2. Suppose that λ in (2.2) is predictable under S in (2.1). Fys is said to be the BLUP of λ, represented by Fys=˜λBLUP, if there exists Fys such that
D(Fys−λ)= min s.t. E(Fys−λ)=0 |
holds in the L¨owner partial ordering. If L=0 and M=0 in λ, Fys is the well-known the best linear unbiased estimator (BLUE) of Kα, represented by ^KαBLUE.
According to the definition of the MSEM given in (1.3), we can give the following lemma, see; [42,43].
Lemma 2.1.. Let ˜λ be a predictor for a general vector of all unknown vectors λ in an LMM. Then,
MSEM(˜λ)=D(˜λ)+D(λ)−cov(˜λ,λ)−cov(λ,˜λ)+[bias(˜λ)][bias(˜λ)]′, | (2.4) |
where bias(˜λ)=E(˜λ−λ). If ˆλ is an estimator for a general vector of all unknown parameters λ in an LMM, then,
MSEM(ˆλ)=D(ˆλ)+[bias(ˆλ)][bias(ˆλ)]′, |
where bias(ˆλ)=E(ˆλ)−λ.
The fundamental BLUP equation and associated properties are summarized in the following lemma; see [25], for different approaches; see also, [37,41,44].
Lemma 2.2. Suppose that λ in (2.2) is predictable under S in (2.1). Consider any two unbiased linear predictors Fys and Gys for λ. Then,
maxE(Gys−λ)=0i+(D(Fys−λ)−D(Gys−λ))=r([F,−It][ΦΛΦ′ΦΛΥ′ΥΛΦ′ΥΛΥ′][XsK]⊥) | (2.5) |
is the maximal positive inertia of D(Fys−λ)−D(Gys−λ) s.t. GXs=K. Hence, D(Fys−λ)=min s.t. E(Fys−λ)=0⇔Fys=˜λBLUP
⇔F[Xs,ΦΛΦ′X⊥s]=[K,ΥΛΦ′X⊥s]. | (2.6) |
This equation in (2.6) is consistent and the BLUP of λ under S can be written as follows by considering the general solution of this equation:
˜λBLUP=Fys=([K,ΥΛΦ′X⊥s]J++UJ⊥)ys, | (2.7) |
where J=[Xs,ΦΛΦ′X⊥s] and U∈Rt,(n+m) is an arbitrary matrix. Further,
D(˜λBLUP)=σ2[K,ΥΛΦ′X⊥s]J+ΦΛΦ′([K,ΥΛΦ′X⊥s]J+)′, | (2.8) |
cov(˜λBLUP,λ)=σ2[K,ΥΛΦ′X⊥s]J+ΦΛΥ′, | (2.9) |
D(λ−˜λBLUP)=σ2([K,ΥΛΦ′X⊥s]J+Φ−Υ)Λ([K,ΥΛΦ′X⊥s]J+Φ−Υ)′, | (2.10) |
and the MSEM of ˜λBLUP is
MSEM(˜λBLUP)=D(λ−˜λBLUP). | (2.11) |
In particular,
1) C(J)=R(n+m),1 ⇔ F is unique and S is consistent ⇔ ˜λBLUP is unique.
2) The following equalities hold:
r(J)=r[Xs,ΦΛΦ′] and C(J)=C[Xs,ΦΛΦ′]. | (2.12) |
Proof of Lemma 2.2. For an unbiased linear predictor Fys of λ in S,
E(Fys−λ)=0⇔FXs=K, i.e., [F,−It][XsK]=0,D(Fys−λ)=σ2[F,−It][ΦΥ]Λ[ΦΥ]′[F,−It]′:=f(F) | (2.13) |
are written. For the other unbiased predictor Gys of λ, the similar expressions can be written as in (2.13) by putting G instead of F. Finding solution F of the consistent equation FXs=K such that
f(F)≼f(G) s.t. GXs=K | (2.14) |
corresponds to find the BLUP of λ under S. According to Lemma 1.3, (2.14) is a typical constrained quadratic matrix-valued function optimization problem in the L¨owner partial ordering. Applying Lemma 1.3 to (2.14), (2.5) is obtained. Using Lemma 1.3, we obtain the fundamental BLUP equation of λ in (2.6). The consistency of the equation in (2.6) is seen from the following column space inclusions:
C(K′)⊆C(X′s) and C(ΦΛΥ′)⊆C(ΦΛ)=C(ΦΛΦ′). | (2.15) |
The well-known general solution of (2.6) s.t. F can be written in the parametric form as in (2.7). (2.8) and (2.9) are directly seen from (2.3) and (2.7). D(λ−˜λBLUP) is written as
D(λ−˜λBLUP)=D(˜λBLUP)+D(λ)−cov(˜λBLUP,λ)−cov(λ,˜λBLUP). | (2.16) |
By setting the equalities in (2.3), (2.8) and (2.9) into (2.16), (2.10) is obtained. (2.4) and (2.16) give (2.11) since bias(˜λBLUP)=0. Item 1 follows from (2.7). The expressions in (2.12) are well-known results; see [45,Lemma 2.1 (a)].
We note that the fundamental equations of BLUPs and BLUEs for unknown vectors in λ and their related results can be derived from 5, by special choices of the matrices K, L and M.
The results of the relationships between the MSEM of any predictor for λ and the MSEM of the BLUP of λ under S are collected in the theorem given below.
Theorem 3.1. Suppose that λ in (2.2) is predictable under S in (2.1). Let ˜λ be any predictor (unbiased or biased) for λ under S. Denote
Δ(˜λ,˜λBLUP)=MSEM(˜λ)−MSEM(˜λBLUP) |
and
E=[ΦΛΦ′ΦΛΥ′XsΥΛΦ′ΥΛΥ′−σ−2MSEM(˜λ)KX′sK′0]=[σ−2D(ys)σ−2cov(ys,λ)Xsσ−2cov(λ,ys)σ−2(D(λ)−MSEM(˜λ))KX′sK′0]. |
Then,
i+(MSEM(˜λ)−MSEM(˜λBLUP))=i−(E)−r(Xs), | (3.1) |
i−(MSEM(˜λ)−MSEM(˜λBLUP))=i+(E)−r[Xs,ΦΛΦ′], | (3.2) |
r(MSEM(˜λ)−MSEM(˜λBLUP))=r(E)−r(Xs)−r[Xs,ΦΛΦ′]. | (3.3) |
In consequence,
1) Δ(˜λ,˜λBLUP)≽0, i.e., MSEM(˜λ)≽MSEM(˜λBLUP)⇔i+(E)=r[Xs,ΦΛΦ′].
2) Δ(˜λ,˜λBLUP)≻0, i.e., MSEM(˜λ)≻MSEM(˜λBLUP) ⇔i−(E)=r(Xs)+t.
3) Δ(˜λ,˜λBLUP)≼0, i.e., MSEM(˜λ)≼MSEM(˜λBLUP) ⇔i−(E)=r(Xs).
4) Δ(˜λ,˜λBLUP)≺0, i.e., MSEM(˜λ)≺MSEM(˜λBLUP) ⇔i+(E)=r[Xs,ΦΛΦ′]+t.
5) Δ(˜λ,˜λBLUP)=0, i.e., MSEM(˜λ)=MSEM(˜λBLUP)⇔r(E)=r[Xs,ΦΛΦ′]+r(Xs).
Proof. Let say C=MSEM(˜λ) and apply (1.11) to the difference between MSEM(˜λ) and MSEM(˜λBLUP) in (2.11). Then, we obtain
i±(Δ(˜λ,˜λBLUP))=i±(MSEM(˜λ)−MSEM(˜λBLUP))=i±(C−D(λ−˜λBLUP))=i±(C−σ2(DJ+Φ−Υ)Λ(DJ+Φ−Υ)′)=i±[ΛΛΦ′(J+)′D′−ΛΥ′DJ+ΦΛ−ΥΛσ−2C]−i±(Λ)=i±([Λ−ΛΥ′−ΥΛσ−2C]+[ΛΦ′00D][0JJ′0]+[ΦΛ00D′])−i±(Λ), | (3.4) |
where D=[K,ΥΛΦ′X⊥s] and J=[Xs,ΦΛΦ′X⊥s]. By using column space equalities and inclusions in (2.12) and (2.15), reapplying (1.11) to (3.4) and also using (1.6) and (1.8) with the elementary block matrix operations, the expression in (3.4) is equivalently written as
i±[0−JΦΛ0−J′00D′ΛΦ′0Λ−ΛΥ′0D−ΥΛσ−2C]−i∓[0JJ′0]−i±(Λ)=i±[−ΦΛΦ′−JΦΛΥ′−J′0D′ΥΛΦ′Dσ−2C−ΥΛΥ′]−r(J). | (3.5) |
By setting D and J into (3.5) and also using (1.7)–(1.9) and (2.12) with the elementary block matrix operations,
i±[−ΦΛΦ′−Xs−ΦΛΦ′X⊥sΦΛΥ′−X′s00K′−X⊥sΦΛΦ′00X⊥sΦΛΥ′ΥΛΦ′KΥΛΦ′X⊥sσ−2C−ΥΛΥ′]−r[Xs,ΦΛΦ′]=i±[−ΦΛΦ′−XsΦΛΥ′−X′s0K′ΥΛΦ′Kσ−2C−ΥΛΥ′]+i±(X⊥sΦΛΦ′X⊥s)−r[Xs,ΦΛΦ′]=i∓[ΦΛΦ′ΦΛΥ′XsΥΛΦ′ΥΛΥ′−σ−2CKX′sK′0]+i±[ΦΛΦ′XsX′s0]−r(Xs)−r[Xs,ΦΛΦ′] | (3.6) |
is obtained. From (1.10),
i+[ΦΛΦ′XsX′s0]=r[Xs,ΦΛΦ′] and i−[ΦΛΦ′XsX′s0]=r(Xs) | (3.7) |
are written and then, after setting C=MSEM(˜λ) in the first matrix in (3.6), (3.1) and (3.2) are obtained from (3.6) and (3.7). According to (1.5), adding the equalities in (3.1) and (3.2) yields (3.3). Applying Lemma 1.1 to (3.1)–(3.3) yields 1–5.
Corollary 3.1. Let S and λ be as given in (2.1) and (2.2), respectively. Suppose that ˜γ, ˜ε and ˜e are predictors for γ, ε and e, respectively. Let say μ=Xsα and ˆμ be any estimator for μ under S. Denote
E1=[ΦΛΦ′0Xs0−σ−2MSEM(ˆμ)XsX′sX′s0],E2=[ΦΛΦ′ΦΛ˘I′pXs˘IpΛΦ′˘IpΛ˘I′p−σ−2MSEM(˜γ)0X′s00],E3=[ΦΛΦ′ΦΛ˘I′nXs˘InΛΦ′˘InΛ˘I′n−σ−2MSEM(˜ε)0X′s00],E4=[ΦΛΦ′ΦΛ˘I′mXs˘ImΛΦ′˘ImΛ˘I′m−σ−2MSEM(˜e)0X′s00], |
where ˘Ip=[Ip,0,0], ˘In=[0,In,0] and ˘Im=[0,0,Im]. Then,
1) MSEM(ˆμ)≽MSEM(ˆμBLUE) ⇔ i+(E1)=r[Xs,ΦΛΦ′].
2) MSEM(ˆμ)≻MSEM(ˆμBLUE) ⇔ i−(E1)=r(Xs)+n+m.
3) MSEM(ˆμ)≼MSEM(ˆμBLUE) ⇔ i−(E1)=r(Xs).
4) MSEM(ˆμ)≺MSEM(ˆμBLUE) ⇔ i+(E1)=r[Xs,ΦΛΦ′]+n+m.
5) MSEM(ˆμ)=MSEM(ˆμBLUE) ⇔ r(E1)=r[Xs,ΦΛΦ′]+r(Xs).
6) MSEM(˜γ)≽MSEM(˜γBLUP) ⇔ i+(E2)=r[Xs,ΦΛΦ′].
7) MSEM(˜γ)≻MSEM(˜γBLUP) ⇔ i−(E2)=r(Xs)+p.
8) MSEM(˜γ)≼MSEM(˜γBLUP) ⇔ i−(E2)=r(Xs).
9) MSEM(˜γ)≺MSEM(˜γBLUP) ⇔ i+(E2)=r[Xs,ΦΛΦ′]+p.
10) MSEM(˜γ)=MSEM(˜γBLUP) ⇔ r(E2)=r[Xs,ΦΛΦ′]+r(Xs).
11) MSEM(˜ε)≽MSEM(˜εBLUP) ⇔ i+(E3)=r[Xs,ΦΛΦ′].
12) MSEM(˜ε)≻MSEM(˜εBLUP) ⇔ i−(E3)=r(Xs)+n.
13) MSEM(˜ε)≼MSEM(˜εBLUP) ⇔ i−(E3)=r(Xs).
14) MSEM(˜ε)≺MSEM(˜εBLUP) ⇔ i+(E3)=r[Xs,ΦΛΦ′]+n.
15) MSEM(˜ε)=MSEM(˜εBLUP) ⇔ r(E3)=r[Xs,ΦΛΦ′]+r(Xs).
16) MSEM(˜e)≽MSEM(˜eBLUP) ⇔ i+(E4)=r[Xs,ΦΛΦ′].
17) MSEM(˜e)≻MSEM(˜eBLUP) ⇔ i−(E4)=r(Xs)+m.
18) MSEM(˜e)≼MSEM(˜eBLUP) ⇔ i−(E4)=r(Xs).
19) MSEM(˜e)≺MSEM(˜eBLUP) ⇔ i+(E4)=r[Xs,ΦΛΦ′]+m.
20) MSEM(˜e)=MSEM(˜eBLUP) ⇔ r(E4)=r[Xs,ΦΛΦ′]+r(Xs).
Now, let us consider the LMM in (1.1) with an exact restriction Sα=s. Then, the model S in (2.1) corresponds the following CLMM:
R:ys=Xsα+Zsγ+εr=Xsα+[Zs,In+m][γεr], | (4.1) |
where ys, Xs and Zs are given as in (2.1) and εr=[ε0]. λ in (2.2) corresponds the vector
r=Kα+Lγ+Mεr=Kα+[L,M][γεr]=Kα+[L,M1,M2][γε0]. | (4.2) |
In this case, the assumptions under R are written as
D(ys)=σ2ΦΛrΦ′,D(r)=σ2ΥΛrΥ′,cov(r,ys)=σ2ΥΛrΦ′, |
where Λr=[Λ1000].
In the theorem given below, the results on the relationships between the MSEM of any predictor for r and the MSEM of the BLUP of r under R are collected. These results are obtained from Theorem 3.1.
Theorem 4.1. Suppose that r in (4.2) is predictable under R in (4.1). Let ˜r be any predictor (unbiased or biased) for r under R. Denote
Er=[ΦΛrΦ′ΦΛrΥ′XsΥΛrΦ′ΥΛrΥ′−σ−2MSEM(˜r)KX′sK′0]=[σ−2D(ys)σ−2cov(ys,r)Xsσ−2cov(r,ys)σ−2(D(r)−MSEM(˜r))KX′sK′0]. |
Then, the following results hold.
1) MSEM(˜r)≽MSEM(˜rBLUP) ⇔ i+(Er)=r[Xs,ΦΛrΦ′].
2) MSEM(˜r)≻MSEM(˜rBLUP) ⇔ i−(Er)=r(Xs)+t.
3) MSEM(˜r)≼MSEM(˜rBLUP) ⇔ i−(Er)=r(Xs).
4) MSEM(˜r)≺MSEM(˜rBLUP) ⇔ i+(Er)=r[Xs,ΦΛrΦ′]+t.
5) MSEM(˜r)=MSEM(˜rBLUP) ⇔ r(Er)=r[Xs,ΦΛrΦ′]+r(Xs).
Now, let us consider the LMM in (1.1) without any restrictions on parameters and show this model U, i.e.,
U:y=Xα+Zγ+ε=Xα+[Z,In][γε], | (4.3) |
and λ in (2.2) corresponds the vector
u=Kα+Lγ+M1ε=Kα+[L,M1][γε]. | (4.4) |
In this case,
D(y)=σ2[Z,In]Λ1[Z,In]′,D(u)=σ2[L,M1]Λ1[L,M1]′,cov(u,y)=σ2[L,M1]Λ1[Z,In]′. |
Theorem 4.2. Suppose that u in (4.4) is predictable under U in (4.3). Let ˜u be any predictor (unbiased or biased) for u under U. Denote
Eu=[Φ1Λ1Φ′1Φ1Λ1Υ′1XΥ1Λ1Φ′1Υ1Λ1Υ′1−σ−2MSEM(˜u)KX′K′0]=[σ−2D(y)σ−2cov(y,u)Xσ−2cov(u,y)σ−2(D(u)−MSEM(˜u))KX′K′0], |
where Φ1=[Z,In] and Υ1=[L,M1]. Then, the following results hold.
1) MSEM(˜u)≽MSEM(˜uBLUP) ⇔ i+(Eu)=r[X,Φ1Λ1Φ′1].
2) MSEM(˜u)≻MSEM(˜uBLUP) ⇔ i−(Eu)=r(X)+t.
3) MSEM(˜u)≼MSEM(˜uBLUP) ⇔ i−(Eu)=r(X).
4) MSEM(˜u)≺MSEM(˜uBLUP) ⇔ i+(Eu)=r[X,Φ1Λ1Φ′1]+t.
5) MSEM(˜u)=MSEM(˜uBLUP) ⇔ r(Eu)=r[X,Φ1Λ1Φ′1]+r(X).
As is well-known, two types of commonly used estimators under a general linear model are the BLUEs and the ordinary least squares estimators (OLSEs). Both of these estimators have a variety of simple and remarkable properties. Therefore, they have attracted statisticians' attention throughout the historical development of regression theory, and numerous results regarding the BLUEs and the OLSEs have been established. Since these estimators are defined by different optimality criteria and thereby their expressions and properties are not necessarily the same, it is natural to seek possible connections between them. As a subject area within the regression analysis, the relationships between the BLUEs and OLSEs have been widely considered in the statistical literature and various identifying conditions for their equivalence or comparisons have been obtained. Based on these explanations, to further clarify the results obtained above, since both the BLUE and the OLSE of Xsα, denoted as ^XsαOLSE, are well-known estimators, we present the following result on the relationship between them.
Let us consider the LMM in (1.1) by setting Z=0 with an exact restriction Sα=s. Then, the model S in (2.1) corresponds the following linear model:
L:ys=Xsα+εr, | (4.5) |
where ys and Xs are given as in (2.1) and εr is given as in (4.1). By taking λ=Xsα in (2.2), the matrix E in Theorem 1 corresponds
El=[Λr0Xs0−σ−2MSEM(^XsαOLSE)XsX′sX′s0], | (4.6) |
where {\boldsymbol{\Lambda}}_{r} = \begin{bmatrix} {\boldsymbol{\Lambda}}_{11} & \mathbf{{0}} \\ \mathbf{{0}} & \mathbf{{0}} \end{bmatrix} and {\rm{D}}({\boldsymbol{\varepsilon}}) = {\boldsymbol{\Lambda}}_{11} . Then,
\begin{equation*} \boldsymbol{i}_{+}( \mathbf{{E}}_{l}) = \boldsymbol{i}_{+}\begin{bmatrix} {\boldsymbol{\Lambda}}_{r} & \mathbf{{0}} & \mathbf{{X}}_{s}\\ \mathbf{{0}} & - \mathbf{{X}}_{s} \mathbf{{X}}_{s}^{+} {\boldsymbol{\Lambda}}_{r} \mathbf{{X}}_{s} \mathbf{{X}}_{s}^{+} & \mathbf{{X}}_{s} \\ \mathbf{{X}}_{s}' & \mathbf{{X}}_{s}' & \mathbf{{0}} \end{bmatrix} = \boldsymbol{r}\begin{bmatrix} {\boldsymbol{\Lambda}}_{r} & \mathbf{{0}} & \mathbf{{X}}_{s}\\ \mathbf{{0}} & - \mathbf{{X}}_{s} \mathbf{{X}}_{s}^{+} {\boldsymbol{\Lambda}}_{r} \mathbf{{X}}_{s} \mathbf{{X}}_{s}^{+} & \mathbf{{X}}_{s} \end{bmatrix} = \boldsymbol{r}\begin{bmatrix} {\boldsymbol{\Lambda}}_{r}, & \mathbf{{X}}_{s} \end{bmatrix} \end{equation*} |
always holds, i.e., {\rm{MSEM}}(\widehat{ \mathbf{{X}}_{s} {\boldsymbol{\alpha}}}_{ {\rm{OLSE}}}) \succcurlyeq {\rm{MSEM}}(\widehat{ \mathbf{{X}}_{s} {\boldsymbol{\alpha}}}_{ {\rm{BLUE}}}) . This inequality is already a well-known result in statistical theory.
We now use Example 4.1.8 from Section 4.1 of [46] to illustrate our theoretical findings in response to the referees' advice. Consider the linear model \mathbf{{y}} = \mathbf{{X}} {\boldsymbol{\alpha}}+ {\boldsymbol{\varepsilon}} with
\begin{equation*} \mathbf{{X}} = \begin{bmatrix} \mathbf{{1}} & \mathbf{{1}} & \mathbf{{0}} & \mathbf{{1}} & \mathbf{{0}}\\ \mathbf{{1}} & \mathbf{{0}} & \mathbf{{1}} & \mathbf{{1}} & \mathbf{{0}} \\ \mathbf{{1}} & \mathbf{{1}} & \mathbf{{0}} & \mathbf{{0}} & \mathbf{{1}} \\ \mathbf{{1}} & \mathbf{{0}} & \mathbf{{1}} & \mathbf{{0}} & \mathbf{{1}} \end{bmatrix} \; \hbox{ and } {\boldsymbol{\alpha}} = \begin{bmatrix} \mu\\ \beta_{1}\\ \beta_{2}\\ \tau_{1}\\ \tau_{2}\\ \end{bmatrix}, \end{equation*} |
where \mathbf{{1}}\in\mathbb{R}_{10, 1} and \mathbf{{0}}\in\mathbb{R}_{10, 1} . Let consider
\begin{equation*} {\rm{D}}( {\boldsymbol{\varepsilon}}) = {\boldsymbol{\Lambda}}_{11} = \begin{bmatrix} 3 \mathbf{{1}} \mathbf{{1}}' & 2 \mathbf{{1}} \mathbf{{1}}' & 2 \mathbf{{1}} \mathbf{{1}}' & \mathbf{{1}} \mathbf{{1}}'\\ 2 \mathbf{{1}} \mathbf{{1}}' &3 \mathbf{{1}} \mathbf{{1}}' & \mathbf{{1}} \mathbf{{1}}' &2 \mathbf{{1}} \mathbf{{1}}' \\ 2 \mathbf{{1}} \mathbf{{1}}' & \mathbf{{1}} \mathbf{{1}}' &3 \mathbf{{1}} \mathbf{{1}}' &2 \mathbf{{1}} \mathbf{{1}}' \\ \mathbf{{1}} \mathbf{{1}}' &2 \mathbf{{1}} \mathbf{{1}}' &2 \mathbf{{1}} \mathbf{{1}}' &3 \mathbf{{1}} \mathbf{{1}}' \end{bmatrix} \end{equation*} |
as in [2]. This designed set-up is typical in agricultural experiments where several treatments are applied to various blocks of land. The experiment is often conducted to assess the differential impact of the treatments. Here, the parameter \mu represents a general effect that is present in all the observations, and the parameters \beta_{1} and \beta_{2} represent the respective effects of two blocks, and the parameters \tau_{1} and \tau_{2} represent the respective effects of two treatments. Suppose that
\begin{equation*} \tau_{1}-\tau_{2} = 0, \; \hbox{i.e.}, \; \mathbf{{S}} {\boldsymbol{\alpha}} = 0, \end{equation*} |
where \mathbf{{S}} = \begin{bmatrix} 0, & 0, & 0, & 1, & -1 \end{bmatrix} . This restriction may be a known fact from the theory or experiment view.
Direct calculations will show that
\begin{equation*} {\rm{MSEM}}(\widehat{ \mathbf{{X}}_{s} {\boldsymbol{\alpha}}}_{ {\rm{OLSE}}}) = \begin{bmatrix} 2.91 \mathbf{{1}} \mathbf{{1}}' & 1.91 \mathbf{{1}} \mathbf{{1}}' & 2.08 \mathbf{{1}} \mathbf{{1}}' &1.08 \mathbf{{1}} \mathbf{{1}}' &0.82 \mathbf{{1}}\\ 1.91 \mathbf{{1}} \mathbf{{1}}' &2.91 \mathbf{{1}} \mathbf{{1}}' &1.08 \mathbf{{1}} \mathbf{{1}}' &2.08 \mathbf{{1}} \mathbf{{1}}'&0.82 \mathbf{{1}} \\ 2.08 \mathbf{{1}} \mathbf{{1}}' &1.08 \mathbf{{1}} \mathbf{{1}}' &2.91 \mathbf{{1}} \mathbf{{1}}' &1.91 \mathbf{{1}} \mathbf{{1}}'&-0.82 \mathbf{{1}} \\ 1.08 \mathbf{{1}} \mathbf{{1}}' &2.08 \mathbf{{1}} \mathbf{{1}}' &1.91 \mathbf{{1}} \mathbf{{1}}' &2.91 \mathbf{{1}} \mathbf{{1}}'&-0.82 \mathbf{{1}} \\ 0.82 \mathbf{{1}}' &0.82 \mathbf{{1}}' &-0.82 \mathbf{{1}}' &-0.82 \mathbf{{1}}' &1.65 \end{bmatrix} \end{equation*} |
and
\begin{equation*} {\rm{MSEM}}(\widehat{ \mathbf{{X}}_{s} {\boldsymbol{\alpha}}}_{ {\rm{BLUE}}}) = \begin{bmatrix} 2.5 \mathbf{{1}} \mathbf{{1}}' & 1.5 \mathbf{{1}} \mathbf{{1}}' & 2.5 \mathbf{{1}} \mathbf{{1}}' &1.5 \mathbf{{1}} \mathbf{{1}}' &5.97\times 10^{-15} \mathbf{{1}}\\ 1.5 \mathbf{{1}} \mathbf{{1}}' &2.5 \mathbf{{1}} \mathbf{{1}}' &1.5 \mathbf{{1}} \mathbf{{1}}' &2.5 \mathbf{{1}} \mathbf{{1}}' &4.02\times 10^{-15} \mathbf{{1}} \\ 2.5 \mathbf{{1}} \mathbf{{1}}' &1.5 \mathbf{{1}} \mathbf{{1}}' &2.5 \mathbf{{1}} \mathbf{{1}}' &1.5 \mathbf{{1}} \mathbf{{1}}' &5.97\times 10^{-15} \mathbf{{1}} \\ 1.5 \mathbf{{1}} \mathbf{{1}}' &2.5 \mathbf{{1}} \mathbf{{1}}' &1.5 \mathbf{{1}} \mathbf{{1}}' &2.5 \mathbf{{1}} \mathbf{{1}}' &4.02\times 10^{-15} \mathbf{{1}} \\ 5.97\times 10^{-15} \mathbf{{1}}' &4.02\times 10^{-15} \mathbf{{1}}' &5.97\times 10^{-15} \mathbf{{1}}' &4.02\times 10^{-15} \mathbf{{1}}' &1.19\times 10^{-28} \end{bmatrix}, \end{equation*} |
when we consider above matrices for the model {\mathcal{L}} in (4.5). Now it is easy to see that
\begin{equation*} \begin{split} &\begin{bmatrix} 2.91 \mathbf{{1}} \mathbf{{1}}' & 1.91 \mathbf{{1}} \mathbf{{1}}' & 2.08 \mathbf{{1}} \mathbf{{1}}' &1.08 \mathbf{{1}} \mathbf{{1}}' &0.82 \mathbf{{1}}\\ 1.91 \mathbf{{1}} \mathbf{{1}}' &2.91 \mathbf{{1}} \mathbf{{1}}' &1.08 \mathbf{{1}} \mathbf{{1}}' &2.08 \mathbf{{1}} \mathbf{{1}}'&0.82 \mathbf{{1}} \\ 2.08 \mathbf{{1}} \mathbf{{1}}' &1.08 \mathbf{{1}} \mathbf{{1}}' &2.91 \mathbf{{1}} \mathbf{{1}}' &1.91 \mathbf{{1}} \mathbf{{1}}'&-0.82 \mathbf{{1}} \\ 1.08 \mathbf{{1}} \mathbf{{1}}' &2.08 \mathbf{{1}} \mathbf{{1}}' &1.91 \mathbf{{1}} \mathbf{{1}}' &2.91 \mathbf{{1}} \mathbf{{1}}'&-0.82 \mathbf{{1}} \\ 0.82 \mathbf{{1}}' &0.82 \mathbf{{1}}' &-0.82 \mathbf{{1}}' &-0.82 \mathbf{{1}}' &1.65 \end{bmatrix} \succcurlyeq \\ & \quad \quad \begin{bmatrix} 2.5 \mathbf{{1}} \mathbf{{1}}' & 1.5 \mathbf{{1}} \mathbf{{1}}' & 2.5 \mathbf{{1}} \mathbf{{1}}' &1.5 \mathbf{{1}} \mathbf{{1}}' &5.97\times 10^{-15} \mathbf{{1}}\\ 1.5 \mathbf{{1}} \mathbf{{1}}' &2.5 \mathbf{{1}} \mathbf{{1}}' &1.5 \mathbf{{1}} \mathbf{{1}}' &2.5 \mathbf{{1}} \mathbf{{1}}' &4.02\times 10^{-15} \mathbf{{1}} \\ 2.5 \mathbf{{1}} \mathbf{{1}}' &1.5 \mathbf{{1}} \mathbf{{1}}' &2.5 \mathbf{{1}} \mathbf{{1}}' &1.5 \mathbf{{1}} \mathbf{{1}}' &5.97\times 10^{-15} \mathbf{{1}} \\ 1.5 \mathbf{{1}} \mathbf{{1}}' &2.5 \mathbf{{1}} \mathbf{{1}}' &1.5 \mathbf{{1}} \mathbf{{1}}' &2.5 \mathbf{{1}} \mathbf{{1}}' &4.02\times 10^{-15} \mathbf{{1}} \\ 5.97\times 10^{-15} \mathbf{{1}}' &4.02\times 10^{-15} \mathbf{{1}}' &5.97\times 10^{-15} \mathbf{{1}}' &4.02\times 10^{-15} \mathbf{{1}}' &1.19\times 10^{-28} \end{bmatrix}, \end{split} \end{equation*} |
i.e.,
\begin{equation*} {\rm{MSEM}}(\widehat{ \mathbf{{X}}_{s} {\boldsymbol{\alpha}}}_{ {\rm{OLSE}}}) \succcurlyeq {\rm{MSEM}}(\widehat{ \mathbf{{X}}_{s} {\boldsymbol{\alpha}}}_{ {\rm{BLUE}}}). \end{equation*} |
Using the numpy library in Python and setting the above findings in the matrix \mathbf{{E}}_{l} in (4.6), we find the eigenvalues of matrix \mathbf{{E}}_{l} as 1.44, 20.98, 20.98, 80.99, -80.99, -20.98, -19.36 and -1.25. Thus, we can see that the number of positive eigenvalues of \mathbf{{E}}_{l} is 4 and the number of negative eigenvalues of \mathbf{{E}}_{l} is 4, i.e., \boldsymbol{i}_{+}(\mathbf{{E}}_{l}) = 4 and \boldsymbol{i}_{-}(\mathbf{{E}}_{l}) = 4 . It is easily seen from (1.5) that \boldsymbol{r}(\mathbf{{E}}_{l}) = 8 . Also, we obtain \boldsymbol{r}\begin{bmatrix} \mathbf{{X}}_s, & {\boldsymbol{\Lambda}}_{r} \end{bmatrix} = 4 and \boldsymbol{r}(\mathbf{{X}}_{s}) = 3 . Therefore, \boldsymbol{i}_{+}(\mathbf{{E}}_{l}) = \boldsymbol{r}\begin{bmatrix} \mathbf{{X}}_s, & {\boldsymbol{\Lambda}}_{r} \end{bmatrix} = 4 \Leftrightarrow {\rm{MSEM}}(\widehat{ \mathbf{{X}}_{s} {\boldsymbol{\alpha}}}_{ {\rm{BLUE}}}) \preccurlyeq {\rm{MSEM}}(\widehat{ \mathbf{{X}}_{s} {\boldsymbol{\alpha}}}_{ {\rm{OLSE}}}) holds. Since \boldsymbol{r}(\mathbf{{X}}_{s}) = 3 , i.e., \boldsymbol{i}_{-}(\mathbf{{E}}_{l})\ne \boldsymbol{r}(\mathbf{{X}}_s) , {\rm{MSEM}}(\widehat{ \mathbf{{X}}_{s} {\boldsymbol{\alpha}}}_{ {\rm{OLSE}}}) \preccurlyeq {\rm{MSEM}}(\widehat{ \mathbf{{X}}_{s} {\boldsymbol{\alpha}}}_{ {\rm{BLUE}}}) is not provided.
The findings of this study, according to the {\rm{MSEM}} criterion, provide a broad perspective on the {\rm{SRLMM}} s in the context of predictor comparability problems using the methodology of block matrix inertias and ranks. We establish the comparison results after converting an explicitly {\rm{SRLMM}} to an implicitly linear stochastically restricted {\rm{LMM}} . We also reduce our results to models {\rm{CLMM}} s and {\rm{ULMM}} s.
We derive the formulas of comparisons of any predictor and the {\rm{BLUP}} for the same general vector of all unknown vectors under {\rm{SRLMM}} by considering the {\rm{MSEM}} sense. Thus, the performance of a new biased or unbiased predictor according to the dispersion matrices of {\rm{BLUP}} can be examined with various inequalities and equalities derived among {\rm{MSEM}} s based on inertias and ranks. These kinds of inequalities or equalities for comparisons have a useful and strong statistical explanation in the theory of linear statistical models and applications. The comparisons of MSEMs to determine the effectiveness of any predictor according to BLUP are explained by the comparisons of some quantities that come from inertias and ranks.
As a result, we can say that our contribution to issues involving inference and prediction under an {\rm{LMM}} with a linear stochastic restriction on unknown parameters in the literature is to serve as a theoretically comprehensive and significant search for the comparison results between any predictors and {\rm{BLUP}} s.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
The authors would like to express their sincere thanks to the handling editor and anonymous reviewers for their helpful comments and suggestions.
The authors state that they have no competing interest.
[1] |
H. Haupt, W. Oberhofer, Stochastic response restrictions, J. Multivariate Anal., 95 (2005), 66–75. https://doi.org/10.1016/j.jmva.2004.08.006 doi: 10.1016/j.jmva.2004.08.006
![]() |
[2] |
X. Ren, Corrigendum to "On the equivalence of the BLUEs under a general linear model and its restricted and stochastically restricted models" [Stat. Probabil. Lett. 90 (2014) 1–10], Stat. Probabil. Lett., 104 (2015), 181–185. https://doi.org/10.1016/j.spl.2015.05.004 doi: 10.1016/j.spl.2015.05.004
![]() |
[3] |
J. K. Baksalary, R. Kala, Best linear unbiased estimation in the restricted general linear model, Series Statistics, 10 (1979), 27–35. https://doi.org/10.1080/02331887908801464 doi: 10.1080/02331887908801464
![]() |
[4] |
J. S. Chipman, M. M. Rao, The treatment of linear restrictions in regression analysis, Econometrica, 32 (1964), 198–209. https://doi.org/10.2307/1913745 doi: 10.2307/1913745
![]() |
[5] |
W. T. Dent, On restricted estimation in linear models, J. Econometrics, 12 (1980), 49–58. https://doi.org/10.1016/0304-4076(80)90052-4 doi: 10.1016/0304-4076(80)90052-4
![]() |
[6] |
N. Güler, M. E. Büyükkaya, Further remarks on constrained over-parameterized linear models, Stat. Papers, 2023 (2023), 01426. https://doi.org/10.1007/s00362-023-01426-z doi: 10.1007/s00362-023-01426-z
![]() |
[7] |
C. R. Hallum, T. O. Lewis, T. L. Boullion, Estimation in the restricted general linear model with a positive semidefinite covariance matrix, Commun. Stat., 1 (1973), 157–166. https://doi.org/10.1080/03610927308827014 doi: 10.1080/03610927308827014
![]() |
[8] |
B. Jiang, Y. Tian, On best linear unbiased estimation and prediction under a constrained linear random-effects model, J. Ind. Manag. Optim., 19 (2023), 852–867. https://doi.org/10.3934/jimo.2021209 doi: 10.3934/jimo.2021209
![]() |
[9] |
H. Jiang, J. Qian, Y. Sun, Best linear unbiased predictors and estimators under a pair of constrained seemingly unrelated regression models, Stat. Probabil. Lett., 158 (2020), 108669. https://doi.org/10.1016/j.spl.2019.108669 doi: 10.1016/j.spl.2019.108669
![]() |
[10] |
W. Li, Y. Tian, R. Yuan, Statistical analysis of a linear regression model with restrictions and superfluous variables, J. Ind. Manag. Optim., 19 (2023), 3107-3127. https://doi.org/10.3934/jimo.2022079 doi: 10.3934/jimo.2022079
![]() |
[11] |
C. Lu, Y. Sun, Y. Tian, A comparison between two competing fixed parameter constrained general linear models with new regressors, Statistics, 52 (2018), 769–781. https://doi.org/10.1080/02331888.2018.1469021 doi: 10.1080/02331888.2018.1469021
![]() |
[12] |
T. Mathew, A note on best linear unbiased estimation in the restricted general linear model, Series Statistics, 14 (1983), 3–6. https://doi.org/10.1080/02331888308801679 doi: 10.1080/02331888308801679
![]() |
[13] |
C. A. McGilchrist, C. W. Aisbett, Restricted BLUP for mixed linear models, Biometrical J., 33 (1991), 131–141. https://doi.org/10.1002/bimj.4710330202 doi: 10.1002/bimj.4710330202
![]() |
[14] |
Y. Tian, J. Wang, Some remarks on fundamental formulas and facts in the statistical analysis of a constrained general linear model, Commun. Stat. Theor. M., 49 (2020), 1201–1216. https://doi.org/10.1080/03610926.2018.1554138 doi: 10.1080/03610926.2018.1554138
![]() |
[15] |
S. J. Haslett, S. Puntanen, A note on the equality of the BLUPs for new observations under two linear models, Acta Comment. Univ. Tartu. Math., 14 (2010), 27–33. https://doi.org/10.12697/ACUTM.2010.14.03 doi: 10.12697/ACUTM.2010.14.03
![]() |
[16] |
H. Theil, On the use of incomplete prior information in regression analysis, J. Am. Stat. Assoc., 58 (1963), 401–414. https://doi.org/10.2307/2283275 doi: 10.2307/2283275
![]() |
[17] |
H. Theil, A. S. Goldberger, On pure and mixed statistical etimation in economics, Int. Econ. Rev., 2 (1961), 65–78. https://doi.org/10.2307/2525589 doi: 10.2307/2525589
![]() |
[18] |
J. Xu, H. Yang, Estimation in singular linear models with stochastic linear restrictions, Commun. Stat. Theor. M., 36 (2007), 1945–1951. https://doi.org/10.1080/03610920601126530 doi: 10.1080/03610920601126530
![]() |
[19] |
M. E. Büyükkaya, Characterizing relationships between BLUPs under linear mixed model and some associated reduced models, Commun. Stat. Simul. C., 2022 (2022), 2115071. https://doi.org/10.1080/03610918.2022.2115071 doi: 10.1080/03610918.2022.2115071
![]() |
[20] |
B. Dong, W. Guo, Y. Tian, On relations between BLUEs under two transformed linear models, J. Multivariate Anal., 131 (2014), 279–292. https://doi.org/10.1016/j.jmva.2014.07.005 doi: 10.1016/j.jmva.2014.07.005
![]() |
[21] |
N. Güler, On relations between BLUPs under two transformed linear random-effects models, Commun. Stat. Simul. C., 51 (2022), 5099–5125. https://doi.org/10.1080/03610918.2020.1757709 doi: 10.1080/03610918.2020.1757709
![]() |
[22] |
N. Güler, M. E. Büyükkaya, Notes on comparison of covariance matrices of BLUPs under linear random-effects model with its two sub-sample models, Iran. J. Sci. Technol. Trans. Sci., 43 (2019), 2993–3002. https://doi.org/10.1007/s40995-019-00785-3 doi: 10.1007/s40995-019-00785-3
![]() |
[23] |
N. Güler, M. E. Büyükkaya, Rank and inertia formulas for covariance matrices of BLUPs in general linear mixed models, Commun. Stat. Theor. M., 50 (2021), 4997–5012. https://doi.org/10.1080/03610926.2019.1599950 doi: 10.1080/03610926.2019.1599950
![]() |
[24] |
N. Güler, M. E. Büyükkaya, Some remarks on comparison of predictors in seemingly unrelated linear mixed models, Appl. Math., 67 (2022), 525–542. https://doi.org/10.21136/AM.2021.0366-20 doi: 10.21136/AM.2021.0366-20
![]() |
[25] |
N. Güler, M. E. Büyükkaya, Inertia and rank approach in transformed linear mixed models for comparison of BLUPs, Commun. Stat. Theor. M., 52 (2023), 3108–3123. https://doi.org/10.1080/03610926.2021.1967397 doi: 10.1080/03610926.2021.1967397
![]() |
[26] |
N. Güler, M. E. Büyükkaya, M. Yiğit, Comparison of covariance matrices of predictors in seemingly unrelated regression models, Indian J. Pure Appl. Math., 53 (2022), 801–809. https://doi.org/10.1007/s13226-021-00174-w doi: 10.1007/s13226-021-00174-w
![]() |
[27] |
Y. Tian, Some equalities and inequalities for covariance matrices of estimators under linear model, Stat. Papers, 58 (2017), 467–484. https://doi.org/10.1007/s00362-015-0707-x doi: 10.1007/s00362-015-0707-x
![]() |
[28] |
Y. Tian, W. Guo, On comparison of dispersion matrices of estimators under a constrained linear model, Stat. Methods Appl., 25 (2016), 623–649. https://doi.org/10.1007/s10260-016-0350-2 doi: 10.1007/s10260-016-0350-2
![]() |
[29] |
M. A. E. Abdelrahman, M. A. Sohaly, S. I. Ammar, Y. F. Alharbi, The deterministic and stochastic solutions for the nonlinear Phi-4 equation, Int. J. Nonlin. Sci. Num., 23 (2022), 823–832. https://doi.org/10.1515/ijnsns-2022-2272 doi: 10.1515/ijnsns-2022-2272
![]() |
[30] |
H. G. Abdelwahed, A. F. Alsarhana, E. K. El-Shewy, M. A. E. Abdelrahman, The stochastic structural modulations in collapsing Maccari's model solitons, Fractal Fract., 7 (2023), 290. https://doi.org/10.3390/fractalfract7040290 doi: 10.3390/fractalfract7040290
![]() |
[31] |
Y. F. Alharbi, E. K. El-Shewy, M. A. E. Abdelrahman, New and effective solitary applications in Schrödinger equation via Brownian motion process with physical coefficients of fiber optics, AIMS Mathematics, 8 (2023), 4126–4140. https://doi.org/10.3934/math.2023205 doi: 10.3934/math.2023205
![]() |
[32] |
F. Mirzaee, S. Rezaei, N. Samadyar, Solving one-dimensional nonlinear stochastic Sine-Gordon equation with a new meshfree technique, Int. J. Numer. Model. El., 34 (2021), e2856. https://doi.org/10.1002/jnm.2856 doi: 10.1002/jnm.2856
![]() |
[33] |
F. Mirzaee, N. Samadyar, Combination of finite difference method and meshless method based on radial basis functions to solve fractional stochastic advection-diffusion equations, Eng. Comput., 36 (2020), 1673–1686. https://doi.org/10.1007/s00366-019-00789-y doi: 10.1007/s00366-019-00789-y
![]() |
[34] |
E. K. El-Shewy, Y. F. Alharbi, M. A. E. Abdelrahman, On the dynamical stochastic electrostatic noise fluctuations in Zakharov model, Chaos Soliton. Fract., 170 (2023), 113324. https://doi.org/10.1016/j.chaos.2023.113324 doi: 10.1016/j.chaos.2023.113324
![]() |
[35] |
Y. Tian, Equalities and inequalities for inertias of Hermitian matrices with applications, Linear Algebra Appl., 433 (2010), 263–296. https://doi.org/10.1016/j.laa.2010.02.018 doi: 10.1016/j.laa.2010.02.018
![]() |
[36] |
Y. Tian, Solving optimization problems on ranks and inertias of some constrained nonlinear matrix functions via an algebraic linearization method, Nonlinear Anal. Theor., 75 (2012), 717–734. https://doi.org/10.1016/j.na.2011.09.003 doi: 10.1016/j.na.2011.09.003
![]() |
[37] |
C. R. Rao, Representations of best linear unbiased estimators in the Gauss-Markoff model with a singular dispersion matrix, J. Multivariate Anal., 3 (1973), 276–292. https://doi.org/10.1016/0047-259X(73)90042-0 doi: 10.1016/0047-259X(73)90042-0
![]() |
[38] |
I. S. Alalouf, G. P. H. Styan, Characterizations of estimability in the general linear model, Ann. Statist., 7 (1979), 194–200. https://doi.org/10.1214/aos/1176344564 doi: 10.1214/aos/1176344564
![]() |
[39] |
C. Lu, S. Gan, Y. Tian, Some remarks on general linear model with new regressors, Stat. Probabil. Lett., 97 (2015), 16–24. https://doi.org/10.1016/j.spl.2014.10.015 doi: 10.1016/j.spl.2014.10.015
![]() |
[40] |
A. S. Goldberger, Best linear unbiased prediction in the generalized linear regression model, J. Am. Stat. Assoc., 57 (1962), 369–375. https://doi.org/10.2307/2281645 doi: 10.2307/2281645
![]() |
[41] | S. Puntanen, G. P. H. Styan, J. Isotalo, Matrix tricks for linear statistical models: Our personal top twenty, 1 Eds., Heidelberg: Springer Berlin, 2011. https://doi.org/10.1007/978-3-642-10473-2 |
[42] | C. R. Rao, Shalabh, H. Toutenburg, C. Heumann, Linear models and generalizations: Least squares and alternatives, 3 Eds., Heidelberg: Springer Berlin, 2008. https://doi.org/10.1007/978-3-540-74227-2 |
[43] |
H. Yang, H. Ye, K. Xue, A further study ofpredictions in linear mixed models, Commun. Stat. Theor. M., 43 (2014), 4241–4252. https://doi.org/10.1080/03610926.2012.725497 doi: 10.1080/03610926.2012.725497
![]() |
[44] | H. Drygas, The coordinate-free approach to Gauss-Markov estimation, 1 Eds., Heidelberg: Springer Berlin, 1970. https://doi.org/10.1007/978-3-642-65148-9 |
[45] |
Y. Tian, On properties of BLUEs under general linear regression models, J. Stat. Plan. Infer., 143 (2013), 771–782. https://doi.org/10.1016/j.jspi.2012.10.005 doi: 10.1016/j.jspi.2012.10.005
![]() |
[46] | D. Sengupta, S. R. Jammalamadaka, Linear models: An integrated approach, Singapore: World Scientific Press, 2003. https://doi.org/10.1142/4674 |
1. | Bo Jiang, Yongge Tian, On a Biased Prediction Based on Optimal Mean Square Error Criterion, 2024, 47, 0126-6705, 10.1007/s40840-024-01773-6 | |
2. | Nesrin Güler, Analysis of Optimal Prediction Under Stochastically Restricted Linear Model and Its Subsample Models, 2024, 13, 2075-1680, 882, 10.3390/axioms13120882 | |
3. | Nesrin Güler, Melek Eriş Büyükkaya, A matrix approach to determine optimal predictors in a constrained linear mixed model, 2024, 22, 2391-5455, 10.1515/math-2024-0114 |