This article offers a general procedure of carrying out estimation and inference under a linear statistical model y=Xββ+εε with an adding-up restriction Ay=b to the observed random vector y. We first propose an available way of converting the adding-up restrictions to a linear matrix equation for ββ and a matrix equality for the covariance matrix of the error term εε, which can help in combining the two model equations in certain consistent form. We then give the derivations and presentations of analytic expressions of the ordinary least-squares estimator (OLSE) and the best linear unbiased estimator (BLUE) of parametric vector Kββ using various analytical algebraic operations of the given vectors and matrices in the model.
Citation: Yongge Tian. An effective treatment of adding-up restrictions in the inference of a general linear model[J]. AIMS Mathematics, 2023, 8(7): 15189-15200. doi: 10.3934/math.2023775
[1] | Bo Jiang, Yongge Tian . Equivalent analysis of different estimations under a multivariate general linear model. AIMS Mathematics, 2024, 9(9): 23544-23563. doi: 10.3934/math.20241144 |
[2] | Nesrin Güler, Melek Eriş Büyükkaya . Statistical inference of a stochastically restricted linear mixed model. AIMS Mathematics, 2023, 8(10): 24401-24417. doi: 10.3934/math.20231244 |
[3] | Ruixia Yuan, Bo Jiang, Yongge Tian . A study of the equivalence of inference results in the contexts of true and misspecified multivariate general linear models. AIMS Mathematics, 2023, 8(9): 21001-21021. doi: 10.3934/math.20231069 |
[4] | Victoria May P. Mendoza, Renier Mendoza, Youngsuk Ko, Jongmin Lee, Eunok Jung . Managing bed capacity and timing of interventions: a COVID-19 model considering behavior and underreporting. AIMS Mathematics, 2023, 8(1): 2201-2225. doi: 10.3934/math.2023114 |
[5] | Junke Kou, Xianmei Chen . Wavelet estimations of a density function in two-class mixture model. AIMS Mathematics, 2024, 9(8): 20588-20611. doi: 10.3934/math.20241000 |
[6] | Muhammad Bilal Khan, Hakeem A. Othman, Gustavo Santos-García, Muhammad Aslam Noor, Mohamed S. Soliman . Some new concepts in fuzzy calculus for up and down λ-convex fuzzy-number valued mappings and related inequalities. AIMS Mathematics, 2023, 8(3): 6777-6803. doi: 10.3934/math.2023345 |
[7] | Dayang Dai, Dabuxilatu Wang . A generalized Liu-type estimator for logistic partial linear regression model with multicollinearity. AIMS Mathematics, 2023, 8(5): 11851-11874. doi: 10.3934/math.2023600 |
[8] | W. B. Altukhaes, M. Roozbeh, N. A. Mohamed . Feasible robust Liu estimator to combat outliers and multicollinearity effects in restricted semiparametric regression model. AIMS Mathematics, 2024, 9(11): 31581-31606. doi: 10.3934/math.20241519 |
[9] | Ronnason Chinram, Aiyared Iampan . Codewords generated by UP-valued functions. AIMS Mathematics, 2021, 6(5): 4771-4785. doi: 10.3934/math.2021280 |
[10] | Sen Ming, Xiaodong Wang, Xiongmei Fan, Xiao Wu . Blow-up of solutions for coupled wave equations with damping terms and derivative nonlinearities. AIMS Mathematics, 2024, 9(10): 26854-26876. doi: 10.3934/math.20241307 |
This article offers a general procedure of carrying out estimation and inference under a linear statistical model y=Xββ+εε with an adding-up restriction Ay=b to the observed random vector y. We first propose an available way of converting the adding-up restrictions to a linear matrix equation for ββ and a matrix equality for the covariance matrix of the error term εε, which can help in combining the two model equations in certain consistent form. We then give the derivations and presentations of analytic expressions of the ordinary least-squares estimator (OLSE) and the best linear unbiased estimator (BLUE) of parametric vector Kββ using various analytical algebraic operations of the given vectors and matrices in the model.
Throughout, let Rm×n stand for the collection of all m×n matrices over the field of real numbers; A′, r(A), and R(A) stand for the transpose, the rank, and the range (column space) of a matrix A∈Rm×n, respectively; and let Im denote the identity matrix of order m. Given an A∈Rm×n, the Moore–Penrose generalized inverse of A, denoted by A+, is defined to be the unique solution G satisfying the four matrix equations AGA=A, GAG=G, (AG)′=AG, and (GA)′=GA. Further, let PA, EA, and FA stand for the three orthogonal projectors (symmetric idempotent matrices) PA=AA+, EA=A⊥=Im−AA+, and FA=In−A+A, which will help in briefly denoting calculation processes related to generalized inverses of matrices. Further information about the orthogonal projectors PA, EA, and FA with their applications in the linear statistical models can be found, e.g., in [5,9]. Two symmetric matrices A and B of the same size are said to satisfy the inequality A≽B in the Löwner partial ordering if A−B is nonnegative definite. For more results on the Löwner partial ordering of real symmetric matrices and its applications in statistical analysis, see e.g., [9,17].
Consider a general linear model
M: y=Xββ+εε, E(εε)=0, Cov(εε)=σ2ΣΣ, | (1.1) |
where it is assumed that y∈Rn×1 is a vector of observable random variables, X∈Rn×p is a known model matrix of arbitrary rank (0≤r(X)≤min{n,p}), ββ∈Rn×1 is a vector of fixed but unknown parameters, σ2 is an arbitrary positive scaling factor, and ΣΣ∈Rn×n is a known nonnegative definite matrix of arbitrary rank (0≤r(ΣΣ)≤n), for example ΣΣ=I.
Below, we present some background details of the work. For a variety of reasons, statisticians may meet with the situation where certain restrictions are imposed on the unknown coefficient vector ββ and the observed random vector y in (1.1). For example, it is a regular case to add a system of linear matrix equations Bββ=c to the unknown parameter vector under the assumption in (1.1), and there have been plenty of approaches and discussions addressing how to carry out statistical inference under linear models with restrictions to their unknown parameters. In addition to the situations imposing restrictions on unknown parameters, it is necessary to take into account as well the situations of adding certain limitations and restrictions upon observed random variables in the model from theoretical and applied points of view. Under the model assumption in (1.1), one of such considerations is assuming that the observed random vector y satisfies a consistent linear matrix equation
Ay=b, | (1.2) |
where it is assumed that A∈Rm×n is a known matrix with rank(A)=k≤min{m,n} and b∈Rm×1 is a known vector with b∈R(A). A matrix equation as given in (1.2) is usually called adding-up restrictions to y in (1.1) in the literature. Clearly, the adding-up restrictions include w′y=a or 1 etc. as its special cases, where w is a column vector. This kind of plausible-sounding restrictions do exist in statistical practice, which were noticed and approached in certain fields of applied statistics and attracted sort of consideration. For example, economists explored some situations of the specific kind where certain adding-up restrictions appeared, in which they presented fitting descriptions of how to think the restrictions, and gave some of their solutions to a number of corresponding estimation and inference problems. We refer the reader to [3,4,10,13] for more information on the appearance of such kind of adding-up restrictions. However, the general situation depicted in (1.2) has not been properly approached in the statistical literature, yet the very process of mathematical and statistical approaches of the adding-up restrictions remains hidden.
For the purpose of making inference in the contexts of (1.1) and (1.2), we merge the two model equations in the following form
N: y=Xββ+εε, Ay=b, E(εε)=0, Cov(εε)=σ2ΣΣ. | (1.3) |
In this case, we can substitute (1.1) into (1.2) to lead to AXββ+Aεε=b, and rewrite (1.3) in the following equivalent form
N: y=Xββ+εε, b=AXββ+Aεε, E(εε)=0, Cov(εε)=σ2ΣΣ. | (1.4) |
Given the model equations as such in (1.3) or (1.4), we are confronted with the task of how to properly merge the adding-up restrictions in the estimation and inference procedure of the unknown parameter vector ββ. There have been several attempts in the past to solve the problem of choosing possible merge procedures. Unfortunately, when faced to such a question, professional statisticians have no definitive answers or rather they have different methods which have been suggested and implemented in the literature, none of them being completely convincing in the sense that it has been shown to be better than the others. The purpose of this article is to focus attention on dealing with the adding-up restrictions in (1.3) with some new thoughts and methodologies. The author will offer a feasible algebraic method to reconcile the second adding-up restrictions with the first regression equation in (1.3), and then use the method to solve some basic estimation and inference problems associated with N.
The rest of this paper is organized as follows. In Section 2, we introduce some basic formulas, facts, and results in matrix theory, as well as two groups of existing results related to the ordinary least-squares estimators (OLSE) and the best linear unbiased estimators (BLUEs) of unknown parametric vectors under (1.1). In Section 3, we show how to transform N in (1.3) into two kinds of linear models with implicit and explicit restrictions to the unknown parameter vector ββ respectively via certain suitable equivalent explanation of the adding-up restrictions. In Sections 4 and 5, we presents the description of the estimability of unknown parametric vector Kββ under the transformed models, and give the definitions and the derivations of analytical expressions of the OLSEs and BLUEs of Kββ through the transformed models. Section 6 gives some concluding remarks and a group of research problems concerning general linear models with adding-up restrictions.
In this section, we introduce some fundamental formulas and facts about matrix operations that have related applications to statistics, especially linear statistical models. It was properly known that the theory of generalized inverses of matrices is a major and dependable source of methods and techniques that was brought into the theory of linear statistical models for regression in 1950s, and thereby it played a key role for carrying out statistical estimation and inference in a wide variety of situations; see e.g., [2,9,14]. In this section, we shall present a group of well-known formulas, facts, and results in linear algebra and matrix theory, which we shall use as resource to simplify various matrix expressions that involve generalized inverses of matrices.
Lemma 2.1 ([7]). Let A∈Rm×n, B∈Rm×k, and C∈Rl×n. Then,
(a) r[A,B]=r(A)+r(EAB)=r(B)+r(EBA).
(b) r[AC]=r(A)+r(CFA)=r(C)+r(AFC).
In particular,
(c) r[A,B]=r(A)⇔EAB=0⇔R(B)⊆R(A).
(d) r[AC]=r(A)⇔CFA=0⇔R(C′)⊆R(A′).
Lemma 2.2 ([8]). Let A∈Rm×n and B∈Rm×k, Then, the linear matrix equation AX=B is solvable for X if and only if r[A,B]=r(A), or equivalently, AA+B=B. In this case, the general solution of the equation can be written in the parametric form X=A+B+(In−A+A)U, where U∈Rn×k is an arbitrary matrix.
There were different established inference theories and methods that we can adopt to estimate the unknown parameter vector ββ in (1.1), the two best-known tools were OLSEs and BLUEs. We turn now to reviewing some basic definitions and existing facts in linear model theory regarding the estimability, as well as the OLSEs and BLUEs of a given unknown parametric vector under (1.1), see e.g., [1,9,11,12,18].
Definition 2.3. Let M be as given in (1.1) and let K∈Rk×p be given. The vector Kββ of parametric functions is said to be estimable under M if there exists an L∈Rk×n such that E(Ly−Kββ)=0 holds for all ββ in M.
Definition 2.4. Let M be as given in (1.1), and let K∈Rk×p be given.
(a) The OLSE of the parametric vector ββ under (1.1), denoted by OLSEM(ββ), is defined to be
^ββ=argminββ(y−Xββ)′(y−Xββ). | (2.1) |
The OLSE of Kββ under (1.1) is defined to be OLSEM(Kββ)=KOLSEM(ββ).
(b) The BLUE of the vector of parametric functions Kββ under (1.1), denoted by BLUEM(Kββ), is defined to be linear statistic Ly, where L is a matrix such that Cov(Ly−Kββ)=min in the Löwner partial ordering subject to E(Ly−Kββ)=0.
As we know that the concepts of OLSE and BLUE have a long history and deep roots in parametric regression analysis, both of which have many nice and optimal algebraic and statistical properties, and therefore are the most welcome linear statistical inference techniques in parametric regression theory and the related applications. The conventionality of OLSEs/BLUEs under linear regression models really attracted statisticians' attention in the development of regression theory, and numerous formulas and facts regarding the OLSEs/BLUEs of ββ and Kββ under (1.1) were established via various precise and analytical algebraic operations of the given vectors and matrices and their generalized inverses. Specifically, the results in the following two lemmas were highly recognized in the domain of linear statistical models.
Lemma 2.5. Let M be as given in (1.1), and let K∈Rk×p be given. Then, the general expression of OLSEs of ββ in M can be written as
OLSEM(ββ)=X+y+FXv, | (2.2) |
where v∈Rp×1 is arbitrary; and the OLSE of Kββ under M can be written as
OLSEM(Kββ)=KX+y+KFXv. | (2.3) |
Lemma 2.6. Let M be as given in (1.1), K∈Rk×p, and suppose Kββ is estimable under M, namely, R(K′)⊆R(X′). Then, the BLUE of Kββ under M can be written as
BLUEM(Kββ)=PK;X;ΣΣy, | (2.4) |
where PK;X;ΣΣ is the solution of the matrix equation
G[X,ΣΣX⊥]=[K,0]. | (2.5) |
This equation is always solvable for G, that is, R([K,0]′)⊆R([X,ΣΣX⊥]′). In this case, the general solution of (2.5) can be expressed as
PK;X;ΣΣ=[K,0][X,ΣΣX⊥]++U[X,ΣΣX⊥]⊥, | (2.6) |
where U∈Rk×n is arbitrary. Moreover, the following results hold.
(a) r[X,ΣΣX⊥]=r[X,ΣΣ] and R[X,ΣΣX⊥]=R[X,ΣΣ].
(b) The product PK;X;ΣΣΣΣ can be uniquely written as PK;X;ΣΣΣΣ=[K,0][X,ΣΣX⊥]+ΣΣ.
(c) The expectation and covariance matrix of BLUEN(Kββ) are given by
E(BLUEM(Kββ))=Kββ and Cov(BLUEM(Kββ))=[K,0][X,ΣΣX⊥]+ΣΣ([K,0][X,ΣΣX⊥]+)′. |
(d) The matrix PK;X;ΣΣ is unique if and only if r[X,ΣΣ]=n.
(e) BLUEM(Kββ) is unique if and only if if and only if y∈R[X,ΣΣ] holds with probability 1.
Apparently, the representations and derivations of the OLSEs and BLUEs in the above two existing lemmas do not require to specify the mathematical forms of the distributions of the error term vector εε, and thereby we are able to actively utilize the clear and exact expressions of the OLSEs and BLUEs to solve many algebraic and computational problems in the context of (1.1).
Because the adding-up equation is directly put on the random vector y in (1.1), where no information of ββ explicitly appears in the equation, we should be careful to include the equation in the estimation and inference procedure of (1.1). In other words, we have to seek some alternative methods to approach estimation and inference problems of unknown parameters in the model. To this purpose, we show in this section how to make use of the complete information associated with the adding-up restrictions and to convert (1.3) into certain ordinary linear models with implicit and explicit restrictions to the unknown parametric vector ββ, respectively.
Given that y in (1.1) is a random vector, we then take the expectation and covariance matrix of both sides of the equation Ay−b=0 with respect to y to obtain
E(Ay−b)=AXββ−b=0 and Cov(Ay−b)=σ2AΣΣA′=0. | (3.1) |
Since the matrix ΣΣ in (1.3) is positive semi-definite, it is easy to verify that the matrix equality AΣΣA′=0 is equivalent to ΣΣ=FAΣΣFA. The adding-up equation in (1.2) thereby suggests the following facts
AXββ=b and ΣΣ=FAΣΣFA. | (3.2) |
This treatment is by no means profound and difficult to understand under the assumption in (1.3), and thus we can view them as the best mathematical interpretation that can be given about the adding-up restrictions. Recognizing the key role of (3.2) in the interpretation of (1.2), we are able subsequently to deal with the adding-up equation in carrying out inference under (1.3). There are basically two algebraic methods to merge the adding-up equation into (1.1) via (3.2). Below, we perspicuously illustrate the algebraic processes.
(I) Firstly, substituting the first equation into the second equation in (1.3) and noting (3.2), we can equivalently rewrite (1.3) in the following implicitly restricted linear model
Na: [yb]=[XAX]ββ+[εεAεε], E[εεAεε]=0, Cov[εεAεε]=σ2[FAΣΣFA000]. | (3.3) |
(II) Also replacing Ay=b and ΣΣ in (1.3) with (3.2) produces with probability 1 the following explicitly restricted linear model
Nb: y=Xββ+εε, AXββ=b, E(εε)=0, Cov(εε)=σ2FAΣΣFA. | (3.4) |
Apparently, the two alternative forms in (3.3) and (3.4) are in compliance with N in (1.3), and thereby they can conveniently help solve various estimation and inference problems under N. It is easily seen that (3.3) and (3.4) are nothing but two ordinary linear statistical models with implicit and explicit restrictions to unknown parameters in the models. This fact enables us to adopt different approach processes to carry out common estimation and inference under (1.3) via (3.3) and (3.4), and thereby this alternative way in fact makes clear insights actionable into the connotation of the model in (1.3).
As a classic subject of study in regression analysis, there has been some general discussion regarding estimation and inference problems of a linear statistical model with implicit and explicit restrictions to unknown parameters in the model; see, e.g., [19] and references therein. In light of the existing theory pertaining to this topic, we are now able to make statistical inference of (1.3) via the two alternative forms in (3.3) and (3.4) through the well-organized employment of ordinary theory and methodology of dealing linear regression models under various assumptions.
For convenience of representation, we adopt the notation
ˆy=[yb], ˆX=[XAX], ^ΣΣ=[FAΣΣFA000] |
in the sequel. We first describe the consistency problem in the context of (3.3). Note that the matrix equality
[ˆX,^ΣΣ][ˆX,^ΣΣ]+[ˆX,^ΣΣ]=[ˆX,^ΣΣ] |
holds from the definition of the Moore–Penrose generalized inverse. Therefore, it turns out under the assumptions in (3.2) that
E([ˆX,^ΣΣ][ˆX,^ΣΣ]+ˆy−ˆy)=[ˆX,^ΣΣ][ˆX,^ΣΣ]+ˆXββ−ˆXββ=0,Cov([ˆX,^ΣΣ][ˆX,^ΣΣ]+ˆy−ˆy)=σ2([ˆX,^ΣΣ][ˆX,^ΣΣ]+−I)^ΣΣ([ˆX,^ΣΣ][ˆX,^ΣΣ]+−I)′=0. |
These two equalities imply [ˆX,^ΣΣ][ˆX,^ΣΣ]+ˆy=ˆy holds with probability 1, or equivalently,
ˆy∈R[ˆX,^ΣΣ] | (4.1) |
holds with probability 1. In view of this fact, we adopt the following definition.
Definition 4.1. Na in (3.3) is said to be consistent if (4.1) holds with probability 1.
The OLSEs and BLUEs of unknown parameters in a given linear statistical model were recognized as two principal estimations in the domain of linear regression models, which were deeply approached and utilized in the development of statistical science. In the following, we introduce the definitions of the OLSEs and BLUEs of vectors of parametric functions, and then presents the exact and analytical formulas for calculating the OLSEs and BLUEs under (3.3).
Definition 4.2. Let Na be as given in (3.3), and let K∈Rk×p be given. The vector Kββ of parametric functions is said to be estimable under Na if there exists an L∈Rk×(n+m) such that E(Lˆy−Kββ)=0 holds for all ββ under Na.
Definition 4.3. Let Na be as given in (3.3), and let K∈Rk×p be given.
(a) The OLSE of the parametric vector ββ under (3.3), denoted by OLSENa(ββ), is defined to be
^ββ=argminββ(ˆy−ˆXββ)′(ˆy−ˆXββ). |
The OLSE of Kββ under (3.3) is defined to be OLSENa(Kββ)=KOLSENa(ββ).
(b) The BLUE of the vector of parametric functions Kββ under N, denoted by BLUENa(Kββ), is defined to be a linear statistic Lˆy, where L is a matrix such that Cov(Lˆy−Kββ)=min in the Löwner partial ordering subject to E(Lˆy−Kββ)=0.
Now applying the above definitions to Na in (3.3), we obtain the following results.
Theorem 4.4. Let Na be as given in (3.3), and let K∈Rk×p be given. Then, Kββ is estimable under Na ⇔ R(K′)⊆R(X′). In particular, Xββ is always estimable under Na.
Proof. It follows from E(Lˆy−Kββ)=0 ⇔ LˆXββ−Kββ=0 for all ββ ⇔ LˆX=K ⇔ R(K′)⊆R(ˆX′) ⇔ R(K′)⊆R(X′) by Lemma 2.2.
Referring to Lemmas 2.5 and 2.6, we obtain the following two results about the OLSEs and BLUEs under (3.3).
Theorem 4.5. Let Na be as given in (3.3) and suppose Kββ is estimable under Na. Then, the OLSE of ββ under Na can be written as OLSENa(ββ)=ˆX+ˆy+FXv, where v∈Rp×1 is arbitrary; and the OLSE of Kββ under Na can be uniquely written as
OLSENa(Kββ)=KˆX+ˆy, E(OLSENa(Kββ))=Kββ, Cov(OLSENa(Kββ))=σ2KˆX+^ΣΣ(KˆX+)′. |
Theorem 4.6. Let Na be as given in (3.3) and suppose Kββ is estimable under Na. Then, the BLUE of Kββ under Na can be written as
BLUENa(Kββ)=PK;ˆX;^ΣΣˆy, |
where PK;ˆX;^ΣΣ is the solution of the matrix equation G[ˆX,^ΣΣˆX⊥]=[K,0]. This equation is always solvable for G, that is, R([K,0]′)⊆R([ˆX,^ΣΣˆX⊥]′). In this case, the general solution of the matrix equation can be expressed as
G=PK;ˆX;^ΣΣ=[K,0][ˆX,^ΣΣˆX⊥]++UE[ˆX,^ΣΣˆX⊥], |
where U∈Rk×(n+m) is arbitrary. Moreover, the following results hold.
(a) r[ˆX,^ΣΣˆX⊥]=r[ˆX,^ΣΣ] and R[ˆX,^ΣΣˆX⊥]=R[ˆX,^ΣΣ].
(b) The product PK;ˆX;^ΣΣ^ΣΣ can be uniquely written as PK;ˆX;^ΣΣ^ΣΣ=[K,0][ˆX,^ΣΣˆX⊥]+^ΣΣ.
(c) The expectation and covariance matrix of BLUENa(Kββ) are given by
E(BLUENa(Kββ))=Kββ and Cov(BLUENa(Kββ))=[K,0][ˆX,^ΣΣˆX⊥]+^ΣΣ([K,0][ˆX,^ΣΣˆX⊥]+)′. |
(d) The matrix PK;ˆX;^ΣΣ is unique if and only if r[ˆX,^ΣΣ]=m+n.
(e) BLUENa(Kββ) is unique if and only if ˆy∈R[ˆX,^ΣΣ] holds with probability 1.
In what follows, we denote ˜A=AX and ~ΣΣ=FAΣΣFA. Recall a well-known fact that the matrix equation ˜Aββ=b is solvable for ββ if and only if b∈R(˜A). By Lemma 2.2, the general solution of ββ and the corresponding Kββ can be written in the following parametric forms
ββ=˜A+b+F˜Aγγ, | (5.1) |
Kββ=K˜A+b+KF˜Aγγ, | (5.2) |
where γγ∈Rp×1 is arbitrary. Substitution of (5.1) into (3.4) yields
˜Nb: z=XF˜Aγγ+εε, E(εε)=0, Cov(εε)=σ2~ΣΣ, | (5.3) |
where z=y−X˜A+b. This is a new linear model with the unknown parameter vector γγ. Hence, the estimablility, the OLSE, and the BLUE of the vector of parametric functions KF˜Aγγ can be obtained from various existing results as follows.
Definition 5.1. Let Nb be as given in (3.4), and let K∈Rk×p be given. The vector Kββ of parametric functions is said to be estimable under Nb if there exist L∈Rk×n and c∈Rk×1 such that E(Ly+c−Kββ)=0 holds under Nb.
Lemma 5.2. Let Nb be as given in (3.4), and let K∈Rk×p be given. Then, Kββ is estimable under Nb if and only if R(K′)⊆R(X′).
Theorem 5.3. Let Nb be as given in (3.4), let K∈Rk×p be given, and suppose Kββ is estimable under (3.4). Then, the OLSE of ββ under Nb can be written as
OLSENb(ββ)=(˜A+−F˜A(XF˜A)+X˜A+)b+F˜A(XF˜A)+y+F˜AFXF˜Au, | (5.4) |
where u∈Rp×1 is arbitrary. The OLSE of Kββ under Nb can be uniquely written as
OLSENb(Kββ)=(K˜A+−KF˜A(XF˜A)+X˜A+)b+KF˜A(XF˜A)+y, | (5.5) |
E(OLSENb(Kββ))=Kββ, Cov(OLSENb(Kββ))=σ2KF˜A(XF˜A)+~ΣΣ(KF˜A(XF˜A)+)′. |
Proof. According to Lemma 2.5, the OLSE of γγ in (5.3) can be written as
^γγ=(XF˜A)+z+FXF˜Au, |
where u∈Rp×1 is arbitrary. Substitution of this formula into (5.1) gives the OLSE of ββ in (3.4):
OLSENb(ββ)=˜A+b+F˜A(XF˜A)+z+F˜AFXF˜Au=(˜A+−F˜A(XF˜A)+X˜A+)b+F˜A(XF˜A)+y+F˜AFXF˜Au, |
establishing (5.4) and (5.5).
Theorem 5.4. Let Nb be as given in (3.4) and suppose Kββ is estimable under (3.4). Then, the BLUE of Kββ under Nb can be written as
BLUENb(Kββ)=(K−PKF˜A;XF˜A;~ΣΣX)˜A+b+PKF˜A;XF˜A;~ΣΣy, | (5.6) |
where PKF˜A;XF˜A;~ΣΣ=[KF˜A,0][XF˜A,~ΣΣEXF˜A]++U1E[XF˜A,~ΣΣEXF˜A], and U1∈Rk×n is arbitrary. Moreover, the following results hold.
(a) r[XF˜A,~ΣΣEXF˜A]=r[XF˜A,~ΣΣ] and R[XF˜A,~ΣΣEXF˜A]=R[XF˜A,~ΣΣ].
(b) The product PKF˜A;XF˜A;~ΣΣ~ΣΣ can be uniquely written as PKF˜A;XF˜A;~ΣΣ~ΣΣ=[KF˜A,0][XF˜A,~ΣΣEXF˜A]+~ΣΣ.
(c) The expectation and covariance matrix of BLUENb(Kββ) are given by
E(BLUENb(Kββ))=Kββ and Cov(BLUENb(Kββ))=PKF˜A;XF˜A;~ΣΣ~ΣΣP′KF˜A;XF˜A;~ΣΣ. |
(d) The matrix PKF˜A;XF˜A;~ΣΣ is unique if and only if r[X~ΣΣ˜A0]=r(˜A)+n.
(e) BLUENb(Kββ) is unique if and only if [yb]∈R[X~ΣΣ˜A0] holds with probability 1.
Proof. According to Lemma 2.6, the BLUE of KF˜Aγγ under (5.3) is given by
BLUENb(KF˜Aγγ)=PKF˜A;XF˜A;~ΣΣz. |
Substitution of this BLUE into the equality in (5.2) gives the BLUE of Kββ under (3.4)
BLUENb(Kββ)=K˜A+b+BLUENb(KF˜Aγγ)=K˜A+b+PKF˜A;XF˜A;~ΣΣ(y−X˜A+b), |
as required for (5.6).
Result (a) follows from Lemma 2.6(a). Result (b) follows from Lemma 2.6(b). Result (c) follows from (5.6).
It can be seen from (5.6) that PKF˜A;XF˜A;~ΣΣ is unique if and only if E[XF˜A,~ΣΣEXF˜A]=0, i.e., r[XF˜A,~ΣΣEXF˜A]=n. Also see from (a) and Lemma 2.1(b) that r[XF˜A,~ΣΣEXF˜A]=r[XF˜A,~ΣΣ] =r[X~ΣΣ˜A0]−r(˜A), so that Result (d) follows.
It can be seen from (5.6) and that BLUENb(Kββ) is unique if and only if E[XF˜A,~ΣΣEXF˜A](y−X˜A+b)=0, i.e.,
r[y−X˜A+b,XF˜A,~ΣΣEXF˜A]=r[XF˜A,~ΣΣ] | (5.7) |
holds with probability 1 by Lemma 2.1(c). In this situation, it is necessary to simplify the rank equality by removing the generalized inverses on both sides of (5.7). In fact by Lemma 2.1(b) and elementary block matrix operations,
r[y−X˜A+b,XF˜A,~ΣΣEXF˜A]=r[y−X˜A+bX~ΣΣ0˜A0]−r(˜A)=r[yX~ΣΣb˜A0]−r(˜A),r[XF˜A,~ΣΣ]=r[XF˜A,~ΣΣ]=r[X~ΣΣ˜A0]−r(˜A). |
So that (5.7) is equivalent to r[yX~ΣΣb˜A0]=r[X~ΣΣ˜A0], i.e., [yb]∈R[X~ΣΣ˜A0] holds by Lemma 2.1(c), as required for Result (e).
The author proposed and investigated some estimation and inference problems regarding a linear statistical model with adding-up restrictions to the observable random variables in the model, and obtained a group of formulas and facts about estimations and inferences in the context of (1.3), including the computational processes to derive the OLSEs and BLUEs under the model assumptions via a series of careful algebraic operations of the given vectors and matrices. Based on the findings obtained in the preceding sections, there is no doubt to say that this approach clearly demonstrates a normal procedure of dealing with adding-up restrictions in the context of linear statistical models. Hopefully, this study will add to what is out there with regard to the subject, and also is compatible with the previous contributions.
Given the resolutions to the adding-up restrictions, we may say some additional remarks pertaining to further research problems under the model assumption. Recall that the OLSEs and BLUEs are defined by different optimality criteria in mathematics and statistics. Therefore, their expressions and properties are not necessarily the same, and thereby it is natural to seek possible connections between these estimation results. It is, in fact, a subject area in regression analysis is to characterize relationships between different OLSEs and BLUEs, which has deep roots with strong statistical explanation and usefulness in the domain of linear statistical models and applications; see, e.g., [6,15,16,20,21] and references therein for the background and study of this subject. Based on the exact and analytical expressions of OLSEs and BLUEs obtained, we can consider in depth various additional topics in the statistical inference of general linear statistical models with adding-up restrictions. Particularly, it is natural to speculate on the relationship between the OLSEs and BLUEs under the two models in (1.1) and (1.3), and therefore we put forward the following five clear and reasonable equalities between the OLSEs and BLUEs under the two models in (1.1) and (1.3):
(a) OLSEM(Kββ)=OLSEN(Kββ),
(b) OLSEM(Kββ)=BLUEN(Kββ),
(c) BLUEM(Kββ)=OLSEN(Kββ),
(d) BLUEM(Kββ)=BLUEN(Kββ),
(e) OLSEN(Kββ)=BLUEN(Kββ).
The five equalities describe the direct relevances of the four estimators under the two models, and thus it would be of interest to make a deep-going study of the equalities from theoretical and practical points of view. As a matter of fact, these kinds of equalities were properly considered with clear objective intention in the statistical inference of linear statistical models in the past several decades. Unquestionably, the equalities in (a)–(e) not only can be classified as certain statistical inference problems, but also can be alternatively converted to certain matrix equality problems in term of the clear and analytical expressions of the OLSEs and BLUEs obtained in the preceding sections. In order to resolve these proposed equivalence problems, we need to prepare a complex of preliminary methods and techniques in matrix algebra, including many formulas and facts about generalized inverses of matrices and the matrix rank methodology. Nevertheless, this sort of future investigations together with the current theoretical and methodological advances in this research area will definitely provide considerable insight into the intrinsic natures hidden behind the adding-up restrictions, so that we believe that the algebraic treatments presented in this article will sufficiently prompt other similar studies regarding different kinds of regression models with adding-up restrictions to observable random variables under various assumptions. The five equalities describe the direct relevances of the four estimators under the two models, and thus it would be of interest to make a deep-going study of the equalities from theoretical and practical points of view. As a matter of fact, these kinds of equalities were properly considered with clear objective intention in the statistical inference of linear statistical models in the past several decades. Unquestionably, the equalities in (a)–(e) not only can be classified as certain statistical inference problems, but also can be alternatively converted to certain matrix equality problems in term of the clear and analytical expressions of the OLSEs and BLUEs obtained in the preceding sections. In order to resolve these proposed equivalence problems, we need to prepare a complex of preliminary methods and techniques in matrix algebra, including many formulas and facts about generalized inverses of matrices and the matrix rank methodology. Nevertheless, this sort of future investigations together with the current theoretical and methodological advances in this research area will definitely provide considerable insight into the intrinsic natures hidden behind the adding-up restrictions, so that we believe that the algebraic treatments presented in this article will sufficiently prompt other similar studies regarding different kinds of regression models with adding-up restrictions to observable random variables under various assumptions.
The author would like to express his sincere thanks to the handling editor and anonymous reviewers for their helpful comments and suggestions.
The author declares that there is no competing interest.
[1] | I. S. Alalouf, G. P. H. Styan, Characterizations of estimability in the general linear model, Ann. Statist., 7 (1979), 194–200. |
[2] |
H. H. Bingham, W. J. Krzanowski, Linear algebra and multivariate analysis in statistics: development and interconnections in the twentieth century, British J. Hist. Math., 37 (2022), 43–63. https://doi.org/10.1080/26375451.2022.2045811 doi: 10.1080/26375451.2022.2045811
![]() |
[3] | H. Haupt, W. Oberhofer, Fully restricted linear regression: A pedagogical note, Econ. Bull., 3 (2002), 1–7. |
[4] |
H. Haupt, W. Oberhofer, Generalized adding-up in systems of regression equations, Econ. Lett., 92 (2006), 263–269. https://doi.org/10.1016/j.econlet.2006.03.001 doi: 10.1016/j.econlet.2006.03.001
![]() |
[5] |
A. Markiewicz, S. Puntanen, All about the ⊥ with its applications in the linear statistical models, Open Math., 13 (2015), 33–50. https://doi.org/10.1515/math-2015-0005 doi: 10.1515/math-2015-0005
![]() |
[6] | A. Markiewicz, S. Puntanen, G. P. H. Styan, The legend of the equality of OLSE and BLUE: highlighted by C. R. Rao in 1967, In: A volume in Honor of C. R. Rao on the occasion of his 100th birthday, Methodol. Appl. Statist., 2021, 51–76. |
[7] | G. Marsaglia, G. P. H. Styan, Equalities and inequalities for ranks of matrices, Linear Multilinear Algebra, 2 (1974), 269–292. |
[8] |
R. Penrose, A generalized inverse for matrices, Proc. Cambridge Philos. Soc., 51 (1955), 406–413. https://doi.org/10.1017/S0305004100030401 doi: 10.1017/S0305004100030401
![]() |
[9] | S. Puntanen, G. P. H. Styan, J. Isotalo, Matrix tricks for linear statistical models: our personal top twenty, Berlin: Springer, 2011. |
[10] | B. Ravikumar, S. Ray, N. E. Savin, Robust Wald tests in SUR systems with adding-up restrictions, Econometrica, 68 (2000), 715–719. |
[11] | C. R. Rao, Unified theory of linear estimation, Sankhyā Indian J. Statist. A, 33 (1971), 371–394. |
[12] |
C. R. Rao, Representations of best linear unbiased estimators in the Gauss-Markoff model with a singular dispersion matrix, J. Multivariate Anal., 3 (1973), 276–292. https://doi.org/10.1016/0047-259X(73)90042-0 doi: 10.1016/0047-259X(73)90042-0
![]() |
[13] |
M. Satchi, A note on adding-up restrictions when modelling trade flows, Econ. Model., 21 (2004), 999–1002. https://doi.org/10.1016/j.econmod.2003.12.002 doi: 10.1016/j.econmod.2003.12.002
![]() |
[14] | S. R. Searle, Matrix algebra useful for statistics, New York: Wiley, 1982. |
[15] |
Y. Tian, Some decompositions of OLSEs and BLUEs under a partitioned linear model, Int. Stat. Rev., 75 (2007), 224–248. https://doi.org/10.1111/j.1751-5823.2007.00018.x doi: 10.1111/j.1751-5823.2007.00018.x
![]() |
[16] | Y. Tian, On equalities of estimations of parametric functions under a general linear model and its restricted models, Metrika, 72 (2010), 313–330. |
[17] |
Y. Tian, Solving optimization problems on ranks and inertias of some constrained nonlinear matrix functions via an algebraic linearization method, Nonlinear Anal., 75 (2012), 717–734. https://doi.org/10.1016/j.na.2011.09.003 doi: 10.1016/j.na.2011.09.003
![]() |
[18] |
Y. Tian, On properties of BLUEs under general linear regression models, J. Statist. Plann. Inference, 143 (2013), 771–782. https://doi.org/10.1016/j.jspi.2012.10.005 doi: 10.1016/j.jspi.2012.10.005
![]() |
[19] |
Y. Tian, M. Beisiegel, E. Dagenais, C. Haines, On the natural restrictions in the singular Gauss-Markov model, Stat. Papers, 49 (2007), 553–564. https://doi.org/10.1007/s00362-006-0032-5 doi: 10.1007/s00362-006-0032-5
![]() |
[20] |
Y. Tian, W. Guo, On comparison of dispersion matrices of estimators under a constrained linear model, Stat. Methods Appl., 25 (2016), 623–649. https://doi.org/10.1007/s10260-016-0350-2 doi: 10.1007/s10260-016-0350-2
![]() |
[21] |
Y. Tian, J. Zhang, Some equalities for estimations of partial coefficients under a general linear regression model, Stat. Papers, 52 (2011), 911–920. https://doi.org/10.1007/s00362-009-0298-5 doi: 10.1007/s00362-009-0298-5
![]() |