
ISSN: 2644-1381
Jaesung Choi*
Received: February 27, 2020; Published: March 11, 2020
*Corresponding author: Department of Statistics, Keimyung University, Korea
DOI: 10.32474/CTBB.2020.02.000140
This paper discusses projection methods to find nonnegative estimates for variance components of random effects in mixed models. The proposed methods are based on the concepts of projections, which are called projection method I, II and III. These three methods produce the same nonnegative estimates for the same data. Even though each method uses orthogonal projections in its own way, the results are the same for the variance components regardless of which method is used. It is shown that quadratic forms of an observation vector are constructed differently in each method. All sums of squares in quadratic forms of the observation vector can be expressed as squared distances of corresponding projections. A projection model is defined and used to evaluate expected values of quadratic forms of observations that are associated with variance components. Hartley’s synthesis is used as a method for finding the coefficients of variance components.
Keywords:Mixed model; projections; quadratic forms; random effects; synthesis
Much literature has been devoted to the estimation of variance
components in random effects or mixed effects models. A variance
component should always be nonnegative by its definition;
however, we sometimes get it as negative [1,2]. illustrated this
with the simple hypothetical data of a one-way classification
having three observations in two classes and insisted that there
was nothing intrinsic in the analysis of variance method to prevent
it. When a negative estimate happens, it is not easy to handle this
situation in interpretation and action. Hence, many papers have
been contributed to strategies to deal with the negative values
as estimates of variance components [3]. suggests that negative
estimates of variance components can occur in certain designs such
as split plot and randomized block designs by random- inaction.
Thompson discusses the interpretation of the negative estimate
and suggests an alternative method when the analysis of variance
method yields negative estimates [4]. also suggest a procedure
for eliminating negative estimates of variance components in
random effects models. The analysis of the variance method is
almost exclusively applied to balanced data for estimating variance
components. However, there are multiple methods for unbalanced
data. Therefore, it is necessary to identify the types of data before
choosing a method. Though balanced data have the same numbers
of observations in each cell, unbalanced data have unequal
numbers of observations in the subclasses made by the levels
of classification factors. Depending on the types of data, many
methods can be applied to the estimation of variance components
in a vector space. Representing data as vectors, the vector space of
an observation vector can be partitioned in many ways, depending
on the data structure. For balanced data, the vector space can
always be partitioned into orthogonal vector subspaces according
to the sources of variation, but it is not true for unbalanced data.
This is the main difference between balanced and unbalanced
data from the view point of a vector space. A random effect is a
random variable representing the effect of a randomly chosen
level from a population of levels that a random factor can assume,
while a fixed effect is an unknown constant denoting the effect of a
predetermined level of a factor. A linear model with these two types
of effects is called a mixed effects model. The primary concern with
the model in this paper is naturally in the nonnegative estimation
of variance components of random effects. A negative estimate can
happen in any method that contributes to the estimation.
Hence, many papers have investigated strategies for
interpretation and alternatives. Such strategies are seen in [5-9].
However, it is necessary to have a method that yields nonnegative estimates despite all such efforts [10].suggested a method that uses
reductions in sums of squares due to fitting both the full model and
different sub-models of it for estimating variance components of
random effects in mixed models. This method is called the fitting
constants method or Henderson’s Method 3. Even though it has
been used extensively for the estimation of variance components in
mixed models, it still has some defects producing negative estimates
in case [11]. synthesis is also used for calculating the coefficients of
variance components in the method. Although this method is very
useful, we should recognize whether quadratic forms for variance
components are in the right form or not. Otherwise, expectations
of the quadratic forms can be different from the real ones. This is
going to be discussed in detail in projection model building. This
paper suggests three methods to produce nonnegative estimates
for variance components in mixed models. They are based on the
concept of projection defined on a vector space. The definition
of a projection and its related concepts are discussed in [12,13].
Quadratic forms in the observations can be obtained as squared
distances of projections defined in proper vector subspaces. Each
method requires that all vector subspaces for projections should
be orthogonal to each other at the stage of fitting sub-models
serially. When the orthogonality is satisfied with vector subspaces,
it is possible to get nonnegative estimates. Hence, we also discuss
how to construct orthogonal vector subspaces from a given mixed
model. Quadratic forms as sums of squares due to random effects
are then used to evaluate their expected values. Hereafter, equating
quadratic forms to their expected values represents available
equations for the estimates. For calculating the coefficients of
variance components, Hartley’s synthesis is applied but in a
different manner, which will be discussed.
Mixed models are used to describe data from experimental situations where some factors are fixed, and others are random. When two types of factors are considered in experiments, one is interested in both parts, that is, the fixed-effects part and the random-effects part, in models. Let α be a vector of all the fixed effects except μ in a mixed model and let δi denote a set of random effects for random factor i for i = 1, 2, r. Then, i δ could be interaction effects or nested-factor effects when they are simply regarded as effects from random factors. The matrix notation of the mixed model for an observation vector y is
where jμ + XFαF is the fixed part of the model and XRδR + ∈ is the random part of the model. δi s are assumed to be independent and identically distributed as N(0, δσi2 I ), and ε is assumed to be distributed as N(0,σ∈2I ) . The mean and variance of y from (1) is
The expectation of any quadratic form in the observations of a vector y is represented as a function of variance components and fixed effects. The variance components of the full model can be estimated by the fitting constants method of using reductions in the sums of squares due to fitting the full model and the submodel of it. This method provides unbiased estimators of the variance components that do not depend on any fixed effects in the model, and it has been widely used for the estimation of variance components for unbalanced data. However, it still has an unsolved problem having negative solutions as estimates. As an alternative, a method which is based on the concepts of projections is suggested [14]. To discuss it, we consider the model (1) as representative. Since there are two parts in the model, we naturally divide the model into a fixed part and a random part. The random part of the model consists of random effects and errors:
where ∈R = Σi=1r Xiδi+ ∈ The general mean μ and fixed effects αF of (5) can be estimated from normal equations. Regarding y as an observation vector in the n-dimensional vector space, it can be decomposed into two component vectors orthogonal to each other. The decomposition of y is done by projecting y onto the vector subspace generated by ( j, XF ) .
Since a method based on the concept of projection is discussed, it will be called the projection method. For a mixed model such as (5), we can decompose y into two components by means of projections. Denoting (j, XF) and (μ, αF)T by Xym and μm , respectively, the projection of y onto the vector subspace spanned by Xm is Xm X my − , where Xm − denotes a Moore-Penrose generalized inverse of XM . Then, y can be decomposed into two vectors, that is, Xm X −my and (I -XMX−M )y − which are orthogonal [15,16]. Instead of the fitting constants method, the projection method is attempted to estimate the nonnegative estimates of the variance components in a mixed model. To explain the method simply, suppose there are two factors A and B for a two-way crossclassified unbalanced data where A is fixed with a levels and B is random with b levels. The model for this is
where y is an observation vector in the n dimensional vector space, αF is a vector of fixed effects of A, δβ and δαβ represent vectors of random effects of B and AB interaction respectively, and XM = ( j, XF ), αM(μ, αF)T and ∈M = Xβδβ + Xαβ δαβ +∈. The second εexpression of (6) represents the fixed-effects part and the random part. The random part SM is obtained by the projection of y onto a vector subspace generated by the XM, which is (I -XM XM−) y. So, y is represented as
where yM= XM XM−y satisfies the two conditions for being the projection of y onto a vector subspace spanned by the columns of XM. The projection should be obtained by the orthogonal projection to the subspace and denoted as a linear combination of the column vectors of XM.XM XM-y conditions. Since yM is orthogonal to eM, the random part eM=(I− XM XM-)y is not affected by the fixed effects and has all the information about the variance components and random error variance. Since there are two random effects and random error terms in the model of (6), we can use eM for finding the related variance components. The model for the estimation of σ2β is
where yB XBXB−eM = is the projection of eM onto the column space of XB ⋅ yB and eB are orthogonal each other. Hence, eB is not affected by the random effects B δ Therefore, eB is used for finding the subspace that has information about σ2αβ. The model for this is
where XAB =(I − XM XM X−BX−B) Xαβ and ∈αβ=(I− XM X−M XB X−B) Hence, the projection of eB onto the subspace generated by XAB is yAB= XAB X−ABeB. Then,
where eAB is (I − XAB X−AB )eB. Finally, we can use eAB for finding the coefficient matrix of the random error vector which generates the error space orthogonal to all the other spaces.
Thus, we can know that eAB has all the information about σ2∈ of the random error vector ∈ . Denoting y as the sum of orthogonal projections and error part,
Each term of (13) can be used to calculate the sums of squares that are quadratic forms in the observations. Since y is partitioned as four terms, there are four available sums of squares. We denote them SSM, SSB, SSAB and SSE where subscripts are corresponding factors. They are defined as
where each SS term is given as the squared length of the projection of y onto its own vector subspace, and XE= (I−XMX−M XBX−B XAB X−AB). All the sums of squares are evaluated by using the eigenvalues and eigenvectors of the projection matrices associated with the quadratic forms in y. Since projections are defined on subspaces that are orthogonal to each other, we can identify the coefficient matrices spanning them.
Since y is made up by the sum of mutual orthogonal projections such as (13), y can be represented by the orthogonal coefficient’s matrices of the effects of the assumed model (6). Temporarily, we denote y as yp for differentiating the model based on projections to the classical model (6). Then, the model for yp is
y = XMαM + XBαβ + XABδαβ + XE ∈ (15)
where yp = y . Since each coefficient matrix of the effects is derived from the corresponding orthogonal projection, the equation of (15) defines a projection model that is different from a classical two-way linear mixed model (6). It is useful for evaluating the coefficients of the variance components in the expectations of the quadratic form of an observation vector yp . In the model, all the coefficient matrices are orthogonal to
each other. δβ, δαβ and s are assumed to be N(0,σ2βIb), N(0,σ2αβIab) and N(0,σ2∈In) respectively. The expectation and the covariance matrix of y p of the projection model (15) is
Equating the three sums of squares, SSB, SSAB, and SSE of (14) to their correspond- ing expectations leads to linear equations in the variance components, the solutions to which are taken as the estimators of those components. Now, the equations are
Solutions from the linear equations (18) are nonnegative estimates of the variance components. Since there are three different ways of getting sums of squares by means of projections, we will differentiate them with projection method I, II, and III. The procedure using the system of linear equations like (18) is called projection method I. The projection method II uses residual vectors after projecting y onto orthogonal subspaces. That is, eM, eB, and eAB are used such as. Then
Since eM has three random components, eMT eM in the quadratic form of yp in which the coefficients matrices of the projection model are orthogonal is available for esti- mating their variance components. Denoting eMT eM as RSSM,
RSSM = eMT eM, (20)
where M RSS measures the variation due to the three random effects, and thus, the quantity is used for the estimation of three variance components σ2β , σ2αβ, and σ2∈. Representing the residual random vector eB as yp , has two random components as follows.
Hence, eTB is used as an variation quantity for two random effects vectors. Denoting eTB as RSSB,
where RSSB is used for estimating the two variance components σ2αβ and σ2∈ since eB has just two random effects. Finally, expressing eAB as yp
eAB = (I − XAB XAB−) eB = XE ∈ (23)
which has just one random component s. Therefore, eABT eAB shows the variation due to the random error vector only, and this quantity is used for estimating the variance component σ2 ∈ . Denoting eABT eAB as RSSAB,
RSSAB = eABT eAB (24)
Hence,RSSM , RSSB , and RSSAB are another set of sums of squares for estimating variance components instead of using sums of squares derived from the projections as an alternative method.RSSM , RSSB , and RSSAB are also evaluated by using the eigenvalues and eigenvectors of the projection matrices associated with the quadratic forms in y. Now, the expected values of the RSS’s are
Then, the linear equations of variance components are obtained
by equating the RSS’s to their expected values, the solutions for
which always produce nonlinear estimates.
That is,
Even though two systems of linear equations are not the same, either system will produce the same estimates of the variance components that are nonnegative. As another method, projection method III is also available for the estimation of variance components. This method is done as follows. For the model of (6), y = Xθ +∈, where X = (j, XF, Xβ, Xαβ) and θ =( μ ,αF, δβ, δαβ )T. This method splits the vector space of an observation vector into two subspaces, one for the projection part and the other for the error part at each step. Then, the projection of y onto the subspace spanned by XX− is given by XX− y , and the error vector in the error vector space is (I − XX− ) y . Therefore, the coefficient matrix of s is derived as (I − XX− )from it. The quadratic form y '(I − XX− ) y denoted by 0 BSS is the sum of squares due to random error only, which has all the information about σ2∈ . For information about both σ2αβ and σ2∈, the vector space of the observation vector can be decomposed into two parts one for the projection part and the other for the error part. For this, the model to be fitted is y = X1 θ1 +∈1 , where X1 (j j, XF, Xβ ) θ1 = (μ, αF, δβ ) and ∈1= Xαβ ααβ +∈. Then, the projection of y onto the subspace spanned by X1 X1− is given by X1 X1− y, and the error vector in the errror vector space is (I-X1 X1−)y . The quadratic form yT(I-X1 X1−)y denoted by BSS1 has information about σ2αβ and σ2∈ . Now, the error vector is represented by
Hence, the coefficient matrix of δβ is given by (I-X2X2− )Xβ. It is necessary to evaluate the expected values of the quadratic forms for constructing the equations for the variance components. They are
The nonnegative estimates of variance components are given as solutions of linear equations of σ2α and σ2αβ σ2∈. The above equations are summarized as follows:
where cij ’s are coefficients of variance components of expected values of quadratic forms of (29).
As a first example of nonnegative estimates of random effects
for a two-way mixed model, Montgomery (2013)’s data are
illustrated. The data are from an experiment for a gauge capability
study where parts are randomly selected, and three operators
are fixed. An instrument or gauge is used to measure a critical
dimension on a part. Twenty parts have been selected from
the production process, and only three operators are assumed
to use the gauge. The assumed model for the data in Table 1 is yijk = μ + αi +γj + (αγ)ij + ∈ijk, where they αi (i = 1, 2, 3) are fixed effects such that Σi=13αi = 0 and γj (j = 1, 2,...., 20), (αγ)ij, and ∈ijk are uncorrelated random variables having zero means
and variances Var(γj)= σγ2, Var((αγ)ij)= σαγ2
Var(∈ijk)= σ2. Under the assumed unrestricted model, estimated variance
components are
σ2ˆγ = 10.2798 , σ2ˆαγ= −0.1399 , and σ2ˆ∈ = 0.9917 . Applying the projection method, I to the data, the linear equations
of variance components are given as follows:
The solutions of the equations are σ2ˆγ = 10.3985, σ2ˆαγ= 0.3559, σ2ˆ∈ = 0.9917. All the variance components
are estimated nonnegatively. When we apply projection method II
to the same data, we get where RSSfixed = 1271.975 , RSSpart = 86.55 , and
RSSpart×operator = 59.5 . The solutions for the equations are σˆ2γ =10.3985, σˆ2αγ
=0.3559, σˆ2∈ =0.9917 which are the
same as the previous solutions. Hence, either one of the projection
methods can be used for the nonnegative estimation of variance
components of random effects in a mixed model. Projection method
III also gives the same result as projection methods I and II for
the data. As a second example, Searle [2]’s hypothetical data are
illustrated. Searle explains why a negative estimate can occur in the
estimation of variance component of random effects in a random
model. The data are shown in Table 1. Since class in Table 2 is a
random factor, the one-way random effects model is assumed. The
assumed model is yij = μ+αi+ ∈ij, where the αi(i=1,2) are
random effects and ∈ij are uncorrelated random errors having
zero means and variances Var(αi)= σ2α and Var(∈ij)= σ2. As a result of the analysis of variance, the estimates of variance
components are given as σˆα2 = −15.33 and σˆ∈2 = 52 . Searledemonstrated how negative estimates could come from the analysis
of variance and insisted that there would be nothing intrinsic in the
method to prevent it. However, the projection methods yield the
same nonnegative estimates as σˆ2α = 2 and
σˆ2∈ = 52 in any method. Variance should be a nonnegative quantity as a measure
of variation in data by its definition. In this work, it shows that
orthogonal projections are very useful for defining a projection
model for nonnegative variance estimation. Although there have
been many attempts in literature to fix the problem of negative
estimates for variance components over decades, those were not
successful. However, the proposed methods in this paper always
produce nonnegative estimates of variance components of the
random effects in a mixed model. The two most important findings
are checked and discussed for the estimation of nonnegative
variance component. One is that a projection model should be
derived from an assumed mixed-effects model. The other is that
expectations of quadratic forms associated with the random
effects should be evaluated from the projection model. This
paper introduces terms such as projection method I, II, and III
related to the methods, and the projection model for emphasizing
projection rather than model fitting. Though they are based on
the same assumed model, three methods are ap- plied differently
in the application. Each method uses in its own way but summing
up all orthogonal projections come to the observation vector.
Depending on the types of projections, each method produces three
different sets of equations for the evaluation of quadratic forms.
Nonetheless, all of them show the same nonnegative estimates for
variance components. It also shows that projection methods can be
used for estimating variance components of the random effects in
either random model or mixed model through examples. It should
be noted that all the matrices associated with the quadratic forms
come from the projection model not from the assumed model. In
such a case, Hartley’s synthesis can yield correct coefficients of
variance components. This work was supported by Basic Science Research Program
through the National Research Foundation of Korea (NRF) funded by
the Ministry of Education under Grant No.2018R1D1A1B07043021.
Conclusion
Funding
References
Bio chemistry
University of Texas Medical Branch, USADepartment of Criminal Justice
Liberty University, USADepartment of Psychiatry
University of Kentucky, USADepartment of Medicine
Gally International Biomedical Research & Consulting LLC, USADepartment of Urbanisation and Agricultural
Montreal university, USAOral & Maxillofacial Pathology
New York University, USAGastroenterology and Hepatology
University of Alabama, UKDepartment of Medicine
Universities of Bradford, UKOncology
Circulogene Theranostics, EnglandRadiation Chemistry
National University of Mexico, USAAnalytical Chemistry
Wentworth Institute of Technology, USAMinimally Invasive Surgery
Mercer University school of Medicine, USAPediatric Dentistry
University of Athens , GreeceThe annual scholar awards from Lupine Publishers honor a selected number Read More...