email   Email Us: info@lupinepublishers.com phone   Call Us: +1 (914) 407-6109   57 West 57th Street, 3rd floor, New York - NY 10019, USA

Lupine Publishers Group

Lupine Publishers

  Submit Manuscript

ISSN: 2644-1381

Current Trends on Biostatistics & Biometrics

Research Article(ISSN: 2644-1381)

Non-Negative Variance Component Estimation for Mixed- Effects Models Volume 2 - Issue 3

Jaesung Choi*

  • Department of Statistics, Keimyung University, Korea

Received: February 27, 2020;   Published: March 11, 2020

*Corresponding author: Department of Statistics, Keimyung University, Korea

DOI: 10.32474/CTBB.2020.02.000140

Abstract PDF

Abstract

This paper discusses projection methods to find nonnegative estimates for variance components of random effects in mixed models. The proposed methods are based on the concepts of projections, which are called projection method I, II and III. These three methods produce the same nonnegative estimates for the same data. Even though each method uses orthogonal projections in its own way, the results are the same for the variance components regardless of which method is used. It is shown that quadratic forms of an observation vector are constructed differently in each method. All sums of squares in quadratic forms of the observation vector can be expressed as squared distances of corresponding projections. A projection model is defined and used to evaluate expected values of quadratic forms of observations that are associated with variance components. Hartley’s synthesis is used as a method for finding the coefficients of variance components.

Keywords:Mixed model; projections; quadratic forms; random effects; synthesis

Introduction

Much literature has been devoted to the estimation of variance components in random effects or mixed effects models. A variance component should always be nonnegative by its definition; however, we sometimes get it as negative [1,2]. illustrated this with the simple hypothetical data of a one-way classification having three observations in two classes and insisted that there was nothing intrinsic in the analysis of variance method to prevent it. When a negative estimate happens, it is not easy to handle this situation in interpretation and action. Hence, many papers have been contributed to strategies to deal with the negative values as estimates of variance components [3]. suggests that negative estimates of variance components can occur in certain designs such as split plot and randomized block designs by random- inaction. Thompson discusses the interpretation of the negative estimate and suggests an alternative method when the analysis of variance method yields negative estimates [4]. also suggest a procedure for eliminating negative estimates of variance components in random effects models. The analysis of the variance method is almost exclusively applied to balanced data for estimating variance components. However, there are multiple methods for unbalanced data. Therefore, it is necessary to identify the types of data before choosing a method. Though balanced data have the same numbers of observations in each cell, unbalanced data have unequal numbers of observations in the subclasses made by the levels of classification factors. Depending on the types of data, many methods can be applied to the estimation of variance components in a vector space. Representing data as vectors, the vector space of an observation vector can be partitioned in many ways, depending on the data structure. For balanced data, the vector space can always be partitioned into orthogonal vector subspaces according to the sources of variation, but it is not true for unbalanced data. This is the main difference between balanced and unbalanced data from the view point of a vector space. A random effect is a random variable representing the effect of a randomly chosen level from a population of levels that a random factor can assume, while a fixed effect is an unknown constant denoting the effect of a predetermined level of a factor. A linear model with these two types of effects is called a mixed effects model. The primary concern with the model in this paper is naturally in the nonnegative estimation of variance components of random effects. A negative estimate can happen in any method that contributes to the estimation.
Hence, many papers have investigated strategies for interpretation and alternatives. Such strategies are seen in [5-9]. However, it is necessary to have a method that yields nonnegative estimates despite all such efforts [10].suggested a method that uses reductions in sums of squares due to fitting both the full model and different sub-models of it for estimating variance components of random effects in mixed models. This method is called the fitting constants method or Henderson’s Method 3. Even though it has been used extensively for the estimation of variance components in mixed models, it still has some defects producing negative estimates in case [11]. synthesis is also used for calculating the coefficients of variance components in the method. Although this method is very useful, we should recognize whether quadratic forms for variance components are in the right form or not. Otherwise, expectations of the quadratic forms can be different from the real ones. This is going to be discussed in detail in projection model building. This paper suggests three methods to produce nonnegative estimates for variance components in mixed models. They are based on the concept of projection defined on a vector space. The definition of a projection and its related concepts are discussed in [12,13]. Quadratic forms in the observations can be obtained as squared distances of projections defined in proper vector subspaces. Each method requires that all vector subspaces for projections should be orthogonal to each other at the stage of fitting sub-models serially. When the orthogonality is satisfied with vector subspaces, it is possible to get nonnegative estimates. Hence, we also discuss how to construct orthogonal vector subspaces from a given mixed model. Quadratic forms as sums of squares due to random effects are then used to evaluate their expected values. Hereafter, equating quadratic forms to their expected values represents available equations for the estimates. For calculating the coefficients of variance components, Hartley’s synthesis is applied but in a different manner, which will be discussed.

Mixed Models

Mixed models are used to describe data from experimental situations where some factors are fixed, and others are random. When two types of factors are considered in experiments, one is interested in both parts, that is, the fixed-effects part and the random-effects part, in models. Let α be a vector of all the fixed effects except μ in a mixed model and let δi denote a set of random effects for random factor i for i = 1, 2, r. Then, i δ could be interaction effects or nested-factor effects when they are simply regarded as effects from random factors. The matrix notation of the mixed model for an observation vector y is

where jμ + XFαF is the fixed part of the model and XRδR + ∈ is the random part of the model. δi s are assumed to be independent and identically distributed as N(0, δσi2 I ), and ε is assumed to be distributed as N(0,σ2I ) . The mean and variance of y from (1) is

The expectation of any quadratic form in the observations of a vector y is represented as a function of variance components and fixed effects. The variance components of the full model can be estimated by the fitting constants method of using reductions in the sums of squares due to fitting the full model and the submodel of it. This method provides unbiased estimators of the variance components that do not depend on any fixed effects in the model, and it has been widely used for the estimation of variance components for unbalanced data. However, it still has an unsolved problem having negative solutions as estimates. As an alternative, a method which is based on the concepts of projections is suggested [14]. To discuss it, we consider the model (1) as representative. Since there are two parts in the model, we naturally divide the model into a fixed part and a random part. The random part of the model consists of random effects and errors:

where ∈R = Σi=1r Xiδi+ ∈ The general mean μ and fixed effects αF of (5) can be estimated from normal equations. Regarding y as an observation vector in the n-dimensional vector space, it can be decomposed into two component vectors orthogonal to each other. The decomposition of y is done by projecting y onto the vector subspace generated by ( j, XF ) .

Projection method

Since a method based on the concept of projection is discussed, it will be called the projection method. For a mixed model such as (5), we can decompose y into two components by means of projections. Denoting (j, XF) and (μ, αF)T by Xym and μm , respectively, the projection of y onto the vector subspace spanned by Xm is Xm X my − , where Xm − denotes a Moore-Penrose generalized inverse of XM . Then, y can be decomposed into two vectors, that is, Xm X −my and (I -XMX−M )y − which are orthogonal [15,16]. Instead of the fitting constants method, the projection method is attempted to estimate the nonnegative estimates of the variance components in a mixed model. To explain the method simply, suppose there are two factors A and B for a two-way crossclassified unbalanced data where A is fixed with a levels and B is random with b levels. The model for this is

where y is an observation vector in the n dimensional vector space, αF is a vector of fixed effects of A, δβ and δαβ represent vectors of random effects of B and AB interaction respectively, and XM = ( j, XF ), αM(μ, αF)T and ∈M = Xβδβ + Xαβ δαβ +∈. The second εexpression of (6) represents the fixed-effects part and the random part. The random part SM is obtained by the projection of y onto a vector subspace generated by the XM, which is (I -XM XM−) y. So, y is represented as

where yM= XM XM−y satisfies the two conditions for being the projection of y onto a vector subspace spanned by the columns of XM. The projection should be obtained by the orthogonal projection to the subspace and denoted as a linear combination of the column vectors of XM.XM XM-y conditions. Since yM is orthogonal to eM, the random part eM=(I− XM XM-)y is not affected by the fixed effects and has all the information about the variance components and random error variance. Since there are two random effects and random error terms in the model of (6), we can use eM for finding the related variance components. The model for the estimation of σ2β is

where yB XBXB−eM = is the projection of eM onto the column space of XB ⋅ yB and eB are orthogonal each other. Hence, eB is not affected by the random effects B δ Therefore, eB is used for finding the subspace that has information about σ2αβ. The model for this is

where XAB =(I − XM XM XBXB) Xαβ and ∈αβ=(I− XM XM XB XB) Hence, the projection of eB onto the subspace generated by XAB is yAB= XAB XABeB. Then,

where eAB is (I − XAB XAB )eB. Finally, we can use eAB for finding the coefficient matrix of the random error vector which generates the error space orthogonal to all the other spaces.

Thus, we can know that eAB has all the information about σ2 of the random error vector ∈ . Denoting y as the sum of orthogonal projections and error part,

Each term of (13) can be used to calculate the sums of squares that are quadratic forms in the observations. Since y is partitioned as four terms, there are four available sums of squares. We denote them SSM, SSB, SSAB and SSE where subscripts are corresponding factors. They are defined as

where each SS term is given as the squared length of the projection of y onto its own vector subspace, and XE= (I−XMXM XBXB XAB XAB). All the sums of squares are evaluated by using the eigenvalues and eigenvectors of the projection matrices associated with the quadratic forms in y. Since projections are defined on subspaces that are orthogonal to each other, we can identify the coefficient matrices spanning them.

Projection model

Since y is made up by the sum of mutual orthogonal projections such as (13), y can be represented by the orthogonal coefficient’s matrices of the effects of the assumed model (6). Temporarily, we denote y as yp for differentiating the model based on projections to the classical model (6). Then, the model for yp is

y = XMαM + XBαβ + XABδαβ + XE ∈ (15)

where yp = y . Since each coefficient matrix of the effects is derived from the corresponding orthogonal projection, the equation of (15) defines a projection model that is different from a classical two-way linear mixed model (6). It is useful for evaluating the coefficients of the variance components in the expectations of the quadratic form of an observation vector yp . In the model, all the coefficient matrices are orthogonal to

each other. δβ, δαβ and s are assumed to be N(0,σ2βIb), N(0,σ2αβIab) and N(0,σ2In) respectively. The expectation and the covariance matrix of y p of the projection model (15) is

Equating the three sums of squares, SSB, SSAB, and SSE of (14) to their correspond- ing expectations leads to linear equations in the variance components, the solutions to which are taken as the estimators of those components. Now, the equations are

Solutions from the linear equations (18) are nonnegative estimates of the variance components. Since there are three different ways of getting sums of squares by means of projections, we will differentiate them with projection method I, II, and III. The procedure using the system of linear equations like (18) is called projection method I. The projection method II uses residual vectors after projecting y onto orthogonal subspaces. That is, eM, eB, and eAB are used such as. Then

Since eM has three random components, eMT eM in the quadratic form of yp in which the coefficients matrices of the projection model are orthogonal is available for esti- mating their variance components. Denoting eMT eM as RSSM,

RSSM = eMT eM, (20)

where M RSS measures the variation due to the three random effects, and thus, the quantity is used for the estimation of three variance components σ2β , σ2αβ, and σ2. Representing the residual random vector eB as yp , has two random components as follows.

Hence, eTB is used as an variation quantity for two random effects vectors. Denoting eTB as RSSB,

where RSSB is used for estimating the two variance components σ2αβ and σ2 since eB has just two random effects. Finally, expressing eAB as yp

eAB = (I − XAB XAB) eB = XE ∈ (23)

which has just one random component s. Therefore, eABT eAB shows the variation due to the random error vector only, and this quantity is used for estimating the variance component σ2 ∈ . Denoting eABT eAB as RSSAB,

RSSAB = eABT eAB (24)

Hence,RSSM , RSSB , and RSSAB are another set of sums of squares for estimating variance components instead of using sums of squares derived from the projections as an alternative method.RSSM , RSSB , and RSSAB are also evaluated by using the eigenvalues and eigenvectors of the projection matrices associated with the quadratic forms in y. Now, the expected values of the RSS’s are

Then, the linear equations of variance components are obtained by equating the RSS’s to their expected values, the solutions for which always produce nonlinear estimates.
That is,

Even though two systems of linear equations are not the same, either system will produce the same estimates of the variance components that are nonnegative. As another method, projection method III is also available for the estimation of variance components. This method is done as follows. For the model of (6), y = Xθ +∈, where X = (j, XF, Xβ, Xαβ) and θ =( μ ,αF, δβ, δαβ )T. This method splits the vector space of an observation vector into two subspaces, one for the projection part and the other for the error part at each step. Then, the projection of y onto the subspace spanned by XX is given by XX y , and the error vector in the error vector space is (I − XX ) y . Therefore, the coefficient matrix of s is derived as (I − XX )from it. The quadratic form y '(I − XX ) y denoted by 0 BSS is the sum of squares due to random error only, which has all the information about σ2 . For information about both σ2αβ and σ2, the vector space of the observation vector can be decomposed into two parts one for the projection part and the other for the error part. For this, the model to be fitted is y = X1 θ1 +∈1 , where X1 (j j, XF, Xβ ) θ1 = (μ, αF, δβ ) and ∈1= Xαβ ααβ +∈. Then, the projection of y onto the subspace spanned by X1 X1 is given by X1 X1 y, and the error vector in the errror vector space is (I-X1 X1)y . The quadratic form yT(I-X1 X1)y denoted by BSS1 has information about σ2αβ and σ2 . Now, the error vector is represented by

Hence, the coefficient matrix of δβ is given by (I-X2X2 )Xβ. It is necessary to evaluate the expected values of the quadratic forms for constructing the equations for the variance components. They are

The nonnegative estimates of variance components are given as solutions of linear equations of σ2α and σ2αβ σ2. The above equations are summarized as follows:

where cij ’s are coefficients of variance components of expected values of quadratic forms of (29).

Examples

As a first example of nonnegative estimates of random effects for a two-way mixed model, Montgomery (2013)’s data are illustrated. The data are from an experiment for a gauge capability study where parts are randomly selected, and three operators are fixed. An instrument or gauge is used to measure a critical dimension on a part. Twenty parts have been selected from the production process, and only three operators are assumed to use the gauge. The assumed model for the data in Table 1 is yijk = μ + αij + (αγ)ij + ∈ijk, where they αi (i = 1, 2, 3) are fixed effects such that Σi=13αi = 0 and γj (j = 1, 2,...., 20), (αγ)ij, and ∈ijk are uncorrelated random variables having zero means and variances Var(γj)= σγ2, Var((αγ)ij)= σαγ2 Var(∈ijk)= σ2. Under the assumed unrestricted model, estimated variance components are σγ = 10.2798 , σαγ= −0.1399 , and σ = 0.9917 . Applying the projection method, I to the data, the linear equations of variance components are given as follows:

The solutions of the equations are σγ = 10.3985, σαγ= 0.3559, σ = 0.9917. All the variance components are estimated nonnegatively. When we apply projection method II to the same data, we get

Table 1: Data for a measurement systems capability study from Montgomery.

Lupinepublishers-openaccess-Biostatistics-Biometrics-journal

where RSSfixed = 1271.975 , RSSpart = 86.55 , and RSSpart×operator = 59.5 . The solutions for the equations are σˆ2γ =10.3985, σˆ2αγ =0.3559, σˆ2 =0.9917 which are the same as the previous solutions. Hence, either one of the projection methods can be used for the nonnegative estimation of variance components of random effects in a mixed model. Projection method III also gives the same result as projection methods I and II for the data. As a second example, Searle [2]’s hypothetical data are illustrated. Searle explains why a negative estimate can occur in the estimation of variance component of random effects in a random model. The data are shown in Table 1. Since class in Table 2 is a random factor, the one-way random effects model is assumed. The assumed model is yij = μ+αi+ ∈ij, where the αi(i=1,2) are random effects and ∈ij are uncorrelated random errors having zero means and variances Vari)= σ2α and Var(∈ij)= σ2. As a result of the analysis of variance, the estimates of variance components are given as σˆα2 = −15.33 and σˆ∈2 = 52 . Searledemonstrated how negative estimates could come from the analysis of variance and insisted that there would be nothing intrinsic in the method to prevent it. However, the projection methods yield the same nonnegative estimates as σˆ2α = 2 and σˆ2 = 52 in any method.

Table 2: Hypothetical data of a one-way classification from Searle and Gruber [2].

Lupinepublishers-openaccess-Biostatistics-Biometrics-journal

Conclusion

Variance should be a nonnegative quantity as a measure of variation in data by its definition. In this work, it shows that orthogonal projections are very useful for defining a projection model for nonnegative variance estimation. Although there have been many attempts in literature to fix the problem of negative estimates for variance components over decades, those were not successful. However, the proposed methods in this paper always produce nonnegative estimates of variance components of the random effects in a mixed model. The two most important findings are checked and discussed for the estimation of nonnegative variance component. One is that a projection model should be derived from an assumed mixed-effects model. The other is that expectations of quadratic forms associated with the random effects should be evaluated from the projection model. This paper introduces terms such as projection method I, II, and III related to the methods, and the projection model for emphasizing projection rather than model fitting. Though they are based on the same assumed model, three methods are ap- plied differently in the application. Each method uses in its own way but summing up all orthogonal projections come to the observation vector. Depending on the types of projections, each method produces three different sets of equations for the evaluation of quadratic forms. Nonetheless, all of them show the same nonnegative estimates for variance components. It also shows that projection methods can be used for estimating variance components of the random effects in either random model or mixed model through examples. It should be noted that all the matrices associated with the quadratic forms come from the projection model not from the assumed model. In such a case, Hartley’s synthesis can yield correct coefficients of variance components.

Funding

This work was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education under Grant No.2018R1D1A1B07043021.

References

  1. Searle SR (1971) Linear models. John Wiley and Sons New York, USA p. 1.
  2. Searle SR, Gruber MHJ (2016) Linear Models. John Wiley and Sons New York, USA.
  3. Nelder J (1954) A The interpretation of negative components of variance. Biometric 41: 544-548.
  4. Milliken GA, Johnson DE (1984) Analysis of messy data volume 1: designed experiments. Van Nostrand Reinhold New York, USA.
  5. Searle SR, Fawcett RF (1970) Expected mean squares in variance components models having finite populations. Biometrics 26(2): 243-254.
  6. Hill BM (1965) Inference about variance components in the one-way mode. J Am Stat Assoc 60(311): 806-825.
  7. Hill BM (1967) Correlated errors in the random model. J Am Stat Assoc 62(320): 1387-1400.
  8. Searle SR, Casella G, McCulloch CE (2009) Variance components. John Wiley and Sons New York, USA.
  9. Harville DA (1969) Variance component estimation for the unbalanced one-way random classification a critique. Aerospace Research Laboratories ARL p. 69.
  10. Henderson CR (1953) Estimation of variance and covariance components. Biometrics 9(2): 226-252.
  11. Hartley HO (1967) Expectations variances and covariances of ANOVA means squares bysynthesis. Biometrics 23: 105-114.
  12. GA (1983) Franklin Matrices with Applications in Statistics. Wadsworth Inc New York, USA.
  13. Johnson RA, Wichern DW (2014) Applied multivariate statistical analysis. Prentice hall Upper Saddle River New Jersey, USA.
  14. Montgomery DC (2013) Design and analysis of experiments. John Wiley and Sons New York, USA.
  15. Thompson WA (1967) Negative estimates of variance components: an introduction Bulletin International Institute of Statistics 34: 1-4.
  16. Thompson WA, Moore JR (1967) Non negative estimates of variance components. Tech no metrics 5(4): 441-449.

https://www.high-endrolex.com/21