通信人家园

标题: 什么是正交化因子?  [查看完整版帖子] [打印本页]

时间:  2009-4-20 11:35
作者: oooooooo     标题: 什么是正交化因子?

正交化因子是什么?
时间:  2009-4-21 16:48
作者: oooooooo

没有人知道吗?
时间:  2009-4-21 17:24
作者: shenhqi

该因子有一个集合,再该集合内所有的因子都是正交的。
时间:  2009-4-21 19:31
作者: cdma1xdj

来个英文的
The Orthogonal Factor Model
The aim of factor analysis is to explain the outcome of  variables in the data matrix  using fewer variables, the so-called factors. Ideally all the information in  can be reproduced by a smaller number of factors. These factors are interpreted as latent (unobserved) common characteristics of the observed . The case just described occurs when every observed  can be written as


(10.1)



Here , for  denotes the factors. The number of factors, , should always be much smaller than . For instance, in psychology  may represent  results of a test measuring intelligence scores. One common latent factor explaining  could be the overall level of ``intelligence''. In marketing studies,  may consist of  answers to a survey on the levels of satisfaction of the customers. These  measures could be explained by common latent factors like the attraction level of the product or the image of the brand, and so on. Indeed it is possible to create a representation of the observations that is similar to the one in (10.1) by means of principal components, but only if the last  eigenvalues corresponding to the covariance matrix are equal to zero. Consider a -dimensional random vector X with mean  and covariance matrix . A model similar to (10.1) can be written for  in matrix notation, namely

(10.2)



where  is the -dimensional vector of the  factors. When using the factor model (10.2) it is often assumed that the factors  are centered, uncorrelated and standardized:  and . We will now show that if the last  eigenvalues of  are equal to zero, we can easily express  by the factor model (10.2).
The spectral decomposition of  is given by . Suppose that only the first  eigenvalues are positive, i.e., . Then the (singular) covariance matrix can be written as







In order to show the connection to the factor model (10.2), recall that the PCs are given by . Rearranging we have , where the components of  are partitioned according to the partition of  above, namely










In other words,  has a singular distribution with mean and covariance matrix equal to zero. Therefore,  implies that  is equivalent to , which can be written as






Defining  and , we obtain the factor model (10.2).
Note that the covariance matrix of model (10.2) can be written as


(10.3)



We have just shown how the variable  can be completely determined by a weighted sum of  (where ) uncorrelated factors. The situation used in the derivation, however, is too idealistic. In practice the covariance matrix is rarely singular.
It is common praxis in factor analysis to split the influences of the factors into common and specific ones. There are, for example, highly informative factors that are common to all of the components of  and factors that are specific to certain components. The factor analysis model used in praxis is a generalization of (10.2):


(10.4)



where  is a  matrix of the (non-random) loadings of the common factors  and  is a  matrix of the (random) specific factors. It is assumed that the factor variables  are uncorrelated random vectors and that the specific factors are uncorrelated and have zero covariance with the common factors. More precisely, it is assumed that:

     
     
   (10.5)
     
     




Define







The generalized factor model (10.4) together with the assumptions given in (10.5) constitute the orthogonal factor model.



Orthogonal Factor Model   
=         
()   () ()   ()   ()   
     = mean of variable  
     = -th specific factor
     = -th common factor
     = loading of the -th variable on the -th factor
                  
The random vectors  and  are unobservable and uncorrelated.


Note that (10.4) implies for the components of  that


(10.6)



Using (10.5) we obtain . The quantity  is called the communality and  the specific variance. Thus the covariance of  can be rewritten as

     
      
    (10.7)




In a sense, the factor model explains the variations of  for the most part by a small number of latent factors  common to its  components and entirely explains all the correlation structure between its components, plus some ``noise''  which allows specific variations of each component to enter. The specific factors adjust to capture the individual variance of each component. Factor analysis relies on the assumptions presented above. If the assumptions are not met, the analysis could be spurious. Although principal components analysis and factor analysis might be related (this was hinted at in the derivation of the factor model), they are quite different in nature. PCs are linear transformations of  arranged in decreasing order of variance and used to reduce the dimension of the data set, whereas in factor analysis, we try to model the variations of  using a linear transformation of a fixed, limited number of latent factors. The objective of factor analysis is to find the loadings  and the specific variance . Estimates of  and  are deduced from the covariance structure (10.7).


Interpretation of the Factors
Assume that a factor model with  factors was found to be reasonable, i.e., most of the (co)variations of the  measures in  were explained by the  fixed latent factors. The next natural step is to try to understand what these factors represent. To interpret , it makes sense to compute its correlations with the original variables  first. This is done for  and for  to obtain the matrix . The sequence of calculations used here are in fact the same that were used to interprete the PCs in the principal components analysis.

The following covariance between  and  is obtained via (10.5),







The correlation is

(10.8)



where . Using (10.8) it is possible to construct a figure analogous to Figure 9.6 and thus to consider which of the original variables  play a role in the unobserved common factors .
Returning to the psychology example where  are the observed scores to  different intelligence tests (the WAIS data set in Table B.12 provides an example), we would expect a model with one factor to produce a factor that is positively correlated with all of the components in . For this example the factor represents the overall level of intelligence of an individual. A model with two factors could produce a refinement in explaining the variations of the  scores. For example, the first factor could be the same as before (overall level of intelligence), whereas the second factor could be positively correlated with some of the tests, , that are related to the individual's ability to think abstractly and negatively correlated with other tests, , that are related to the individual's practical ability. The second factor would then concern a particular dimension of the intelligence stressing the distinctions between the ``theoretical'' and ``practical'' abilities of the individual. If the model is true, most of the information coming from the  scores can be summarized by these two latent factors. Other practical examples are given below.



Invariance of Scale
What happens if we change the scale of  to  with ? If the -factor model (10.6) is true for  with , , then, since







the same -factor model is also true for  with  and . In many applications, the search for the loadings  and for the specific variance  will be done by the decomposition of the correlation matrix of  rather than the covariance matrix . This corresponds to a factor analysis of a linear transformation of  (i.e., . The goal is to try to find the loadings  and the specific variance  such that

(10.9)



In this case the interpretation of the factors  immediately follows from (10.8) given the following correlation matrix:

(10.10)



Because of the scale invariance of the factors, the loadings and the specific variance of the model, where  is expressed in its original units of measure, are given by










It should be noted that although the factor analysis model (10.4) enjoys the scale invariance property, the actual estimated factors could be scale dependent. We will come back to this point later when we discuss the method of principal factors.


Non-Uniqueness of Factor Loadings
The factor loadings are not unique! Suppose that  is an orthogonal matrix. Then  in (10.4) can also be written as







This implies that, if a -factor of  with factors  and loadings  is true, then the -factor model with factors  and loadings  is also true. In practice, we will take advantage of this non-uniqueness. Indeed, referring back to Section 2.6 we can conclude that premultiplying a vector  by an orthogonal matrix corresponds to a rotation of the system of axis, the direction of the first new axis being given by the first row of the orthogonal matrix. It will be shown that choosing an appropriate rotation will result in a matrix of loadings  that will be easier to interpret. We have seen that the loadings provide the correlations between the factors and the original variables, therefore, it makes sense to search for rotations that give factors that are maximally correlated with various groups of variables.

From a numerical point of view, the non-uniqueness is a drawback. We have to find loadings  and specific variances  satisfying the decomposition , but no straightforward numerical algorithm can solve this problem due to the multiplicity of the solutions. An acceptable technique is to impose some chosen constraints in order to get--in the best case--an unique solution to the decomposition. Then, as suggested above, once we have a solution we will take advantage of the rotations in order to obtain a solution that is easier to interprete.

An obvious question is: what kind of constraints should we impose in order to eliminate the non-uniqueness problem? Usually, we impose additional constraints where


(10.11)



or

(10.12)



How many parameters does the model (10.7) have without constraints?











Hence we have to determine  parameters! Conditions (10.11) respectively (10.12) introduce  constraints, since we require the matrices to be diagonal. Therefore, the degrees of freedom of a model with  factors is:










If , then the model is undetermined: there are infinitly many solutions to (10.7). This means that the number of parameters of the factorial model is larger than the number of parameters of the original model, or that the number of factors  is ``too large'' relative to . In some cases : there is an unique solution to the problem (except for rotation). In practice we usually have that :there are more equations than parameters, thus an exact solution does not exist. In this case approximate solutions are used. An approximation of , for example, is . The last case is the most interesting since the factorial model has less parameters than the original one. Estimation methods are introduced in the next section.

Evaluating the degrees of freedom, , is particularly important, because it already gives an idea of the upper bound on the number of factors we can hope to identify in a factor model. For instance, if , we could not identify a factor model with 2 factors (this results in  which has infinitly many solutions). With , only a one factor model gives an approximate solution (). When , models with 1 and 2 factors provide approximate solutions and a model with 3 factors results in an unique solution (up to the rotations) since . A model with 4 or more factors would not be allowed, but of course, the aim of factor analysis is to find suitable models with a small number of factors, i.e., smaller than . The next two examples give more insights into the notion of degrees of freedom.



EXAMPLE 10.1   Let  and , then  and






with  and . Note that here the constraint (10.8) is automatically verified since . We have






and






In this particular case (), the only rotation is defined by , so the other solution for the loadings is provided by .



EXAMPLE 10.2   Suppose now  and , then  and






We have infinitely many solutions: for any  , a solution is provided by







The solution in Example 10.1 may be unique (up to a rotation), but it is not proper in the sense that it cannot be interpreted statistically. Exercise 10.5 gives an example where the specific variance  is negative.





1mm

Even in the case of a unique solution , the solution may be inconsistent with statistical interpretations.


Summary

The factor analysis model aims to describe how the original  variables in a data set depend on a small number of latent factors , i.e., it assumes that . The (-dimensional) random vector  contains the common factors, the (-dimensional)  contains the specific factors and  contains the factor loadings.

It is assumed that  and  are uncorrelated and have zero means, i.e., ,  where  is diagonal matrix and .
This leads to the covariance structure .

The interpretation of the factor  is obtained through the correlation .

A normalized analysis is obtained by the model . The interpretation of the factors is given directly by the loadings .

The factor analysis model is scale invariant. The loadings are not unique (only up to multiplication by an orthogonal matrix).

Whether a model has an unique solution or not is determined by the degrees of freedom .
时间:  2009-4-21 19:33
作者: cdma1xdj

图没搞过来




通信人家园 (https://www.txrjy.com/) Powered by C114