Partially Observed Linear Systems

De Leeuw [2004] discusses the problem of finding an approximate solution to the homogeneous linear system when there are cone and orthonormality restrictions on the columns of and when some elements of are restricted to known values, most commonly to zero. Think of the columns of as variables or sets of variables, and think of as regression coefficients or weights.

The loss function used by De Leeuw [2004] is with and with coding the constraints on . Note that the computation of the optimal in is a least squares problem, and even with linear inequality constraints on it is still a straightforward quadratic programming problem.

The function in is the pointwise minimum of linear functions in , and thus it is a concave function of . This means we are in the "aspects of correlation matrices" framework discussed in the previous section.

In particular we define then the subgradient of at is The subgradient inequality now says that for all correlation matrices and we have for all .

The constraints on discussed in De Leeuw [2004] make it possible to fit a wide variety of multivariate analysis techniques. Columns of , or variables, are partitioned into blocks. Some blocks contain only a single variable, such variables are called single. Some blocks are constrained to be orthoblocks, which means that the variabes in the block are required to be orthonormal. Single variables may be cone-constrained, which means the corresponding column of is constrained to be in a cone in . And orthoblocks may be subspace-constrained, which means all columns must be in the same subspace.

We mention some illustrative special cases here. Common factor analysis of a data matrix means finding an approximate solution to the system with , , , and diagonal. The common factor scores are in , the unique factor scores in , the factor loadings in and the uniquenesses in This example can be generalized to cover structural equation models

Homeogeneity analysis [Gifi 1990] is the linear system where is and orthoblock of object scores, while the are orthoblocks in the subspaces defined by the indicator matrices (or B-spline bases) of variable . For single variables only has a single column, which can be cone-constrained. For multiple correspondence analysis and all have the same number of columns. For nonlinear principal component analysis all variables are single and the are .

In both examples the majorization algorithm is actually an alternating least squares algorithm. In the factor analysis example the loss functon is and in homogeneity analysis it is