I.6.5.2: Optimal Scaling with ORDINALS

In LINEALS (section x.x.x) we try to find quantifications of the variables that linearize all bivariate regressions. De Leeuw [1988] has suggested to find standardized quantifications in such a way that the loss function is minimized.

A more general loss function is which must be minimized over both and . The are vectors, called regression targets, and target has elements.

To see that this loss function generalizes suppose we constrain by requiring that is proportional to , i.e. Then, using , This is minimized over by , and the minimum is precisely the loss function . Thus and is an augmentation of . Block relaxation for alternates minimization over for fixed , which we have shown to be easy, and minimization over for fixed , which is a modified eigenvalue problem of the kind discussed in BRAS3, section x.x.x. This is not necessarily simpler than the direct minimum eigenvalue problem for minimizing in section x.x.x.

The major advantage from augmenting is that it now becomes simple to incorporate quite general restrictions on the For example, they can be required to be monotone with the original data, or a spline transformation, or a monotone spline. Or a mixture of these options. Thus we can constrain each individual regression functions to have one of a pre-determined number of shapes.

In ordinals.R we implement the three standard options of the Gifi system. A vector is treated as nominal, ordinal, or numerical. If it is nominal then it is unconstrained, except for the normalization. In that case the are also unconstrained for all . If is treated as ordinal is must be monotone with the data, and so must all . And a numerical must be linear with the data, together with its targets . Of course if all variables are numerical there is nothing to optimize, and we just compute correlations. If all variables are nominal there is nothing to optimize either, because we immediately get zero loss from any starting point.


[Insert ordinals.R Here](../code/ordinals.R)