In training, the local PCA mixture models NGPCA, NGPCAconstV, and
MPPCAext were used (chapter 3). On the pattern association task, they were compared to kernel PCA (chapter 5), to a lookup table, and to a multilayer perceptron. All of them used the same preprocessed pattern set.
Section 3.3 mentioned that MPPCAext comprises the following two modifications: a correction for `empty' units and the use of an online PCA algorithm, which allows that noise can be added to each presented training pattern. On the visuomotor data, these two modifications turned out to be essential because otherwise, the algorithm became numerically unstable; some eigenvalues dropped to zero.
For the mixture models, 120 units and four principal components were used. The number of principal components was chosen after inspecting the local dimensionality of the pattern distribution. As described in section 4.5.1, the ratio of successive eigenvalues, averaged from a PCA in the neighborhood of each training pattern, has a peak at the local dimensionality of the distribution (Philipona et al., 2003). On the collected data, the first peak is at three (figure 6.7, left). This matches the expectation, since the brick had three degrees of freedom: two for the position and one for the orientation.

Different from the kinematic arm model in section 4.5, however, figure 6.7 shows a second peak at five dimensions. The reason is probably that in the real robot task, the neighborhood of a pattern also covers the turns and twists of the underlying manifold because the data are much more sparse (3371 training patterns lie in a 68dimensional space, compared to 50 000 patterns in ten dimensions for the kinematic arm model). A turn increases the local variance (figure 6.7, right). Therefore, for the mixture model, four principal components were chosen instead of three (this also improved the performance).
For NGPCA and NGPCAconstV, two sets of training parameters were used. Set 1 had (0) = 1.0, (t_{max}) = 0.00001, (0) = 0.1, (t_{max}) = 0.01, and t_{max} = 400 000. Set 2 had the same parameters as the kinematic arm model (section 4.5): (0) = 10.0, (t_{max}) = 0.0001, (0) = 0.5, (t_{max}) = 0.001, and t_{max} = 400 000.
Kernel PCA extracted 150 eigenvectors and used an inversemultiquadratic kernel with width = 7.0 (see section 2.4.3). In the 68dimensional space, distances are larger than in the previous applications (chapter 5). Thus, the inversemultiquadratic function was of advantage since it does not decline as quickly as a Gaussian function. Moreover, a reduced set with m = 1000 was used (appendix B.2).
The lookuptable method chooses a pattern from the training set whose visual part has the smallest Euclidean distance to the presented input. The multilayer perceptron maps the visual part to the postural part. A structure with one hiddenlayer, containing 20 neurons, was used. The hidden neurons had sigmoid activation functions. The weights were initialized with random values drawn uniformly from the interval [ 0.5;0.5]. For training, 3 000 epochs of resilient propagation were used (Riedmiller and Braun, 1993).