Research Article

Online Coregularization for Multiview Semisupervised Learning

Table 4

Mean test error rates on the rotating two-moons-two-lines synthetic data sequence. The error rates are reported for three different sparse approximations. For gradient ascent, we choose a stationary step size . The result shows that our derived online co-regularization algorithms are able to track the changes in the sequence and maintain a smaller error rate compared with batch learning algorithms. Specially, ADA (Buffer- ) performs better than the other online co-regularization algorithms.

Online co-regularization algorithmsGradient ascentADA ADA Buffer- ADA Buffer-

Error rates (%)
 No sparse approximation12.757.756.4320.35
 Absolute threshold ( )11.5510.759.2718.75
-MC ( )11.5510.359.4518.35