Research Article

Online Coregularization for Multiview Semisupervised Learning

Table 2

Mean test error rates on the two-moons-two-lines synthetic data set. The error rates are reported for three different sparse approximations. For gradient ascent, we choose a decaying step size . The result shows that our derived online co-regularization algorithms achieve test accuracy comparable to offline co-regularization (CoLapSVM). The online co-regularization algorithms based on aggressive dual ascending procedures perform better than those based on gradient ascent.

Online co-regularization
algorithms
Gradient ascentADA ADA ADA Buffer- ADA Buffer-

Error rates (%)
 No sparse approximation12.158.953.954.254.05
 Absolute threshold ( )10.809.653.954.054.05
-MC ( )13.309.853.957.005.35