Robust Subspace Mixture Models using \(t\)-distributions

Motivation(s)

Mixture models have been used to approximate a manifold using a relatively small number of localized subspaces (e.g. PCA, factor analysis). One outstanding problem is robustness to outliers. To address this, recent research have explored mixture of t-distributions and PPCA.

Proposed Solution(s)

The authors assert that for manifold learning, finding an exact local PCA solution is not necessary, as long as the main axes of the densities are aligned with the manifold. The Gaussian prior can be replaced by the t-distribution, which in turn is a convolution of a Gaussian and the gamma prior.

Evaluation(s)

MTS beat MPPCA in the synthetic noisy data and digit recognition experiments. MTS was able to learn the manifold structure despite a higher level of noise. The proposed EM algorithm assumes that the convolution of two t-distributions is a t-distribution, which is a good approximation only for large \(\nu\) or for small \(\sigma^2\).

Future Direction(s)

  • Instead of an approximation, derive the true EM algorithm and compare its performance.

Question(s)

  • While I like the usage of indicator variables to remove a summation, I’m not sure if (9) is valid probabilistically. Then again, if one doesn’t have to define such a distribution, then maybe it’s alright?

Analysis

Hidden indicator variables can be used to reformulate summations for easier learning.

The idea is great, but the reasoning for not going through with a full derivation is weak. The interesting thing is that this overview is clearer than this paper itself.

References

dRF03

Dick de Ridder and Vojtech Franc. Robust subspace mixture models using t-distributions. In BMVC, 1–10. 2003.