Abstract
The research on robust principal component analysis (RPCA) has been attracting much atten-tion recently. The original RPCA model assumes sparse noise, and use the L1-norm to characterize the error term. In practice, however, the noise is much more complex and it is not appropriate to simply use a certain Lp-norm for noise modeling. We propose a generative RPCA model under the Bayesian framework by modeling data noise as a mixture of Gaussians (MoG). The MoG is a uni-versal approximator to continuous distributions and thus our model is able to fit a wide range of noises such as Laplacian, Gaussian, sparse noises and any combinations of them. A variational Bayes algorithm is presented to infer the posterior of the proposed model. All involved parameters can be recursively updated in closed form. The advantage of our method is demonstrated by extensive experiments on synthetic data, face modeling and background subtraction.
Original language | English |
---|---|
Title of host publication | 31st International Conference on Machine Learning, ICML 2014 |
Publisher | International Machine Learning Society (IMLS) |
Pages | 1216-1226 |
Number of pages | 11 |
Volume | 2 |
ISBN (Electronic) | 9781634393973 |
Publication status | Published - 1 Jan 2014 |
Event | 31st International Conference on Machine Learning, ICML 2014 - Beijing, China Duration: 21 Jun 2014 → 26 Jun 2014 |
Conference
Conference | 31st International Conference on Machine Learning, ICML 2014 |
---|---|
Country/Territory | China |
City | Beijing |
Period | 21/06/14 → 26/06/14 |
ASJC Scopus subject areas
- Artificial Intelligence
- Computer Networks and Communications
- Software