arXiv Analytics

Sign in

arXiv:1804.10696 [cs.DS]AbstractReferencesReviewsResources

Low Rank Approximation in the Presence of Outliers

Aditya Bhaskara, Srivatsan Kumar

Published 2018-04-27Version 1

We consider the problem of principal component analysis (PCA) in the presence of outliers. Given a matrix $A$ ($d \times n$) and parameters $k, m$, the goal is to remove a set of at most $m$ columns of $A$ (known as outliers), so as to minimize the rank-$k$ approximation error of the remaining matrix. While much of the work on this problem has focused on recovery of the rank-$k$ subspace under assumptions on the inliers and outliers, we focus on the approximation problem above. Our main result shows that sampling-based methods developed in the outlier-free case give non-trivial guarantees even in the presence of outliers. Using this insight, we develop a simple algorithm that has bi-criteria guarantees. Further, unlike similar formulations for clustering, we show that bi-criteria guarantees are unavoidable for the problem, under appropriate complexity assumptions.

Related articles: Most relevant | Search more
arXiv:1909.13384 [cs.DS] (Published 2019-09-29)
Optimal Sketching for Kronecker Product Regression and Low Rank Approximation
arXiv:1207.6365 [cs.DS] (Published 2012-07-26, updated 2013-04-05)
Low Rank Approximation and Regression in Input Sparsity Time
arXiv:1811.00414 [cs.DS] (Published 2018-10-31)
Quantum-inspired classical algorithms for principal component analysis and supervised clustering