arXiv Analytics

Sign in

arXiv:2302.12715 [cs.LG]AbstractReferencesReviewsResources

Hiding Data Helps: On the Benefits of Masking for Sparse Coding

Muthu Chidambaram, Chenwei Wu, Yu Cheng, Rong Ge

Published 2023-02-24Version 1

Sparse coding refers to modeling a signal as sparse linear combinations of the elements of a learned dictionary. Sparse coding has proven to be a successful and interpretable approach in many applications, such as signal processing, computer vision, and medical imaging. While this success has spurred much work on sparse coding with provable guarantees, work on the setting where the learned dictionary is larger (or \textit{over-realized}) with respect to the ground truth is comparatively nascent. Existing theoretical results in the over-realized regime are limited to the case of noise-less data. In this paper, we show that for over-realized sparse coding in the presence of noise, minimizing the standard dictionary learning objective can fail to recover the ground-truth dictionary, regardless of the magnitude of the signal in the data-generating process. Furthermore, drawing from the growing body of work on self-supervised learning, we propose a novel masking objective and we prove that minimizing this new objective can recover the ground-truth dictionary. We corroborate our theoretical results with experiments across several parameter regimes, showing that our proposed objective enjoys better empirical performance than the standard reconstruction objective.

Related articles: Most relevant | Search more
arXiv:2011.11805 [cs.LG] (Published 2020-11-24)
The Interpretable Dictionary in Sparse Coding
arXiv:1503.00778 [cs.LG] (Published 2015-03-02)
Simple, Efficient, and Neural Algorithms for Sparse Coding
arXiv:1209.0738 [cs.LG] (Published 2012-09-04, updated 2014-06-16)
Sparse coding for multitask and transfer learning