arXiv Analytics

Sign in

arXiv:2011.11805 [cs.LG]AbstractReferencesReviewsResources

The Interpretable Dictionary in Sparse Coding

Edward Kim, Connor Onweller, Andrew O'Brien, Kathleen McCoy

Published 2020-11-24Version 1

Artificial neural networks (ANNs), specifically deep learning networks, have often been labeled as black boxes due to the fact that the internal representation of the data is not easily interpretable. In our work, we illustrate that an ANN, trained using sparse coding under specific sparsity constraints, yields a more interpretable model than the standard deep learning model. The dictionary learned by sparse coding can be more easily understood and the activations of these elements creates a selective feature output. We compare and contrast our sparse coding model with an equivalent feed forward convolutional autoencoder trained on the same data. Our results show both qualitative and quantitative benefits in the interpretation of the learned sparse coding dictionary as well as the internal activation representations.

Related articles: Most relevant | Search more
arXiv:1607.04917 [cs.LG] (Published 2016-07-17)
Piecewise convexity of artificial neural networks
arXiv:2006.02909 [cs.LG] (Published 2020-06-03)
Assessing Intelligence in Artificial Neural Networks
arXiv:2102.02153 [cs.LG] (Published 2021-02-03)
Fast Concept Mapping: The Emergence of Human Abilities in Artificial Neural Networks when Learning Embodied and Self-Supervised