arXiv Analytics

Sign in

arXiv:2211.01407 [cs.LG]AbstractReferencesReviewsResources

On the Informativeness of Supervision Signals

Ilia Sucholutsky, Raja Marjieh, Nori Jacoby, Thomas L. Griffiths

Published 2022-11-02Version 1

Learning transferable representations by training a classifier is a well-established technique in deep learning (e.g., ImageNet pretraining), but it remains an open theoretical question why this kind of task-specific pre-training should result in ''good'' representations that actually capture the underlying structure of the data. We conduct an information-theoretic analysis of several commonly-used supervision signals from contrastive learning and classification to determine how they contribute to representation learning performance and how the dynamics of learning are affected by training parameters such as the number of labels, classes, and dimensions in the training dataset. We validate these results empirically in a series of simulations and conduct a cost-benefit analysis to establish a tradeoff curve that enables users to optimize the cost of supervising representation learning on their own datasets.

Related articles: Most relevant | Search more
arXiv:2207.08735 [cs.LG] (Published 2022-07-18)
An Information-Theoretic Analysis of Bayesian Reinforcement Learning
arXiv:1403.5341 [cs.LG] (Published 2014-03-21, updated 2015-06-08)
An Information-Theoretic Analysis of Thompson Sampling
arXiv:2302.04452 [cs.LG] (Published 2023-02-09)
An Information-Theoretic Analysis of Nonstationary Bandit Learning