arXiv Analytics

Sign in

arXiv:1212.2466 [cs.LG]AbstractReferencesReviewsResources

On Information Regularization

Adrian Corduneanu, Tommi S. Jaakkola

Published 2012-10-19Version 1

We formulate a principle for classification with the knowledge of the marginal distribution over the data points (unlabeled data). The principle is cast in terms of Tikhonov style regularization where the regularization penalty articulates the way in which the marginal density should constrain otherwise unrestricted conditional distributions. Specifically, the regularization penalty penalizes any information introduced between the examples and labels beyond what is provided by the available labeled examples. The work extends Szummer and Jaakkola's information regularization (NIPS 2002) to multiple dimensions, providing a regularizer independent of the covering of the space used in the derivation. We show in addition how the information regularizer can be used as a measure of complexity of the classification task with unlabeled data and prove a relevant sample-complexity bound. We illustrate the regularization principle in practice by restricting the class of conditional distributions to be logistic regression models and constructing the regularization penalty from a finite set of unlabeled examples.

Comments: Appears in Proceedings of the Nineteenth Conference on Uncertainty in Artificial Intelligence (UAI2003)
Categories: cs.LG, stat.ML
Related articles: Most relevant | Search more
arXiv:1904.11717 [cs.LG] (Published 2019-04-26)
Classification from Pairwise Similarities/Dissimilarities and Unlabeled Data via Empirical Risk Minimization
arXiv:1809.03207 [cs.LG] (Published 2018-09-10)
Beyond the Selected Completely At Random Assumption for Learning from Positive and Unlabeled Data
arXiv:1911.08696 [cs.LG] (Published 2019-11-20)
Where is the Bottleneck of Adversarial Learning with Unlabeled Data?