arXiv Analytics

Sign in

arXiv:2206.01646 [cs.CV]AbstractReferencesReviewsResources

Rethinking Positive Sampling for Contrastive Learning with Kernel

Benoit Dufumier, Carlo Alberto Barbano, Robin Louiset, Edouard Duchesnay, Pietro Gori

Published 2022-06-03Version 1

Data augmentation is a crucial component in unsupervised contrastive learning (CL). It determines how positive samples are defined and, ultimately, the quality of the representation. While efficient augmentations have been found for standard vision datasets, such as ImageNet, it is still an open problem in other applications, such as medical imaging, or in datasets with easy-to-learn but irrelevant imaging features. In this work, we propose a new way to define positive samples using kernel theory along with a novel loss called decoupled uniformity. We propose to integrate prior information, learnt from generative models or given as auxiliary attributes, into contrastive learning, to make it less dependent on data augmentation. We draw a connection between contrastive learning and the conditional mean embedding theory to derive tight bounds on the downstream classification loss. In an unsupervised setting, we empirically demonstrate that CL benefits from generative models, such as VAE and GAN, to less rely on data augmentations. We validate our framework on vision datasets including CIFAR10, CIFAR100, STL10 and ImageNet100 and a brain MRI dataset. In the weakly supervised setting, we demonstrate that our formulation provides state-of-the-art results.

Related articles: Most relevant | Search more
arXiv:2005.10243 [cs.CV] (Published 2020-05-20)
What makes for good views for contrastive learning
arXiv:2106.09958 [cs.CV] (Published 2021-06-18)
Novelty Detection via Contrastive Learning with Negative Data Augmentation
arXiv:2206.12925 [cs.CV] (Published 2022-06-26)
Vision Transformer for Contrastive Clustering