{ "id": "2211.03782", "version": "v1", "published": "2022-11-07T18:57:20.000Z", "updated": "2022-11-07T18:57:20.000Z", "title": "On minimal variations for unsupervised representation learning", "authors": [ "Vivien Cabannes", "Alberto Bietti", "Randall Balestriero" ], "comment": "5 pages, 1 figure; 1 table", "categories": [ "cs.LG", "cs.AI", "stat.ML" ], "abstract": "Unsupervised representation learning aims at describing raw data efficiently to solve various downstream tasks. It has been approached with many techniques, such as manifold learning, diffusion maps, or more recently self-supervised learning. Those techniques are arguably all based on the underlying assumption that target functions, associated with future downstream tasks, have low variations in densely populated regions of the input space. Unveiling minimal variations as a guiding principle behind unsupervised representation learning paves the way to better practical guidelines for self-supervised learning algorithms.", "revisions": [ { "version": "v1", "updated": "2022-11-07T18:57:20.000Z" } ], "analyses": { "subjects": [ "68Q32", "G.3" ], "keywords": [ "downstream tasks", "better practical guidelines", "unsupervised representation learning paves", "raw data", "unveiling minimal variations" ], "note": { "typesetting": "TeX", "pages": 5, "language": "en", "license": "arXiv", "status": "editable" } } }