{ "id": "2402.15441", "version": "v2", "published": "2024-02-13T09:19:05.000Z", "updated": "2024-03-12T07:44:00.000Z", "title": "Active Few-Shot Fine-Tuning", "authors": [ "Jonas Hübotter", "Bhavya Sukhija", "Lenart Treven", "Yarden As", "Andreas Krause" ], "categories": [ "cs.LG", "cs.AI" ], "abstract": "We study the active few-shot fine-tuning of large neural networks to downstream tasks. We show that few-shot fine-tuning is an instance of a generalization of classical active learning, transductive active learning, and we propose ITL, short for information-based transductive learning, an approach which samples adaptively to maximize the information gained about specified downstream tasks. Under general regularity assumptions, we prove that ITL converges uniformly to the smallest possible uncertainty obtainable from the accessible data. To the best of our knowledge, we are the first to derive generalization bounds of this kind, and they may be of independent interest for active learning. We apply ITL to the few-shot fine-tuning of large neural networks and show that ITL substantially improves upon the state-of-the-art.", "revisions": [ { "version": "v2", "updated": "2024-03-12T07:44:00.000Z" } ], "analyses": { "keywords": [ "active few-shot fine-tuning", "large neural networks", "active learning", "general regularity assumptions", "specified downstream tasks" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }