arXiv:2402.15441 [cs.LG]AbstractReferencesReviewsResources
Active Few-Shot Fine-Tuning
Jonas Hübotter, Bhavya Sukhija, Lenart Treven, Yarden As, Andreas Krause
Published 2024-02-13, updated 2024-03-12Version 2
We study the active few-shot fine-tuning of large neural networks to downstream tasks. We show that few-shot fine-tuning is an instance of a generalization of classical active learning, transductive active learning, and we propose ITL, short for information-based transductive learning, an approach which samples adaptively to maximize the information gained about specified downstream tasks. Under general regularity assumptions, we prove that ITL converges uniformly to the smallest possible uncertainty obtainable from the accessible data. To the best of our knowledge, we are the first to derive generalization bounds of this kind, and they may be of independent interest for active learning. We apply ITL to the few-shot fine-tuning of large neural networks and show that ITL substantially improves upon the state-of-the-art.