arXiv Analytics

Sign in

arXiv:2402.15441 [cs.LG]AbstractReferencesReviewsResources

Active Few-Shot Fine-Tuning

Jonas Hübotter, Bhavya Sukhija, Lenart Treven, Yarden As, Andreas Krause

Published 2024-02-13, updated 2024-03-12Version 2

We study the active few-shot fine-tuning of large neural networks to downstream tasks. We show that few-shot fine-tuning is an instance of a generalization of classical active learning, transductive active learning, and we propose ITL, short for information-based transductive learning, an approach which samples adaptively to maximize the information gained about specified downstream tasks. Under general regularity assumptions, we prove that ITL converges uniformly to the smallest possible uncertainty obtainable from the accessible data. To the best of our knowledge, we are the first to derive generalization bounds of this kind, and they may be of independent interest for active learning. We apply ITL to the few-shot fine-tuning of large neural networks and show that ITL substantially improves upon the state-of-the-art.

Related articles: Most relevant | Search more
arXiv:2011.14015 [cs.LG] (Published 2020-11-27)
Active Learning in CNNs via Expected Improvement Maximization
arXiv:1905.12782 [cs.LG] (Published 2019-05-29)
Active Learning in the Overparameterized and Interpolating Regime
arXiv:1206.4647 [cs.LG] (Published 2012-06-18)
Active Learning for Matching Problems