arXiv Analytics

Sign in

arXiv:2103.08493 [cs.LG]AbstractReferencesReviewsResources

How Many Data Points is a Prompt Worth?

Teven Le Scao, Alexander M. Rush

Published 2021-03-15Version 1

When fine-tuning pretrained models for classification, researchers either use a generic model head or a task-specific prompt for prediction. Proponents of prompting have argued that prompts provide a method for injecting task-specific guidance, which is beneficial in low-data regimes. We aim to quantify this benefit through rigorous testing of prompts in a fair setting: comparing prompted and head-based fine-tuning in equal conditions across many tasks and data sizes. By controlling for many sources of advantage, we find that prompting does indeed provide a benefit, and that this benefit can be quantified per task. Results show that prompting is often worth 100s of data points on average across classification tasks.

Comments: NAACL HLT 2021
Categories: cs.LG
Related articles: Most relevant | Search more
arXiv:1902.00033 [cs.LG] (Published 2019-01-31)
Compressed Diffusion
arXiv:1802.03936 [cs.LG] (Published 2018-02-12)
On the Needs for Rotations in Hypercubic Quantization Hashing
arXiv:2106.15662 [cs.LG] (Published 2021-06-29)
Exponential Weights Algorithms for Selective Learning