Most active learning research has focused on methods which perform well when many labels are available, but can be dramatically worse than random selection when label budgets are small. Other methods have focused on the low-budget regime, but do poorly as label budgets increase. As the line between “low” and “high” budgets varies by problem, this is a serious issue in practice. We propose uncertainty coverage, an objective which generalizes a variety of low- and high-budget objectives, as well as natural, hyperparameter-light methods to smoothly interpolate between low- and high-budget regimes. We call greedy optimization of the estimate Uncertainty Herding; this simple method is computationally fast, and we prove that it nearly optimizes the distribution-level coverage. In experimental validation across a variety of active learning tasks, our proposal matches or beats state-of-the-art performance in essentially all cases; it is the only method of which we are aware that reliably works well in both low- and high-budget settings.
Bibtex
@misc{bae2024uncertaintyherdingactivelearning,
title={Uncertainty Herding: One Active Learning Method for All Label Budgets},
author={Wonho Bae and Gabriel L. Oliveira and Danica J. Sutherland},
year={2024},
eprint={2412.20644},
archivePrefix={arXiv},
primaryClass={cs.LG},
url={https://arxiv.org/abs/2412.20644},
}
Related Research
-
Training foundation models up to 10x more efficiently with Memory-Mapped Datasets
Training foundation models up to 10x more efficiently with Memory-Mapped Datasets
T. Badamdorj, and M. Anand.
Research
-
DeepRRTime: Robust Time-series Forecasting with a Regularized INR Basis
DeepRRTime: Robust Time-series Forecasting with a Regularized INR Basis
C.S. Sastry, M. Gilany, K. Y. C. Lui, M. Magill, and A. Pashevich. Transactions on Machine Learning Research (TMLR)
Publications
-
Radar: Fast Long-Context Decoding for Any Transformer
Radar: Fast Long-Context Decoding for Any Transformer
Y. Hao, M. Zhai, H. Hajimirsadeghi, S. Hosseini, and F. Tung. International Conference on Learning Representations (ICLR)
Publications