CUED Publications database

Bayesian batch active learning as sparse subset approximation

Pinsler, R and Gordon, J and Nalisnick, E and Hernández-Lobato, JM (2019) Bayesian batch active learning as sparse subset approximation. In: UNSPECIFIED.

Full text not available from this repository.


© 2019 Neural information processing systems foundation. All rights reserved. Leveraging the wealth of unlabeled data produced in recent years provides great potential for improving supervised models. When the cost of acquiring labels is high, probabilistic active learning methods can be used to greedily select the most informative data points to be labeled. However, for many large-scale problems standard greedy procedures become computationally infeasible and suffer from negligible model change. In this paper, we introduce a novel Bayesian batch active learning approach that mitigates these issues. Our approach is motivated by approximating the complete data posterior of the model parameters. While naive batch construction methods result in correlated queries, our algorithm produces diverse batches that enable efficient active learning at scale. We derive interpretable closed-form solutions akin to existing active learning procedures for linear models, and generalize to arbitrary models using random projections. We demonstrate the benefits of our approach on several large-scale regression and classification tasks.

Item Type: Conference or Workshop Item (UNSPECIFIED)
Uncontrolled Keywords: stat.ML stat.ML cs.LG
Divisions: Div F > Computational and Biological Learning
Depositing User: Cron Job
Date Deposited: 01 Oct 2019 03:38
Last Modified: 17 Sep 2020 02:57