arXiv Analytics

Sign in

arXiv:1407.7644 [stat.ML]AbstractReferencesReviewsResources

Estimating the Accuracies of Multiple Classifiers Without Labeled Data

Ariel Jaffe, Boaz Nadler, Yuval Kluger

Published 2014-07-29, updated 2014-10-30Version 2

In various situations one is given only the predictions of multiple classifiers over a large unlabeled test data. This scenario raises the following questions: Without any labeled data and without any a-priori knowledge about the reliability of these different classifiers, is it possible to consistently and computationally efficiently estimate their accuracies? Furthermore, also in a completely unsupervised manner, can one construct a more accurate unsupervised ensemble classifier? In this paper, focusing on the binary case, we present simple, computationally efficient algorithms to solve these questions. Furthermore, under standard classifier independence assumptions, we prove our methods are consistent and study their asymptotic error. Our approach is spectral, based on the fact that the off-diagonal entries of the classifiers' covariance matrix and 3-d tensor are rank-one. We illustrate the competitive performance of our algorithms via extensive experiments on both artificial and real datasets.

Related articles: Most relevant | Search more
arXiv:1606.04316 [stat.ML] (Published 2016-06-14)
Time for a change: a tutorial for comparing multiple classifiers through Bayesian analysis
arXiv:2408.07796 [stat.ML] (Published 2024-08-14)
Ranking and Combining Latent Structured Predictive Scores without Labeled Data
arXiv:1409.7495 [stat.ML] (Published 2014-09-26)
Unsupervised Domain Adaptation by Backpropagation