Binary Classification under Sample Selection Bias
Binary Classification under Sample Selection Bias
This chapter examines the problem of binary classification under sample selection bias from a decision-theoretic perspective. Starting from a derivation of the necessary and sufficient conditions for equivalence of the Bayes classifiers of training and test distributions, it provides the conditions under which sample selection bias does not affect the performance of a classifier. From this viewpoint, there are fundamental differences between classifiers of low and high capacity, in particular the ones that are Bayes consistent. The second part of the chapter provides means to modify existing learning algorithms such that they are more robust to sample selection bias in the case where one has access to an unlabeled sample of the test data. This is achieved by constructing a graph-based regularization functional. The close connection of this approach to semisupervised learning is also highlighted.
Keywords: Bayes classifiers, sample selection bias, learning algorithms, regularization functional
MIT Press Scholarship Online requires a subscription or purchase to access the full text of books within the service. Public users can however freely search the site and view the abstracts and keywords for each book and chapter.
Please, subscribe or login to access full text content.
If you think you should have access to this title, please contact your librarian.
To troubleshoot, please check our FAQs, and if you can't find the answer there, please contact us.