Authors
Shai Ben-David, John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, Jennifer Wortman Vaughan
Publication date
2010/5
Journal
Machine learning
Volume
79
Pages
151-175
Publisher
Springer US
Description
Discriminative learning methods for classification perform well when training and test data are drawn from the same distribution. Often, however, we have plentiful labeled training data from a source domain but wish to learn a classifier which performs well on a target domain with a different distribution and little or no labeled training data. In this work we investigate two questions. First, under what conditions can a classifier trained from source data be expected to perform well on target data? Second, given a small amount of labeled target data, how should we combine it during training with the large amount of labeled source data to achieve the lowest target error at test time?
We address the first question by bounding a classifier’s target error in terms of its source error and the divergence between the two domains. We give a classifier-induced divergence measure that can be estimated from finite …
Total citations
20112012201320142015201620172018201920202021202220232024313654495789126169297476610665685294
Scholar articles
S Ben-David, J Blitzer, K Crammer, A Kulesza… - Machine learning, 2010