Authors
John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, Jennifer Wortman
Publication date
2007
Conference
Advances in neural information processing systems
Pages
129-136
Description
Empirical risk minimization offers well-known learning guarantees when training and test data come from the same domain. In the real world, though, we often wish to adapt a classifier from a source domain with a large amount of training data to different target domain with very little training data. In this work we give uniform convergence bounds for algorithms that minimize a convex combination of source and target empirical risk. The bounds explicitly model the inherent trade-off between training on a large but inaccurate source data set and a small but accurate target training set. Our theory also gives results when we have multiple source domains, each of which may have a different number of instances, and we exhibit cases in which minimizing a non-uniform combination of source risks can achieve much lower target error than standard empirical risk minimization.
Total citations
200820092010201120122013201420152016201720182019202020212022202320241517312323302530242624455464575342
Scholar articles
J Blitzer, K Crammer, A Kulesza, F Pereira, J Wortman - Advances in neural information processing systems, 2007