A Practical Approach to Reduce the Learning Bias Under Covariate Shift

  • Van-Tinh TranEmail author
  • Alex Aussem
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 9285)


Covariate shift is a specific class of selection bias that arises when the marginal distributions of the input features X are different in the source and the target domains while the conditional distributions of the target Y given X are the same. A common technique to deal with this problem, called importance weighting, amounts to reweighting the training instances in order to make them resemble the test distribution. However this usually comes at the expense of a reduction of the effective sample size. In this paper, we show analytically that, while the unweighted model is globally more biased than the weighted one, it may locally be less biased on low importance instances. In view of this result, we then discuss a manner to optimally combine the weighted and the unweighted models in order to improve the predictive performance in the target domain. We conduct a series of experiments on synthetic and real-world data to demonstrate the efficiency of this approach.


Mean Square Error Weighting Scheme Target Domain Importance Weighting Weighted Model 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Bickel, S., Brückner, M., Scheffer, T.: Discriminative learning under covariate shift. The Journal of Machine Learning Research 10, 2137–2155 (2009)Google Scholar
  2. 2.
    Cortes, C., Mansour, Y., Mohri, M.: Learning bounds for importance weighting. In: Lafferty, J., Williams, C., Shawe-Taylor, J., Zemel, R., Culotta, A. (eds.) Advances in Neural Information Processing Systems, vol. 23, pp. 442–450. Curran Associates, Inc. (2010)Google Scholar
  3. 3.
    Gretton, A., Smola, A., Huang, J., Schmittfull, M., Borgwardt, K., Schölkopf, B.: Covariate shift by kernel mean matching (2009)Google Scholar
  4. 4.
    Hernán, M.A., Hernández-Díaz, S., Robins, J.M.: A structural approach to selection bias. Epidemiology 15(5), 615–625 (2004)CrossRefGoogle Scholar
  5. 5.
    Huang, J., Smola, A.J., Gretton, A., Borgwardt, K.M., Schölkopf, B.: Correcting sample selection bias by unlabeled data. In: Schölkopf, B., Platt, J., Hoffman, T. (eds.) NIPS, pp. 601–608. MIT Press (2006)Google Scholar
  6. 6.
    Kanamori, T., Hido, S., Sugiyama, M.: A least-squares approach to direct importance estimation. J. Mach. Learn. Res. 10, 1391–1445 (2009)MathSciNetzbMATHGoogle Scholar
  7. 7.
    Kanamori, T., Suzuki, T., Sugiyama, M.: Statistical analysis of kernel-based least-squares density-ratio estimation. Machine Learning 86(3), 335–367 (2012)MathSciNetCrossRefGoogle Scholar
  8. 8.
    Moreno-Torres, J.G., Raeder, T., Alaiz-Rodríguez, R., Chawla, N.V., Herrera, F.: A unifying view on dataset shift in classification. Pattern Recognition 45(1), 521–530 (2012)CrossRefGoogle Scholar
  9. 9.
    Nguyen, X., Wainwright, M.J., Jordan, M.I.: Estimating divergence functionals and the likelihood ratio by convex risk minimization. IEEE Transactions on Information Theory 56(11), 5847–5861 (2010)MathSciNetCrossRefzbMATHGoogle Scholar
  10. 10.
    Shimodaira, H.: Improving predictive inference under covariate shift by weighting the log-likelihood function. Journal of Statistical Planning and Inference 90(2), 227–244 (2000)MathSciNetCrossRefzbMATHGoogle Scholar
  11. 11.
    Sugiyama, M., Nakajima, S., Kashima, H., von Bünau, P., Kawanabe, M.: Direct importance estimation with model selection and its application to covariate shift adaptation. In: Platt, J., Koller, D., Singer, Y., Roweis, S. (eds.) NIPS (2007)Google Scholar
  12. 12.
    Wen, J., Yu, C.-N., Greiner, R.: Robust learning under uncertain test distributions: relating covariate shift to model misspecification. In: Jebara, T., Xing, E.P. (eds.) Proceedings of the 31st International Conference on Machine Learning (ICML 2014), pp. 631–639 (2014)Google Scholar
  13. 13.
    Zadrozny, B.: Learning and evaluating classifiers under sample selection bias. In: Greiner, R., Schuurmans, D. (eds.) Proceedings of the 21st International Conference on Machine Learning (ICML 2004) (2004)Google Scholar

Copyright information

© Springer International Publishing Switzerland 2015

Authors and Affiliations

  1. 1.LIRIS, UMR 5205, University of Lyon 1LyonFrance

Personalised recommendations