Local Regression Transfer Learning for Users’ Personality Prediction
Some research has been done to predict users’ personality based on their web behaviors. They usually use supervised learning methods to model on training dataset and predict on test dataset. However, when training dataset has different distributions from test dataset, which doesn’t meet independently identical distribution condition, traditional supervised learning models may perform not well on test dataset. Thus, we introduce a new regression transfer learning framework to deal with this problem, and propose two local regression instance-transfer methods. We use clustering and k-nearest-neighbor to reweight importance of each training instance to adapt to test dataset distribution, and then train a weighted risk regression model for prediction. We perform experiments on the condition that users dataset are from different genders and from different districts, and the results indicate that our methods can reduce mean square error about 30% to the most compared with non-transfer methods and be better than other transfer method in the whole.
KeywordsLocal Regression Transfer Learning Importance Reweighting Personality Prediction
Unable to display preview. Download preview PDF.
- 1.Burger, J.: Personality, 7th edn. Thomson Higher Education, Belmont (2008)Google Scholar
- 3.Li, L., Li, A., Hao, B., Guan, Z., Zhu, T.: Predicting active users personality based on micro-blogging behaviors. PLoS ONE 9(1) (2014)Google Scholar
- 4.Pan, S., Yang, Q.: A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering 99, 1041–4347 (2009)Google Scholar
- 5.Huang, J., Smola, A., Gretton, A., Borgwardt, K., Scholkopf, B.: Correcting sample selection bias by unlabeled data. In: Proc. 19th Ann. Conf. Neural Information Processing Systems (2007)Google Scholar
- 6.Sugiyama, M., Nakajima, S., Kashima, H., Buenau, P., Kawanabe, M.: Direct importance estimation with model selection and its application to covariate shift adaptation. In: Proc. 20th Ann. Conf. Neural Information Processing Systems (2008)Google Scholar
- 7.Kanamori, T., Hido, S., Sugiyama, M.: Efficient direct density ratio estimation for non-stationarity adaptation and outlier detection. Advances in Neural Information Processing Systems 20, 809–816 (2008)Google Scholar
- 8.Loog, M.: Nearest neighbor-based importance weighting. In: IEEE International Workshop on Machine Learning for Signal Processing, Santander, Spain (2012)Google Scholar
- 9.Pardoe, D., Stone, P.: Boosting for regression transfer. In: Proceedings of the 27th International Conference on Machine Learning (2010)Google Scholar
- 10.Dai, W., Yang, Q., Xue, G., Yu, Y.: Boosting for transfer learning. In: Proceedings of the 24th International Conference on Machine Learning (2007)Google Scholar
- 14.Hastie, T., Tibshirani, R., Friedman, J.: The Elements of Statistical Learning, 2nd edn. Springer (2008)Google Scholar