Abstract
To realize trans-domain behavioral targeting, which targets interested potential users of a source domain (e.g. E-Commerce) based on their behaviors in a target domain (e.g. Ad-Network), heterogeneous transfer learning (HeTL) is a promising technique for modeling behavior linkage between domains. It is required for HeTL to learn three functionalities: representation alignment, distribution alignment, and classification. In our previous work, we prototyped and evaluated two typical transfer learning algorithms, but neither of them jointly learns the three desired functionalities. Recent advances in transfer learning include a domain-adversarial neural network (DANN), which jointly learns distribution alignment and classification. In this paper, we extended DANN to be able to learn representation alignment by simply replacing its shared encoder with domain-specific types, so that it jointly learns the three desired functionalities. We evaluated the effectiveness of the joint learning of the three functionalities using real-world data of two domains: E-Commerce, which is set as the source domain, and Ad Network, which is set as the target domain.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
The promise of first-party data. Econsultancy. https://cdn2.hubspot.net/hubfs/370829/Campaigns_and_Emails/Archived_Emails/2015_Emails/2015_Econsultancy_Promise_of_First_Party_Data_June/The_Promise_of_First_Party_Data_Signal_Econsultancy_Report.pdf. Accessed 15 Oct 2018
Shi, X., Liu, Q., Fan, W., Yang, Q., Yu, P.S.: Predictive modeling with heterogeneous sources. In: Proceedings of the 2010 SIAM International Conference on Data Mining, pp. 814–825 (2010)
Yamada, M., Suzuki, T., Kanamori, T., Hachiya, H., Sugiyama, M.: Relative density-ratio estimation for robust distribution comparison. In: Advances in Neural Information Processing Systems, vol. 24, pp. 594–602 (2011)
Ganin, Y., et al.: Domain-adversarial training of neural networks. J. Mach. Learn. Res. 17, 1–35 (2016)
Kulis, B., Saenko, K., Darrell, T.: What you saw is not what you get: domain adaptation using asymmetric kernel transforms. In: Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, pp. 1785–1792 (2011)
Feuz, K.D., Cook, D.J.: Transfer learning across feature-rich heterogeneous feature spaces via feature-space remapping (FSR). ACM Trans. Intell. Syst. Technol. 6, 1–27 (2015)
Shi, X., Liu, Q., Fan, W., Yu, P.S., Zhu, R.: Transfer learning on heterogenous feature spaces via spectral transformation. In: Proceedings of the 10th IEEE International Conference on Data Mining, pp. 1049–1054 (2010)
Duan, L., Tsang, I.W.: Learning with augmented features for heterogeneous domain. In: Proceedings of the 29th International Conference on Machine Learning (2012)
Li, W., Duan, L., Xu, D., Tsang, I.W.: Learning with augmented features for supervised and semi-supervised heterogeneous domain adaptation. IEEE Trans. Pattern Anal. Mach. Intell. 36, 1134–1148 (2014)
Chen, T., Guestrin, C.: Xgboost: a scalable tree boosting system. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data mining, pp. 785–794 (2016)
Rehurek, R., Sojka, P.: Software framework for topic modelling with large corpora. In: LREC 2010 Workshop on New Challenges for NLP Frameworks (2010)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Proceedings of the International Conference on Learning Representations (2014)
Kurokawa, M., et al.: Virtual touch-point: trans-domain behavioral targeting via transfer learning. In: Workshop on Big Data Transfer Learning in conjunction with IEEE International Conference on Big Data (2018)
Acknowledgment
This research was partially supported by JST CREST Grant Number J181401085, Japan.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendix
Appendix
-
PM1: HeMap + HEGS + Xgboost
Modeling Behavior Linkage.
HeMap learns a common latent space from the source and target features. The optimization objective related to the common latent space is as follows (the same equation appears in [7]):
where, \( {\text{B}}_{\text{S}} \), \( {\text{B}}_{\text{T}} \) are the projected source and target instances onto the common latent space respectively, and \( {\text{P}}_{\text{S}} \), \( {\text{P}}_{\text{T}} \) are the projection matrices from the common latent space onto the source and target space, respectively.
Then, HEGS selects the source instances in \( {\text{B}}_{\text{S}} \) similar to the target instances in \( {\text{B}}_{\text{T}} \). Since HEGS is originally a domain adaptation method for regression, we modified HEGS by hiring logistic regression to unify the labels in both domains. Finally, Xgboost learns the binary classification model \( F_{Xgb} \left( \cdot \right) \) using the selected instances and labels. Hyper-parameters of HeMap, HEGS, and Xgboost are tuned using cross validation. The resulting behavior linkage model \( {\mathbf{\mathcal{M}}} = \left( {{\text{P'}}_{\text{T}} ,F_{Xgb} } \right) \), where \( {\text{P'}}_{\text{T}} \) is a pseudo-inverse of the projection matrix \( {\text{P}}_{\text{T}} \) obtained by HeMap.
Applying Behavior Linkage Model.
We compute the predictive probability of conversion \( {\text{P}}\left( {y_{j} |\varvec{x}_{j}^{{\left( {\text{T}} \right)}} } \right) = F_{Xgb} \left( {{\text{P'}}_{\text{T}} \varvec{ x}_{j}^{{\left( {\text{T}} \right)}} } \right) \) for each user \( j \) in the target domain. The resulting target users are defined by \( \left\{ {j; {\text{P}}\left( {y_{j} |\varvec{x}_{j}^{{\left( {\text{T}} \right)}} } \right) > \theta } \right\} \) with an arbitrary threshold \( 0 < \theta < 1 \).
-
PM2: HFA
Modeling Behavior Linkage.
HFA learns a multiple kernel classifier \( F_{HFA} \left( \cdot \right) \) on a common latent space using the source and target features and the labels. The projection matrices \( {\text{P}}, {\text{Q}} \) from the source and target space onto the common latent space are coupled as \( {\text{H}} = \left[ {{\text{P}}, {\text{Q}}} \right] '\left[ {{\text{P}}, {\text{Q}}} \right] \), and then kernel matrices are computed and optimized based on \( {\text{H}} \). Hyper-parameters of HFA are tuned using cross validation. The resulting behavior linkage model \( {\mathbf{\mathcal{M}}} = \left( {F_{HFA} } \right) \).
Applying Behavior Linkage Model.
Similar to PM1, we compute the predictive probability of conversion \( {\text{P}}\left( {y_{j} |\varvec{x}_{j}^{{\left( {\text{T}} \right)}} } \right) = F_{HFA} \left( {\varvec{ x}_{j}^{{\left( {\text{T}} \right)}} } \right) \) for each user \( j \) in the target domain. The resulting target users are defined by \( \left\{ {j; {\text{P}}\left( {y_{j} |\varvec{x}_{j}^{{\left( {\text{T}} \right)}} } \right) > \theta } \right\} \) with an arbitrary threshold \( 0 < \theta < 1 \).
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Yonekawa, K. et al. (2019). A Heterogeneous Domain Adversarial Neural Network for Trans-Domain Behavioral Targeting. In: U., L., Lauw, H. (eds) Trends and Applications in Knowledge Discovery and Data Mining. PAKDD 2019. Lecture Notes in Computer Science(), vol 11607. Springer, Cham. https://doi.org/10.1007/978-3-030-26142-9_24
Download citation
DOI: https://doi.org/10.1007/978-3-030-26142-9_24
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-26141-2
Online ISBN: 978-3-030-26142-9
eBook Packages: Computer ScienceComputer Science (R0)