Exploring Heterogeneous Product Networks for Discovering Collective Marketing Hyping Behavior
Online spam comments often misguide users during online shopping. Existing online spam detection methods rely on semantic clues, behavioral footprints, and relational connections between users in review systems. Although these methods can successfully identify spam activities, evolving fraud strategies can successfully escape from the detection rules by purchasing positive comments from massive random users, i.e., user Cloud. In this paper, we study a new problem, Collective Marketing Hyping detection, for spam comments detection generated from the user Cloud. It is defined as detecting a group of marketing hyping products with untrustful marketing promotion behaviour. We propose a new learning model that uses heterogenous product networks extracted from product review systems. Our model aims to mining a group of hyping activities, which differs from existing models that only detect a single product with hyping activities. We show the existence of the Collective Marketing Hyping behavior in real-life networks. Experimental results demonstrate that the product information network can effectively detect fraud intentional product promotions.
KeywordsRegularization Term Product Network User Cloud Online Store Spam Detection
This work was supported by Australia ARC Discovery Project (DP140102206) and Australia Linkage Project (LP150100671).
- 1.Fei, G., Mukherjee, A., Liu, B., Hsu, M., Castellanos, M., Ghosh, R.: Exploiting burstiness in reviews for review spammer detection. ICWSM 13, 175–184 (2013)Google Scholar
- 2.Feng, S., Xing, L., Gogar, A., Choi, Y.: Distributional footprints of deceptive product reviews (2012)Google Scholar
- 3.Grabocka, J., Schilling, N., Wistuba, M., Schmidt-Thieme, L.: Learning time-series shapelets. In: KDD 2014, pp. 392–401. ACM (2014)Google Scholar
- 4.Jindal, N., Liu, B.: Review spam detection. In: WWW 2007, pp. 1189–1190. ACM (2007)Google Scholar
- 5.Jindal, N., Liu, B.: Opinion spam and analysis. In: WSDM 2008, pp. 219–230. ACM (2008)Google Scholar
- 6.Jindal, N., Liu, B., Lim, E.-P.: Finding unusual review patterns using unexpected rules, pp. 1549–1552. ACM (2010)Google Scholar
- 7.Li, F., Huang, M., Yang, Y., Zhu, X.: Learning to identify review spam. In: IJCAI 2011, pp. 2488–2493. AAAI Press (2011)Google Scholar
- 8.Lines, J., Davis, L.M., Hills, J., Bagnall, A.: A shapelet transform for time series classification. In: KDD 2012, pp. 289–297. ACM (2012)Google Scholar
- 9.McAuley, J., Leskovec, J.: Hidden factors and hidden topics: understanding rating dimensions with review text. In: RecSys 2013, pp. 165–172. ACM (2013)Google Scholar
- 10.McAuley, J.J., Leskovec, J., Jurafsky, D.: Learning attitudes and attributes from multi-aspect reviews, abs/1210.3926 (2012)Google Scholar
- 11.McAuley, J.J., Leskovec, J.: From amateurs to connoisseurs: Modeling the evolution of user expertise through online reviews. In: WWW 2013, pp. 897–908. International World Wide Web Conferences Steering Committee (2013)Google Scholar
- 12.Xie, S., Wang, G., Lin, S., Yu, P.S.: Review spam detection via temporal pattern discovery, pp. 823–831. ACM (2012)Google Scholar