Skip to main content

Zero-Inflated Poisson Tensor Factorization for Sparse Purchase Data in E-Commerce Markets

  • Conference paper
  • First Online:
Industrial Engineering and Applications – Europe (ICIEA-EU 2024)

Abstract

Nonnegative tensor factorization (NTF) plays a crucial role in extracting latent factors and predicting future sales from purchase data consisting of user and item attributes. However, the increase in these attributes leads to tensor data becoming sparse, causing a reduction in decomposition accuracy. For example, when there are numerous combinations of unavailable item genres and prices, the purchase history data becomes sparse and follows a distribution where all its elements are zero. To address this issue, we propose a novel NTF method assuming zero-inflated Poisson (ZIP) distribution based on Expectation-Maximization (EM) algorithm. This enables us to effectively handle sparsity in high-dimensional multiway data and identify combinations of user and item attributes that are potentially not likely to be purchased. We verified the effectiveness of the proposed approach through numerical experiments using real-world e-commerce data. The results showed our proposed ZIP model outperforms existing methods in both in-sample and out-of-sample experiments. Moreover, the proposed method qualitatively demonstrated the effectiveness of handling sparsity.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 64.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://github.com/tokyotech-nakatalab/tensor-decomposition.git.

  2. 2.

    https://www.rakuten.com/.

References

  1. Lee, D., Seung, H.S.: Algorithms for non-negative matrix factorization. In: Advances in Neural Information Processing Systems, vol. 13. MIT Press (2000)

    Google Scholar 

  2. Rafailidis, D., Daras, P.: The TFC model: tensor factorization and tag clustering for item recommendation in social tagging systems. IEEE Trans. Syst., Man Cybern.: Syst. 43(3), 673–688 (2012)

    Article  Google Scholar 

  3. Taneja, A., Arora, A.: Cross domain recommendation using multidimensional tensor factorization. Expert Syst. Appl. 92, 304–316 (2018)

    Article  Google Scholar 

  4. Yin, H., et al.: SPTF: a scalable probabilistic tensor factorization model for semantic-aware behavior prediction. In: 2017 IEEE International Conference on Data Mining (ICDM), pp. 585–594. IEEE (2017)

    Google Scholar 

  5. Chou, S.-Y., Jang, J.-S.R., Yang, Y.-H.: Fast tensor factorization for large-scale context-aware recommendation from implicit feedback. IEEE Trans. Big Data 6(1), 201–208 (2018)

    Article  Google Scholar 

  6. Narita, A., Hayashi, K., Tomioka, R., Kashima, H.: Tensor factorization using auxiliary information. Data Min. Knowl. Disc. 25, 298–324 (2012)

    Article  MathSciNet  Google Scholar 

  7. Zhou, P., Lu, C., Lin, Z., Zhang, C.: Tensor factorization for low-rank tensor completion. IEEE Trans. Image Process. 27(3), 1152–1163 (2017)

    Article  MathSciNet  Google Scholar 

  8. Lambert, D.: Zero-inflated Poisson regression, with an application to defects in manufacturing. Technometrics 34(1), 1–14 (1992)

    Article  Google Scholar 

  9. Coxe, S., West, S.G., Aiken, L.S.: The analysis of count data: a gentle introduction to Poisson regression and its alternatives. J. Pers. Assess. 91(2), 121–136 (2009)

    Article  Google Scholar 

  10. Hiroyasu, A., Hiroshi, Y.: A non-negative matrix factorization model based on the zero-inflated Tweedie distribution. Comput. Statistics 32(2), 475–499 (2017)

    Article  MathSciNet  Google Scholar 

  11. Cichocki, A., Rafal, Z., Amari, S.: Nonnegative matrix and tensor factorization [lecture notes]. IEEE Signal Process. Mag. 25(1), 142–145 (2008)

    Article  Google Scholar 

  12. Mørup, M.: Applications of tensor (multiway array) factorizations and decompositions in data mining. Wiley Interdisc. Rev.: Data Min. Knowl. Discovery 1(1), 24–40 (2011)

    Google Scholar 

  13. Chen, Y., He, W., Yokoya, N., Huang, T.-Z.: Hyperspectral image restoration using weighted group sparsity-regularized low-rank tensor decomposition. IEEE Trans. Cybern. 50(8), 3556–3570 (2020)

    Article  Google Scholar 

  14. Li, S., Dian, R., Fang, L., Bioucas-Dias, J.M.: Fusing hyperspectral and multispectral images via coupled sparse tensor factorization. IEEE Trans. Image Process. 27(8), 4118–4130 (2018)

    Article  MathSciNet  Google Scholar 

  15. Mitsufuji, Y., Roebel, A.: Sound source separation based on non-negative tensor factorization incorporating spatial cue as prior knowledge. In: 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 71–75 (2013)

    Google Scholar 

  16. Gligorijević, V., Panagakis, Y., Zafeiriou, S.: Non-negative matrix factorizations for multiplex network analysis. IEEE Trans. Pattern Anal. Mach. Intell. 41(4), 928–940 (2018)

    Article  Google Scholar 

  17. Jain, P., Oh, S.: Provable tensor factorization with missing data. In: Advances in Neural Information Processing Systems, vol. 27. Curran Associates, Inc., (2014)

    Google Scholar 

  18. Lopez, O., Lehoucq, R., Dunlavy, D.: Zero-truncated Poisson tensor decomposition for sparse count data. Sandia National Lab.(SNL-NM), Albuquerque, NM (United States), Technical Report (2022)

    Google Scholar 

  19. Yu, H.F., Rao, N., Dhillon, I.S.: Temporal regularized matrix factorization for high-dimensional time series prediction. In: Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., (2016)

    Google Scholar 

  20. Dempster, A.P., Laird, N.M., Rubin, D.B.: Maximum likelihood from incomplete data via the EM algorithm. J. Roy. Stat. Soc.: Ser. B (Methodological) 39(10), 1–22 (1977)

    MathSciNet  Google Scholar 

  21. Kossaifi, J., Panagakis, Y., Anandkumar, A., Pantic, M.: TensorLy: tensor learning in python. J. Mach. Learn. Res. 20, 1–6 (2016)

    Google Scholar 

  22. Daly, F., Gaunt, R.E.: The Conway-Maxwell-Poisson distribution: distributional theory and approximation. ALEA 13, 635–658 (2016)

    Article  MathSciNet  Google Scholar 

  23. Gladys, F.L., Barriga, D.C.: The zero-inflated Conway-Maxwell-Poisson distribution: Bayesian inference, regression modeling and influence diagnostic. Stat. Methodol. 21, 23–34 (2014)

    Article  MathSciNet  Google Scholar 

  24. Shmueli, G., Minka, T.P., Kadane, J.B., Borle, S., Boatwright, P.: A useful distribution for fitting discrete data: revival of the Conway-Maxwell-Poisson distribution. J. Roy. Stat. Soc.: Ser. C: Appl. Stat. 54(1), 127–142 (2005)

    Article  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ken Kobayashi .

Editor information

Editors and Affiliations

Appendices

A Computational Time

We compare the seconds per step of the above experiments to evaluate computational time. Figure 5 (a) shows the average computational time of 10 trial runs for the four factorization models. The initial values are uniformly sampled across all models for each trial. Due to the complexity of parameters and update algorithms, our proposed method resulted in the slowest computational time.

Fig. 5.
figure 5

Computational time per step; (a) Each tensor factorization model. (b) Backends of implementation.

Additionally, we compared the computational time of implementations using our JAX + JIT compiler method with Tensorly in Fig. 5 (b). Both of these methods assume a normal distribution on the tensor and use the same algorithm. Our proposed algorithm significantly mitigates the increase in computational time with the increase in rank compared to the alternative.

B ZICMP Distribution Model

We demonstrate the Zero-Inflated Conway-Maxwell-Poisson (ZICMP) distribution tensor factorization model, which is a modified version of the Conway-Maxwell-Poisson (CMP) distribution in place of the Poisson distribution [22,23,24]. This enables to represent over and under dispersion which was not able by the ZIP distribution. When random variable y follows this distribution, the probability mass function is defined as follows:

$$\begin{aligned} f(y;\lambda , \nu , p) &= {\left\{ \begin{array}{ll}p+(1-p) \frac{1}{Z(\lambda , \nu )} &{} \text{ if } y=0 , \\ (1-p) \frac{\lambda ^y}{(y !)^\nu } \frac{1}{Z(\lambda , \nu )} &{} \text {otherwise} , \end{array}\right. } \quad \text {where}~Z(\lambda , \nu )&=\sum _{h=0}^{\infty } \frac{\lambda ^h}{(h !)^\nu }. \end{aligned}$$

p is the mixture ratio and \(\lambda \) is the expected value of the ZICMP distribution. \(\nu \) is a nonnegative real number and controls the variance. For \(\nu = 1\), its distribution aligns with the Poisson distribution.

Table 4. Evaluation metrics for each algorithm

In contrast to the ZIP model, here we update \(\boldsymbol{A}^{(1)},\dots , \boldsymbol{A}^{(n)}\) and \(\nu \) using Newton’s method at each step of factor matrices updating phase in Algorithm 1. A magnitude of Newton step can be derived by taking the first and second derivatives of the log-likelihood function with respect to \(a^{(k)}_{i_kr}\) and \(\nu \).

$$\begin{aligned} \varDelta a^{(k)}_{i_{k}r} &= \frac{- \displaystyle \sum _{i_1,\dots ,i_{k-1},i_{k+1},\dots ,i_{n}}\left( 1-z_{i_1 \dots i_n} \right) \left( -\frac{Z^{\prime }}{Z}+\frac{y_{i_1 \dots i_n}}{\lambda }\right) \prod _{\begin{array}{c} l=1 \\ l \ne k \end{array}}^n a^{(l)}_{i_l r}}{\displaystyle \sum _{i_1,\dots ,i_{k-1},i_{k+1},\dots ,i_{n}}\left( 1-z_{i_1 \dots i_n} \right) \Biggl (-\frac{Z^{\prime \prime }}{Z}+\left( \frac{Z^{\prime }}{Z}\right) ^{2}-\frac{y_{i_1 \dots i_n}}{\lambda ^2} \Biggl ) \prod _{\begin{array}{c} l=1 \\ l \ne k \end{array}}^n (a^{(l)}_{i_l r})^2}, \\ \varDelta \nu & =- \frac{\displaystyle \sum _{i_1 \dots i_n}\left( 1-z_{i_1 \dots i_n} \right) \Biggl (-\frac{\frac{\partial Z}{\partial \nu }}{Z}-\log \left( y_{i_1 \dots i_n} !\right) \Biggl )}{\displaystyle \sum _{i_1 \dots i_n}\left( 1-z_{i_1 \dots i_n} \right) \Biggl (-\frac{\frac{\partial ^2 Z}{{\partial }{\nu ^2}}}{Z}+\left( \frac{\frac{\partial Z}{\partial \nu }}{Z}\right) ^2\Biggl )}, \\ \text {where} &~ Z^{\prime } = \sum _{h=0}^{\infty } \frac{h \lambda ^{h-1}}{(h !)^\nu }, ~Z^{\prime \prime }= \sum _{h=0}^{\infty } \frac{h(h-1) \lambda ^{h-2}}{(h !)^\nu }, ~\lambda = \sum _r \prod _{l = 1}^n a^{(l)}_{i_l r}. \end{aligned}$$

We present the results of the numerical experiments for tensor factorization assuming the ZICMP distribution in Table 4. Our algorithm encountered overflow with large tensor sizes or element values. To address this, we reduced the tensor size to a 5th-mode tensor by eliminating the region mode.

As a result, RMSPE outperformed the other algorithms, but not in other evaluation metrics. While it can be assumed that the ZICMP distribution improved the expressiveness of the distribution, the fitting to sparsity did not work.

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Mizutani, K. et al. (2024). Zero-Inflated Poisson Tensor Factorization for Sparse Purchase Data in E-Commerce Markets. In: Sheu, SH. (eds) Industrial Engineering and Applications – Europe. ICIEA-EU 2024. Lecture Notes in Business Information Processing, vol 507. Springer, Cham. https://doi.org/10.1007/978-3-031-58113-7_14

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-58113-7_14

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-58112-0

  • Online ISBN: 978-3-031-58113-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics