Skip to main content

Interactive Generalized Dirichlet Mixture Allocation Model

  • Conference paper
  • First Online:
Structural, Syntactic, and Statistical Pattern Recognition (S+SSPR 2022)

Abstract

A lot of efforts have been put in recent times for research in the field of natural language processing. Extracting topics is undoubtedly one of the most important tasks in this area of research. Latent Dirichlet allocation (LDA) is a widely used model that can perform this task in an unsupervised manner efficiently. It has been proved recently that using priors other than Dirichlet can be advantageous in extracting better quality topics from the data. Hence, in our paper we introduce the interactive latent generalized Dirichlet allocation model to extract topics. The model infers better topics using little information provided by the users through interactive learning. We use a variational algorithm for efficient inference. The model is validated against text datasets based on extracting topics related to news categories and types of emotions to test its efficiency.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 89.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 119.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    http://mlg.ucd.ie/datasets/bbc.html.

References

  1. Bakhtiari, A.S., Bouguila, N.: A variational Bayes model for count data learning and classification. Eng. Appl. Artif. Intell. 35, 176–186 (2014)

    Article  Google Scholar 

  2. Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent dirichlet allocation. J. Mach. Learn. Res. 3, 993–1022 (2003)

    MATH  Google Scholar 

  3. Bouguila, N.: Deriving kernels from generalized dirichlet mixture models and applications. Inf. Process. Manage. 49(1), 123–137 (2013)

    Article  Google Scholar 

  4. Chien, J.T., Lee, C.H., Tan, Z.H.: Latent dirichlet mixture model. Neurocomputing 278, 12–22 (2018)

    Article  Google Scholar 

  5. Fan, W., Bouguila, N.: Online variational learning of generalized dirichlet mixture models with feature selection. Neurocomputing 126, 166–179 (2014)

    Article  Google Scholar 

  6. Fan, W., Bouguila, N.: Topic novelty detection using infinite variational inverted dirichlet mixture models. In: 2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA), pp. 70–75. IEEE (2015)

    Google Scholar 

  7. Fan, W., Bouguila, N., Ziou, D.: Variational learning for finite dirichlet mixture models and applications. IEEE Trans. Neural Netw. Learn. Syst. 23(5), 762–774 (2012)

    Article  Google Scholar 

  8. Liu, Y., Du, F., Sun, J., Jiang, Y.: ILDA: an interactive latent dirichlet allocation model to improve topic quality. J. Inf. Sci. 46(1), 23–40 (2020)

    Article  Google Scholar 

  9. Mcauliffe, J., Blei, D.: Supervised topic models. In: 20th Proceedings of the Conference on Advances in Neural Information Processing Systems (2007)

    Google Scholar 

  10. Mimno, D., Wallach, H.M., Talley, E., Leenders, M., McCallum, A.: Optimizing semantic coherence in topic models. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 262–272. EMNLP ’11, Association for Computational Linguistics, USA (2011)

    Google Scholar 

  11. Opper, M., Saad, D.: Mean-field theory of learning: from dynamics to statics (2001)

    Google Scholar 

  12. Saravia, E., Liu, H.C.T., Huang, Y.H., Wu, J., Chen, Y.S.: CARER: contextualized affect representations for emotion recognition. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 3687–3697. Association for Computational Linguistics, Brussels, Belgium, October–November 2018

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kamal Maanicshah .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Maanicshah, K., Amayri, M., Bouguila, N. (2022). Interactive Generalized Dirichlet Mixture Allocation Model. In: Krzyzak, A., Suen, C.Y., Torsello, A., Nobile, N. (eds) Structural, Syntactic, and Statistical Pattern Recognition. S+SSPR 2022. Lecture Notes in Computer Science, vol 13813. Springer, Cham. https://doi.org/10.1007/978-3-031-23028-8_4

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-23028-8_4

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-23027-1

  • Online ISBN: 978-3-031-23028-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics