Abstract
A lot of efforts have been put in recent times for research in the field of natural language processing. Extracting topics is undoubtedly one of the most important tasks in this area of research. Latent Dirichlet allocation (LDA) is a widely used model that can perform this task in an unsupervised manner efficiently. It has been proved recently that using priors other than Dirichlet can be advantageous in extracting better quality topics from the data. Hence, in our paper we introduce the interactive latent generalized Dirichlet allocation model to extract topics. The model infers better topics using little information provided by the users through interactive learning. We use a variational algorithm for efficient inference. The model is validated against text datasets based on extracting topics related to news categories and types of emotions to test its efficiency.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bakhtiari, A.S., Bouguila, N.: A variational Bayes model for count data learning and classification. Eng. Appl. Artif. Intell. 35, 176–186 (2014)
Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent dirichlet allocation. J. Mach. Learn. Res. 3, 993–1022 (2003)
Bouguila, N.: Deriving kernels from generalized dirichlet mixture models and applications. Inf. Process. Manage. 49(1), 123–137 (2013)
Chien, J.T., Lee, C.H., Tan, Z.H.: Latent dirichlet mixture model. Neurocomputing 278, 12–22 (2018)
Fan, W., Bouguila, N.: Online variational learning of generalized dirichlet mixture models with feature selection. Neurocomputing 126, 166–179 (2014)
Fan, W., Bouguila, N.: Topic novelty detection using infinite variational inverted dirichlet mixture models. In: 2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA), pp. 70–75. IEEE (2015)
Fan, W., Bouguila, N., Ziou, D.: Variational learning for finite dirichlet mixture models and applications. IEEE Trans. Neural Netw. Learn. Syst. 23(5), 762–774 (2012)
Liu, Y., Du, F., Sun, J., Jiang, Y.: ILDA: an interactive latent dirichlet allocation model to improve topic quality. J. Inf. Sci. 46(1), 23–40 (2020)
Mcauliffe, J., Blei, D.: Supervised topic models. In: 20th Proceedings of the Conference on Advances in Neural Information Processing Systems (2007)
Mimno, D., Wallach, H.M., Talley, E., Leenders, M., McCallum, A.: Optimizing semantic coherence in topic models. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 262–272. EMNLP ’11, Association for Computational Linguistics, USA (2011)
Opper, M., Saad, D.: Mean-field theory of learning: from dynamics to statics (2001)
Saravia, E., Liu, H.C.T., Huang, Y.H., Wu, J., Chen, Y.S.: CARER: contextualized affect representations for emotion recognition. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 3687–3697. Association for Computational Linguistics, Brussels, Belgium, October–November 2018
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Maanicshah, K., Amayri, M., Bouguila, N. (2022). Interactive Generalized Dirichlet Mixture Allocation Model. In: Krzyzak, A., Suen, C.Y., Torsello, A., Nobile, N. (eds) Structural, Syntactic, and Statistical Pattern Recognition. S+SSPR 2022. Lecture Notes in Computer Science, vol 13813. Springer, Cham. https://doi.org/10.1007/978-3-031-23028-8_4
Download citation
DOI: https://doi.org/10.1007/978-3-031-23028-8_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-23027-1
Online ISBN: 978-3-031-23028-8
eBook Packages: Computer ScienceComputer Science (R0)