Abstract
Business sentiment analysis (BSA) is one of the significant and popular topics of natural language processing. It is one kind of sentiment analysis techniques for business purpose. Different categories of sentiment analysis techniques like lexicon-based techniques and different types of machine learning algorithms are applied for sentiment analysis on different languages like English, Hindi, Spanish, etc. In this paper, long short-term memory (LSTM) is applied for business sentiment analysis, where recurrent neural network is used. LSTM model is used in a modified approach to prevent the vanishing gradient problem rather than applying the conventional recurrent neural network (RNN). To apply the modified RNN model, product review dataset is used. In this experiment, 70% of the data is trained for the LSTM and the rest 30% of the data is used for testing. The result of this modified RNN model is compared with other conventional RNN models and a comparison is made among the results. It is noted that the proposed model performs better than the other conventional RNN models. Here, the proposed model, i.e., modified RNN model approach has achieved around 91.33% of accuracy. By applying this model, any business company or e-commerce business site can identify the feedback from their customers about different types of product that customers like or dislike. Based on the customer reviews, a business company or e-commerce platform can evaluate its marketing strategy.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Vishwanathan, S.V.M., Murty, M.N.: SSVM: A simple SVM algorithm. In: Proceedings of the 2002 International Joint Conference on Neural Networks. IJCNN’02 (Cat. No. 02CH37290) vol. 3, pp. 2393–2398. IEEE (2002)
Wright, R.E.: Logistic regression (1995)
Zhang, H.: The optimality of Naive Bayes. AA, 1(2), 3 (2004)
Breiman, L.: Random Forests. Mach. Learn. 45(1), 5–32 (2001)
Agatonovic-Kustrin, S., Beresford, R.: Basic concepts of artificial neural network (ANN) modeling and its application in pharmaceutical research. J. Pharm. Biomed. Anal. 22(5), 717–727 (2000)
Mikolov, T., Karafiát, M., Burget, L., Černocký, J, Khudanpur, S.: Recurrent neural network based language model. In: Eleventh Annual Conference of the International Speech Communication Association (2010)
Huang, W., Qiao, Y., Tang, X.: Robust scene text detection with convolution neural network induced mser trees. In: European Conference on Computer Vision, pp. 497–511. Springer, Cham (2014)
Liang, J., Koperski, K., Dhillon, N.S., Tusk, C., Bhatti, S., Evri Inc.: NLP-based entity recognition and disambiguation. U.S. Patent 8,594,996 (2013)
Bhatia, V., Rani, R.: Dfuzzy: a deep learning-based fuzzy clustering model for large graphs. Knowl. Inf. Syst. 57(1), 159–181 (2018)
Schmidt, W.F., Kraaijveld, M.A., Duin, R.P.: Feed forward neural networks with random weights. In: International Conference on Pattern Recognition p. 1. IEEE Computer Society Press (1992)
Xu, Y., Wu, X., Wang, Q.: Sentiment analysis of yelp’s ratings based on text reviews (2014)
Rani, S., Kumar, P.: Deep learning based sentiment analysis using convolution neural network. Arabian J. Sci. Eng. 44(4), 3305–3314 (2019)
Mikolov, T., Yih, W.T., Zweig, G.: Linguistic regularities in continuous space word representations. In: Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 746–751 (2013)
Graves, A.: Long short-term memory. In: Supervised Sequence Labelling with Recurrent Neural Networks, pp. 37–45. Springer, Berlin, (2012)
Zhang, X., Lu, L., Lapata, M.: Top-down tree long short-term memory networks. arXiv preprint, arXiv:1511.00060 (2015)
Graves, A.: Supervised sequence labelling. In: Supervised Sequence Labelling with Recurrent Neural Networks, pp. 5–13. Springer, Berlin, (2012)
Soutner, D., Müller, L.: Application of LSTM neural networks in language modelling. In International Conference on Text, Speech and Dialogue, pp. 105–112. Springer, Berlin, (2013)
Sundermeyer, M., Tüske, Z., Schlüter, R., Ney, H.: Lattice decoding and rescoring with long-span neural network language models. In: Fifteenth Annual Conference of the International Speech Communication Association (2014)
Sundermeyer, M., Ney, H., Schlüter, R.: From feedforward to recurrent LSTM neural networks for language modeling. IEEE/ACM Trans. Audio, Speech. Lang. Process. 23(3), 517–529 (2015)
He, R., McAuley, J.: Modeling the visual evolution of fashion trends with one-class collaborative filtering. WWW (2016)
Perkel, J.M.: Why Jupyter is data scientists’ computational notebook of choice. Nature 563(7732), 145–147 (2018)
Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., Devin, M., Ghemawat, S., Irving, G., Isard, M., Kudlur, M.: Tensorflow: a system for large-scale machine learning. In: 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), 265–283 (2016)
Gulli, A., Pal, S.: Deep Learning with Keras. Packt Publishing Ltd (2017)
Huq, M.R., Ali, A., Rahman, A.: Sentiment analysis on Twitter data using KNN and SVM. IJACSA. Int. J. Adv. Comput. Sci. Appl. 8(6), 19–25 (2017)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Jahidul Islam Razin, M., Abdul Karim, M., Mridha, M.F., Rafiuddin Rifat, S.M., Alam, T. (2021). A Long Short-Term Memory (LSTM) Model for Business Sentiment Analysis Based on Recurrent Neural Network. In: Karuppusamy, P., Perikos, I., Shi, F., Nguyen, T.N. (eds) Sustainable Communication Networks and Application. Lecture Notes on Data Engineering and Communications Technologies, vol 55. Springer, Singapore. https://doi.org/10.1007/978-981-15-8677-4_1
Download citation
DOI: https://doi.org/10.1007/978-981-15-8677-4_1
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-8676-7
Online ISBN: 978-981-15-8677-4
eBook Packages: EngineeringEngineering (R0)