Skip to main content
Log in

An improved gated recurrent unit based on auto encoder for sentiment analysis

  • Original Research
  • Published:
International Journal of Information Technology Aims and scope Submit manuscript

Abstract

Sentiment analysis is a particularly common task for determining user thoughts and has been widely used in Natural Language Processing (NLP) applications. Gated Recurrent Unit (GRU) was already effectively integrated into the NLP process with comparatively excellent results. GRU networks outperform traditional recurrent neural networks in sequential learning tasks and solve gradient vanishing and explosion limitations of RNNs. This paper introduces a new method called Normalize Auto-Encoded GRU (NAE-GRU) to address data dimensionality reduction using an Auto-Encoder and to improve performance through batch normalization. Empirically, we demonstrate that with slight adjustments to hyperparameters and optimization of statistic vectors, the proposed model achieves excellent results in sentiment classification on benchmark datasets. The developed NAE-GRU approach outperformed other various traditional approaches in terms of accuracy and convergence rate. The experimental results showed that the developed NAE-GRU approach accomplished better sentiment analysis accuracy of 91.32%, 82.27%, 87.43%, and 84.49% on IMDB, SSTb, Amazon review, and Yelp review datasets respectively. Furthermore, experimental results have shown that the developed approach is proficient in reducing the loss function and capturing long-term relationships with an effective design that achieved excellent results compared to state-of-the-art methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Algorithm 1:
Fig. 2
Algorithm 2:
Fig. 3
Fig. 4
Fig. 5
Algorithm 3:
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Data availability

The following information was supplied regarding data availability: The code is available at GitHub: https://github.com/zunimalik777/Improved-DNNs-Autoencoder-GRU-Sentiment-Analysis.git. The collected datasets are available on the following links: https://www.kaggle.com/datasets/lakshmi25npathi/imdb-dataset-of-50k-movie-reviews. https://snap.stanford.edu/data/web-Amazon.html. https://www.tensorflow.org/datasets/catalog/yelp_polarity_reviews. https://www.kaggle.com/competitions/sentiment-analysis-on-movie-reviews/data.

Data transparency

Authors will ensure data transparency.

References

  1. Tang D, Wei F, Qin B, Yang N, Liu T, Zhou M (2015) Sentiment embeddings with applications to sentiment analysis. IEEE Trans Knowl Data Eng 28(2):496–509

    Article  Google Scholar 

  2. Zulqarnain M, Ghazali R, Hassim YMM, Rehan M (2020) A comparative review on deep learning models for text classification. Indones J Electr Eng Comput Sci 19(1):1856–1866

    Google Scholar 

  3. Ghazali R, Husaini NA, Ismail LH, Herawan T, Hassim YMM (2014) The performance of a Recurrent HONN for temperature time series prediction. In: 2014 International Joint Conference on Neural Networks (IJCNN), IEEE, pp 518–524

  4. Abbas Q (2019) MADeep-automatic microaneurysms detection on retinal fundus by using region growing and deep neural networks. Int J Comput Sci Netw Secur 19(1):161–166

    Google Scholar 

  5. Wadawadagi R, Pagi V (2022) Polarity enriched attention network for aspect-based sentiment analysis. Int J Inf Technol 14(6):2767–2778

    Google Scholar 

  6. Hughes M, Li I, Kotoulas S, Suzumura T (2017) Medical text classification using convolutional neural networks. Informatics for health: connected citizen-led wellness and population health. IOS Press, pp 246–250

    Google Scholar 

  7. Zulqarnain M, Alsaedi AKZ, Ghazali R, Ghouse MG, Sharif W, Husaini NA (2021) A comparative analysis on question classification task based on deep learning approaches. PeerJ Comput Sci 7:e570

    Article  Google Scholar 

  8. Priya CSR, Deepalakshmi P (2023) Sentiment analysis from unstructured hotel reviews data in social network using deep learning techniques. Int J Inf Technol. https://doi.org/10.1007/s41870-023-01419-z

    Article  Google Scholar 

  9. Al-harbi O (2019) A comparative study of feature selection methods for dialectal arabic sentiment classification using support vector machine. Int J Comput Sci Netw Secur 19(1):167–176

    Google Scholar 

  10. Bengio Y, Courville A, Vincent P (2012) Representation learning: a review and new perspectives, pp 1–30

  11. Kalchbrenner N, Grefenstette E, Blunsom P (2014) A convolutional neural network for modelling sentences. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 655–665

  12. Tang D, Qin B, Liu T (2015) Document modeling with gated recurrent neural network for sentiment classification. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, pp 1422–1432

  13. Song J, Qin S, Zhang P (2016) Chinese text categorization based on deep belief networks. IEEE ICIS 2016:1–5

    Google Scholar 

  14. Soni J, Mathur K (2022) Sentiment analysis based on aspect and context fusion using attention encoder with LSTM. Int J Inf Technol 14(7):3611–3618

    Google Scholar 

  15. Cho K et al (2014) Learning phrase representations using rnn encoder-decoder for statistical machine translation. ArXiv, pp 1–15

  16. Divate MS (2021) Sentiment analysis of Marathi news using LSTM. Int J Inf Technol 13(5):2069–2074

    Google Scholar 

  17. Kumar A (2018) Self-attention enhanced recurrent neural networks for sentence classification. In: IEEE Symposium Series on Computational Intelligence, pp 905–911

  18. Yang T, Tseng T, Chen C (2016) Recurrent neural network-based language models with variation in net topology, language, and granularity. Int Conf Asian Lang Process 3:71–74

    Google Scholar 

  19. Rumelhart DE, Todd PM (1993) Learning and connectionist representations. Attention Perform XIV 2:3–30 (Synergies in Experimental Psychology, Artificial Intelligence, and Cognitive Neuroscience)

    Article  Google Scholar 

  20. Aamir M, Nawi NM, Mahdin HB, Naseem R, Zulqarnain M (2020) Auto-encoder variants for solving handwritten digits classification problem. Int J Fuzzy Logic Intell Syst 20(1):8–16

    Article  Google Scholar 

  21. Chung J, Gulcehre C, Cho K, Bengio Y (2014) Empirical Evaluation of gated recurrent neural networks on sequence modeling, pp 1–9

  22. Zulqarnain M, Ghazali R, Hassim YMM, Rehan M (2020) Text classification based on gated recurrent unit combines with support vector machine. Int J Electr Comput Eng 10:2088–8708

    Google Scholar 

  23. Lipton ZC, Berkowitz J, Elkan C (2015) A critical review of recurrent neural networks for sequence learning, pp. 1–38. arXiv: 1506.00019v4 [ cs . LG ]

  24. Noaman HM, Sarhan SS, Rashwan MA (2012) Enhancing recurrent neural network-based language models by word tokenization. Hum-Centric Comput Inf Sci 8(1):12

    Article  Google Scholar 

  25. Hao Y, Sheng Y, Wang J (2019) Variant gated recurrent units with encoders to preprocess packets for payload-aware intrusion detection. IEEE Access 7:49985–49998. https://doi.org/10.1109/ACCESS.2019.2910860

    Article  Google Scholar 

  26. Davidson DW (2016) Modeling missing data in clinical time series with RNNs. Proc Mach Learn Healthc 58(4):725–737

    Google Scholar 

  27. Justus D, Brennan J, Bonner S, McGough AS (2019) Predicting the computational cost of deep learning models. In: Proceedings-2018 IEEE International Conference on Big Data, Big Data, pp 3873–3882

  28. Maas AL, Daly RE, Pham PT, Huang D, Ng AY, Potts C (2011) Learning word vectors for sentiment analysis. Proc Annu Assoc Comput Linguist Hum Lang Technol 1:142–150

    Google Scholar 

  29. Pang B, Lee L (2005) Seeing stars: exploiting class relationships for sentiment categorization with respect to rating scales. Proc Annu Meet Assoc Comput Linguist 3(1):115–124

    Google Scholar 

  30. Socher R, Perelygin A, Wu J (2013) Recursive deep models for semantic compositionality over a sentiment Treebank. In: Proceedings, pp 1631–1642

  31. Kingma DP, Ba JL (2015) A method for stochastic optimization. arXiv, pp 1–15

  32. Hinton G (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 15:1929–1958

    MathSciNet  Google Scholar 

  33. Zulqarnain M, Ghazali R, Aamir M, Hassim YMM (2022) An efficient two-state GRU based on feature attention mechanism for sentiment analysis. Multimedia Tools Appl. https://doi.org/10.1007/s11042-022-13339-4

    Article  Google Scholar 

  34. Pennington J, Socher R, Manning CD (2014) GloVe : global vectors for word representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, pp 1532–1543

  35. Conneau A, Schwenk H, Barrault L, Lecun Y (2016) Very deep convolutional networks for text classification. arXiv preprint arXiv:1606.01781

  36. Deep pyramid convolutional neural networks for text categorization (2017). In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 562–570

  37. Yang Z, Dai Z, Yang Y, Carbonell J, Salakhutdinov RR, Le QV (2019) Xlnet: generalized autoregressive pretraining for language understanding. Advances in neural information processing systems. Springer, Cham, pp 5754–5764

    Google Scholar 

  38. Liu B (2020) Text sentiment analysis based on CBOW model and deep learning in big data environment. J Ambient Intell Humaniz Comput 11(2):451–458. https://doi.org/10.1007/s12652-018-1095-6

    Article  Google Scholar 

  39. Shen D et al (2018) Baseline needs more love: on simple word-embedding-based models and associated pooling mechanisms. ACL Annu Meet Assoc Comput Linguist Proc Conf (Long Paper) 1:440–450. https://doi.org/10.18653/v1/p18-1041

    Article  Google Scholar 

  40. Devlin J, Chang M-W, Lee K, Toutanova K (2018) Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805

  41. Sun C, Qiu X, Xu Y, Huang X (2019) Howto fine-tune bert for text classification? China national conference on chinese computational linguistics. Springer, Cham, pp 194–206

    Google Scholar 

  42. Ren H, Lu H (2018) Compositional coding capsule network with k-means routing for text classification. arXiv preprint arXiv:1810.09177

  43. Xu J, Du Q (2019) A deep investigation into fasttext. Proc IEEE Int Conf High Perform Comput Commun. https://doi.org/10.1109/HPCC/SmartCity/DSS.2019.00234

    Article  Google Scholar 

  44. Zhang X, Zhao J, Le Cun Y (2015) Character-level convolutional networks for text classification. Advances in neural information processing systems. Springer, pp 649–657

    Google Scholar 

  45. Socher R et al (2013) Recursive deep models for semantic compositionality over a sentiment treebank. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)

  46. Socher R, Pennington J, Huang EH, Ng AY, Manning CD (2011) Semi-supervised recursive autoencoders for predicting sentiment distributions. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp 151–161

  47. Wang T, Liu L, Zhang H, Zhang L, Chen X (2020) Joint character-level convolutional and generative adversarial networks for text classification. Complexity 2020:1–11. https://doi.org/10.1155/2020/8516216

    Article  Google Scholar 

  48. Ma Y, Fan H, Zhao C (2019) Feature-based fusion adversarial recurrent neural networks for text sentiment classification. IEEE Access 7:132542–132551. https://doi.org/10.1109/ACCESS.2019.2940506

    Article  Google Scholar 

  49. Fu X, Yang J, Li J, Fang M, Wang H (2018) Lexicon-enhanced LSTM with attention for general sentiment analysis. IEEE Access 6:71884–71891. https://doi.org/10.1109/ACCESS.2018.2878425

    Article  Google Scholar 

  50. Camacho-Collados J, Pilehvar MT (2018) On the role of text preprocessing in neural network architectures: an evaluation study on text categorization and sentiment analysis. ArXivPreprint. https://doi.org/10.18653/v1/w18-5406

    Article  Google Scholar 

Download references

Funding

The authors did not receive support from organization for the submitted work.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Muhammad Zulqarnain.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict interest.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zulqarnain, M., Alsaedi, A.K.Z., Sheikh, R. et al. An improved gated recurrent unit based on auto encoder for sentiment analysis. Int. j. inf. tecnol. 16, 587–599 (2024). https://doi.org/10.1007/s41870-023-01600-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s41870-023-01600-4

Keywords

Navigation