Skip to main content
Log in

Neural Network Sentiment Classification of Russian Sentences into Four Classes

  • Published:
Automatic Control and Computer Sciences Aims and scope Submit manuscript

Abstract

This work is devoted to the sentiment classification of Russian sentences into four classes: positive, negative, mixed, and neutral. Unlike most modern works in this area, a mixed-sentiment class of sentences is introduced. Mixed-sentiment sentences contain both positively and negatively inflected speech at the same time. To solve this problem, the following tools were applied: an attention-based LSTM neural network, a dual attention-based GRU neural network, a and BERT neural network with several output layer modifications providing classification into four classes. Experiments comparing the effectiveness of various neural networks have been carried out on three corpora of Russian-language sentences. Two corpora are made up of users reviews: one with clothing reviews and the other with hotel reviews. The third corpus is made up of news articles from Russian publications. The best average weighted F-measure in the experiments of 0.90 is achieved by a BERT model on the clothing review corpus. The best F-measures for positive and negative sentences are noted on the same corpus, amounting to 0.92 and 0.93, respectively. The best classification results for neutral and mixed-sentiment sentences are achieved on the news corpus. For them, the F-measure is 0.72 and 0.58, respectively. As a result of the experiments, a significant superiority of BERT transfer neural networks over the previous generation LSTM and GRU neural networks is shown, which is most pronounced when classifying texts with weakly expressed sentiments. An analysis of the errors shows that “adjacent” sentiment classes (positive/negative and mixed) account for a larger proportion of errors in classification using BERT than “opposite” classes (positive and negative, neutral and mixed).

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1.
Fig. 2.
Fig. 3.
Fig. 4.
Fig. 5.
Fig. 6.
Fig. 7.
Fig. 8.

Notes

  1. https://www.trivago.ru

  2. http://opencorpora.org

REFERENCES

  1. Potts, C., Wu, Z., Geiger, A., and Kiela, D., Dynasent: A dynamic benchmark for sentiment analysis, Proc. 59th Annu. Meeting of the Association for Computational Linguistics and the 11th Int. Joint Conf. on Natural Language Processing, Zong, Ch., Xia, F., Li, W., and Navigli, R., Eds., Association for Computational Linguistics, 2020, vol. 1, pp. 2388–2404. https://doi.org/10.18653/v1/2021.acl-long.186

  2. Hamborg, F. and Donnay, K., NewsMTSC: A dataset for (multi-)target-dependent sentiment classification in political news articles, Proc. 16th Conf. of the European Chapter of the Association for Computational Linguistics: Main Volume, Merlo, P., Tiedemann, J., and Tsarfaty, R., Eds., Association for Computational Linguistics, 2021, pp. 1663–1675. https://doi.org/10.18653/v1/2021.eacl-main.142

  3. Liu, B., Sentiment Analysis and Opinion Mining, Synthesis Lectures on Human Language Technologies, vol. 5, Cham: Springer, 2012. https://doi.org/10.1007/978-3-031-02145-9

  4. Habimana, O., Li, Yu., Li, R., Gu, X., and Yu, G., Sentiment analysis using deep learning approaches: An overview, Sci. China Inf. Sci., 2020, vol. 63, no. 1, p. 111102. https://doi.org/10.1007/s11432-018-9941-6

    Article  Google Scholar 

  5. Smetanin, S. and Komarov, M., Deep transfer learning baselines for sentiment analysis in Russian, Inf. Process. Manage., 2021, vol. 58, no. 3, p. 102484. https://doi.org/10.1016/j.ipm.2020.102484

    Article  Google Scholar 

  6. Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., and Sutskever, I., Language models are unsupervised multitask learners, OpenAI, 2019.

    Google Scholar 

  7. Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., and Le, Q.V., XLNet: Generalized autoregressive pretraining for language understanding, Adv. Neural Inf. Process. Syst., 2019, vol. 32, pp. 5754–5764. https://proceedings.neurips.cc/paper/2019/hash/dc6a7e655d7e5840e66733e9ee67cc69-Abstract.html.

  8. Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K., BERT: Pre-training of deep bidirectional transformers for language understanding, Proc. 2019 Conf. of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis: Association for Computational Linguistics, 2018, vol. 1, pp. 4171–4186. https://doi.org/10.18653/v1/N19-1423

  9. Kalchbrenner, N., Grefenstette, E., and Blunsom, P., A convolutional neural network for modelling sentences, Proc. 52nd Annu. Meeting of the Association for Computational Linguistics, Baltimore, Md.: Association for Computational Linguistics, 2014, vol. 1, pp. 655–665. https://doi.org/10.3115/v1/p14-1062

  10. Paramonov, I. and Poletaev, A., Adaptation of semantic rule-based sentiment analysis approach for russian language, 30th Conf. of Open Innovations Association FRUCT, Oulu, Finland, 2021, IEEE, 2021, pp. 155–164. https://doi.org/10.23919/fruct53335.2021.9599992

  11. Kenyon-Dean, K., Ahmed, E., Fujimoto, S., Georges-Filteau, J., Glasz, C., Kaur, B., Lalande, A., Bhanderi, S., Belfer, R., Kanagasabai, N., Sarrazingendron, R., Verma, R., and Ruths, D., Sentiment Analysis: It’s Complicated!, Proc. 2018 Conf. of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans: Association for Computational Linguistics, 2018, vol. 1, pp. 1886–1895. https://doi.org/10.18653/v1/n18-1171

  12. Tan, X., Cai, Yi., and Zhu, C., Recognizing conflict opinions in aspect-level sentiment classification with dual attention networks, Proc. 2019 Conf. on Empirical Methods in Natural Language Processing and the 9th Int. Joint Conf. on Natural Language Processing (EMNLP-IJCNLP), Hong Kong: Association for Computational Linguistics, 2019, pp. 3426–3431. https://doi.org/10.18653/v1/d19-1342

  13. Soleymani, M., Garcia, D., Jou, B., Schuller, B., Chang, Sh.-F., and Pantic, M., A survey of multimodal sentiment analysis, Image Vision Comput., 2017, vol. 65, pp. 3–14. https://doi.org/10.1016/j.imavis.2017.08.003

    Article  Google Scholar 

  14. Oberländer, L.A.M. and Klinger, R., An analysis of annotated corpora for emotion classification in text, Proc. 27th Int. Conf. on Computational Linguistics, Santa Fe, N.M.: 2018, pp. 2104–2119. https://aclanthology.org/ C18-1179.

    Google Scholar 

  15. Radford, A., Jozefowicz, R., and Sutskever, I., Learning to generate reviews and discovering sentiment, 6th Int. Conf. on Learning Representations, Vancouver: 2017. https://openreview.net/forum?id=SJ71VXZAZ.

  16. Wang, Ye., Huang, M., Zhu, X., and Zhao, L., Attention-based LSTM for aspect-level sentiment classification, Proc. 2016 Conf. on Empirical Methods in Natural Language Processing, Austin, Texas: Association for Computational Linguistics, 2016, pp. 606–615. https://doi.org/10.18653/v1/d16-1058

  17. Chen, P., Sun, Z., Bing, L., and Yang, W., Recurrent attention network on memory for aspect sentiment analysis, Proc. 2017 Conf. on Empirical Methods in Natural Language Processing, Copenhagen: Association for Computational Linguistics, 2017, pp. 452–461. https://doi.org/10.18653/v1/d17-1047

  18. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, Ł., and Polosukhin, I., Attention is all you need, Adv. Neural Inf. Process. Syst., 2017, vol. 30, pp. 5998–6008.

    Google Scholar 

  19. Socher, R., Perelygin, A., Wu, J., Chuang, J., Manning, C.D., Ng, A., and Po, C., Recursive deep models for semantic compositionality over a sentiment treebank, Proc. 2013 Conference on Empirical Methods in Natural Language Processing, Seattle, Wash.: 2013, pp. 1631–1642. https://aclanthology.org/D13-1170.

  20. Smetanin, S. and Komarov, M., Sentiment analysis of product reviews in Russian using convolutional neural networks, 2019 IEEE 21st Conf. on Business Informatics (CBI), Moscow, 2019, IEEE, 2019, vol. 1, pp. 482–486. https://doi.org/10.1109/cbi.2019.00062

Download references

Funding

This work was supported by ongoing institutional funding. No additional grants to carry out or direct this particular research were obtained.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to M. A. Kosterin.

Ethics declarations

The authors of this work declare that they have no conflicts of interest.

Additional information

Translated by A. Kolemesin

Publisher’s Note.

Allerton Press remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Kosterin, M.A., Paramonov, I.V. Neural Network Sentiment Classification of Russian Sentences into Four Classes. Aut. Control Comp. Sci. 57, 727–739 (2023). https://doi.org/10.3103/S0146411623070052

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.3103/S0146411623070052

Keywords:

Navigation