Text classification is a fundamental task in natural language processing, and with a huge and rapidly growing body of research devoted to it. However, there has been little work on investigating noise robustness for the developed approaches. In this work, we are bridging this gap, introducing results on noise robustness testing of modern text classification architectures for English and Russian languages. We benchmark the CharCNN and SentenceCNN models and introduce a new model, called RoVe (Robust Vectors), that we show to be the most robust to noise.
Similar content being viewed by others
References
A. Joulin, E. Grave, P. Bojanowski, and T. Mikolov, “Bag of Tricks for Efficient Text Classification,” arXiv:1607.01759 (2016).
V. Malykh, “Robust Word Vectors: Embeddings for Noisy Texts,” arXiv:1607.01759 (2018).
Y. Kim, Y. Jernite, D. Sontag, and A. M. Rush, “Character-Aware Neural Language Models,” in: Proceedings of the AAAI Conference on Artificial Intelligence (2016), pp. 2741– 2749.
A. L. Maas, R. E. Daly, P. T. Pham, D. Huang, A. Y. Ng, and Ch. Potts, “Learning Word Vectors for Sentiment Analysis,” in: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics (2011), pp. 142–150.
S. Cucerzan and E. Brill, “Spelling correction as an iterative process that exploits the collective knowledge of web users,” in: Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing (2004).
A. Joulin, E. Grave, P. Bojanowski, and T. Mikolov, “Bag of tricks for efficient text classification,” arXiv:1607.01759 (2016).
J. Howard and S. Ruder, “Fine-tuned Language Models for Text Classification,” arXiv:1801.06146 (2018).
A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, E. Kaiser, and I. Polosukhin, “Attention is all you need,” in: Advances in Neural Information Processing Systems (2017), pp. 6000–6010.
X. Zhang, J. J. Zhao, and Y. LeCun, “Character-level Convolutional Networks for Text Classification,” arXiv:1509.01626 (2017).
Y. Kim, “Convolutional Neural Networks for Sentence Classification,” arXiv:1408.5882 (2014).
K. Cho, B. van Merrienboer, D. Bahdanau, and Y. Bengio, “On the Properties of Neural Machine Translation: Encoder-Decoder Approaches,” arXiv:1409.1259 (2014).
D. Bahdanau, K. Cho, and Y. Bengio, “Neural Machine Translation by Jointly Learning to Align and Translate,” arXiv:1409.0473 (2014).
Y. Bengio, P. Simard, and P. Frasconi, “Learning long-term dependencies with gradient descent is difficult,” IEEE transactions on neural networks, 5, No. 2, 157–166 (1994).
E. Tutubalina and S. Nikolenko, “Inferring sentiment-based priors in topic models,” in: Mexican International Conference on Artificial Intelligence (2015), pp. 92–104.
J. Niu, Y. Yang, S. Zhang, Z. Sun, and W. Zhang, “Multi-task Character-Level Attentional Networks for Medical Concept Normalization,” Neural Processing Letters (2018), pp. 1–18.
Y. Li, T. Cohn, and Y. Baldwin, “Learning robust representations of text,” in: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing (2016), pp. 1979–1985.
N. V. Loukachevitch, et al., “SentiRuEval: Testing Object-oriented Sentiment Analysis Systems in Russian,” in: Proceedings of International Conference “Dialog” (2015).
V. Malykh, “Generalizable Architecture for Robust Word Vectors Tested by Noisy Paraphrases,” in: Supplementary Proceedings of the Sixth International Conference on Analysis of Images, Social Networks and Texts (AIST 2017), Moscow, Russia (2017).
Author information
Authors and Affiliations
Corresponding author
Additional information
Published in Zapiski Nauchnykh Seminarov POMI, Vol. 499, 2021, pp. 236–247.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Malykh, V., Lyalin, V. Improving Classification Robustness for Noisy Texts with Robust Word Vectors. J Math Sci 273, 605–613 (2023). https://doi.org/10.1007/s10958-023-06522-x
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10958-023-06522-x