Skip to main content
Log in

Fine-grained semantic textual similarity measurement via a feature separation network

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Semantic text similarity (STS), which measures the semantic similarity of sentences, is an important task in the field of NLP. It has a wide range of applications, such as machine translation (MT), semantic search, and summarization. In recent years, with the development of deep neural networks, the existing semantic similarity measurement has made great progress. In particular, pretraining models, such as BERT-based models, which have been good representations of sentence features, have set a new state-of-the-art on STS tasks. Although a large amount of corpus data are used in the pretraining stage, there is no fine-grained semantic analysis. We observe that many sentences, such as user reviews and the QA corpus, can be abstractly regarded as including two core parts: a) this sentence states a certain attribute; and b) this attribute is described by descriptive words. This feature is particularly prominent in the corpus of reviews. Motivated by the above observations, in this paper, we propose a feature separation network (FSN) model, which can further separate and extract attribute features and description features and then measure the semantic similarity according to the separated features. To better verify the effectiveness of our model, we propose an unsupervised approach to construct the semantic similarity dataset in the review domain. Experimental results demonstrate that our method outperforms the general semantic similarity measurement method.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Algorithm 1
Algorithm 2
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

Notes

  1. https://www.autohome.com.cn/

  2. https://storage.googleapis.com/cluebenchmark/tasks/afqmc_public.zip

  3. https://storage.googleapis.com/bert_models/2018_11_03/chinese_L-12_H-768_A-12.zip

  4. https://github.com/ymcui/Chinese-XLNet

References

  1. Agirre E, Banea C, Cer D, Diab M, Gonzalez Agirre A, Mihalcea R, Rigau Claramunt G, Wiebe J (2016) Semeval-2016 task 1: semantic textual similarity, monolingual and cross-lingual evaluation. In: SemEval-2016. 10th international workshop on semantic evaluation; 2016 Jun 16–17; San Diego, CA. Stroudsburg (PA): ACL; 2016. ACL (Association for Computational Linguistics), pp 497–511

  2. Xiao Y, Liu L, Huang G, Cui Q, Huang S, Shi S, Chen J (2022) BiTIIMT: a bilingual text-infilling method for interactive machine translation. In: Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), pp 1958–1969

  3. See A, Liu PJ, Manning CD (2017) Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th annual meeting of the association for computational linguistics (vol 1: Long Papers), pp 1073–1083

  4. Lei X, Qian X, Zhao G (2016) Rating prediction based on social sentiment from textual reviews. IEEE trans multimed 18(9):1910–1921

    Article  Google Scholar 

  5. Kuo Y, Fu M, Tsai W, Lee K, Chen L (2016) Integrated microblog sentiment analysis from users’ social interaction patterns and textual opinions. Appl Intell 44(2):399–413

    Article  Google Scholar 

  6. Kim HK, Kim M (2016) Model-induced term-weighting schemes for text classification. Appl Intell 45(1):30–43

    Article  Google Scholar 

  7. Zhao G, Liu Z, Chao Y, Qian X (2021) CAPER: context-aware personalized emoji recommendation. IEEE Trans Knowl Data Eng 33(9):3160–3172

    Article  Google Scholar 

  8. Lopez-Gazpio I, Maritxalar M, Gonzalez-Agirre A, Rigau G, Uria L, Agirre E (2017) Interpretable semantic textual similarity: finding and explaining differences between sentences. Knowl-Based Syst 119:186–199

    Article  Google Scholar 

  9. Yuan Z, Sun S, Duan L, Li C, Wu X, Xu C (2020) Adversarial multimodal network for movie story question answering. IEEE Trans Multimed 23:1744–1756

    Article  Google Scholar 

  10. Janda HK, Pawar A, Du S, Mago V (2019) Syntactic, semantic and sentiment analysis: the joint effect on automated essay evaluation. IEEE Access 7:108486–108503

    Article  Google Scholar 

  11. Dong J, Li X, Xu D (2018) Cross-media similarity evaluation for web image retrieval in the wild. IEEE Trans Multimed 20(9):2371–2384

    Article  Google Scholar 

  12. Zhou Y, Zhao G, Qian X (2022) Dialogue state tracking based on hierarchical slot attention and contrastive learning. In: Hasan MA, Xiong L (eds) Proceedings of the 31st ACM international conference on information & knowledge management, pp 4737–4741

  13. Wu Y, Liao L, Zhang G, Lei W, Zhao G, Qian X, Chua T-S (2022) State graph reasoning for multimodal conversational recommendation. IEEE Trans Multimed, 1–1. https://doi.org/10.1109/TMM.2022.3155900

  14. Kenton JDM-WC, Toutanova LK (2019) Bert: pre-training of deep bidirectional transformers for language understanding, 4171– 4186

  15. Chen K, Zhang Z, Long J, Zhang H (2016) Turning from TF-IDF to TF-IGM for term weighting in text classification. Expert Syst Appl 66:245–260

    Article  Google Scholar 

  16. Wang T, Guo J, Wu Z, Xu T (2021) IFTA: iterative filtering by using TF-AICL algorithm for Chinese encyclopedia knowledge refinement. Appl Intell 51(8):6265–6293

    Article  Google Scholar 

  17. Yang Y, Yuan S, Cer D, Kong S-y, Constant N, Pilar P, Ge H, Sung Y-H, Strope B, Kurzweil R (2018) Learning semantic textual similarity from conversations, 164–174

  18. Lan W, Xu W (2018) Neural network models for paraphrase identification, semantic textual similarity, natural language inference, and question answering. In: Proceedings of the 27th international conference on computational linguistics, pp 3890–3902

  19. Wang T, Li J, Guo J (2021) A scalable parallel chinese online encyclopedia knowledge denoising method based on entry tags and spark cluster. Appl Intell 51(10):7573–7599

    Article  Google Scholar 

  20. Zhu J, He Y, Zhao G, Bo X, Qian X (2022) Joint reason generation and rating prediction for explainable recommendation. IEEE Trans Knowl Data Eng

  21. Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Levy O, Lewis M, Zettlemoyer L, Stoyanov V (2019) Roberta: a robustly optimized bert pretraining approach. arXiv:1907.11692

  22. Yang Z, Dai Z, Yang Y, Carbonell J, Salakhutdinov RR, Le QV (2019) Xlnet: generalized autoregressive pretraining for language understanding. Adv Neural Inf Process Syst, 32

  23. Sun Y, Wang S, Li Y, Feng S, Tian H, Wu H, Wang H. (2020) Ernie 2.0: a continual pre-training framework for language understanding. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 8968–8975

  24. Lan Z, Chen M, Goodman S, Gimpel K, Sharma P, Soricut R (2019) Albert: a lite bert for self-supervised learning of language representations

  25. May C, Wang A, Bordia S, Bowman SR, Rudinger R (2019) On measuring social biases in sentence encoders, 622–628

  26. Zhang T, Kishore V, Wu F, Weinberger KQ, Artzi Y (2019) Bertscore: evaluating text generation with Bert, 1–43

  27. Qiao Y, Xiong C, Liu Z, Liu Z (2019) Understanding the behaviors of bert in ranking. arXiv:1904.07531

  28. Zheng X, Zhao G, Zhu L, Qian X (2022) PERD: personalized emoji recommendation with dynamic user preference. In: Proceedings of the 45th international ACM SIGIR conference on research and development in information retrieval, pp 1922–1926

  29. Reimers N, Gurevych I (2019) Sentence-bert: sentence embeddings using siamese bert-networks, 3982–3992

  30. Das A, Yenala H, Chinnakotla MK, Shrivastava M (2016) Together we stand: Siamese networks for similar question retrieval. In: Proceedings of the 54th annual meeting of the association for computational linguistics, ACL 2016, August 7–12, 2016, Berlin, Germany, vol 1: long papers. The Association for Computer Linguistics

  31. Ranasinghe T, Orǎsan C, Mitkov R (2019) Semantic textual similarity with siamese neural networks. In: Proceedings of the international conference on recent advances in natural language processing (RANLP 2019), pp 1004–1011

  32. Pontes EL, Huet S, Linhares AC, Torres-Moreno J-M (2018) Predicting the semantic textual similarity with siamese CNN and LSTM. arXiv:1810.10641

  33. Cranias L, Papageorgiou H, Piperidis S (1994) A matching technique in example-based machine translation. In: COLING 1994 vol 1: the 15th international conference on computational linguistics

  34. Salton G, Wong A, Yang C-S (1975) A vector space model for automatic indexing. Commun ACM 18(11):613–620

    Article  MATH  Google Scholar 

  35. Qiang J, Li Y, Yuan Y, Wu X (2018) Short text clustering based on Pitman-Yor process mixture model. Appl Intell 48(7):1802–1812

    Article  Google Scholar 

  36. Zhao G, Lou P, Qian X, Hou X (2020) Personalized location recommendation by fusing sentimental and spatial context. Knowl-Based Syst 196:105849

    Article  Google Scholar 

  37. Zhang D, Zhu Z, Kang S, Zhang G, Liu P (2021) Syntactic and semantic analysis network for aspect-level sentiment classification. Appl Intell 51(8):6136–6147

    Article  Google Scholar 

  38. Majumder G, Pakray P, Das R, Pinto D (2021) Interpretable semantic textual similarity of sentences using alignment of chunks with classification and regression. Appl Intell 51(10):7322– 7349

    Article  Google Scholar 

  39. Wang T, Liu L, Liu N, Zhang H, Zhang L, Feng S (2020) A multi-label text classification method via dynamic semantic representation model and deep neural network. Appl Intell 50(8):2339–2351

    Article  Google Scholar 

  40. Zheng X, Zhao G, Zhu L, Zhu J, Qian X (2022) What you like, what I am: online dating recommendation via matching individual preferences with features. IEEE Trans Knowl Data Eng

  41. Vo A, Nguyen Q, Ock C (2020) Semantic and syntactic analysis in learning representation based on a sentiment analysis model. Appl Intell 50(3):663–680

    Article  Google Scholar 

  42. Kim Y (2014) Convolutional neural networks for sentence classification. In: Moschitti A, Pang B, Daelemans W (eds) Proceedings of the 2014 conference on empirical methods in natural language processing, EMNLP 2014, October 25–29, 2014, Doha, Qatar, a meeting of SIGDAT, a special interest group of the ACL. ACL, pp 1746–1751

  43. Hu B, Lu Z, Li H, Chen Q (2014) Convolutional neural network architectures for matching natural language sentences. In: Advances in neural information processing systems, pp 2042–2050

  44. Mueller J, Thyagarajan A (2016) Siamese recurrent architectures for learning sentence similarity. In: Proceedings of the AAAI conference on artificial intelligence, vol 30

  45. Zhao G, Lei X, Qian X, Mei T (2018) Exploring users’ internal influence from reviews for social recommendation. IEEE Trans Multimed 21(3):771–781

    Article  Google Scholar 

  46. Bahdanau D, Cho KH, Bengio Y (2015) Neural machine translation by jointly learning to align and translate

  47. Wang S, Jiang J (2016) Learning natural language inference with LSTM. In: Proceedings of the 2016 conference of the north american chapter of the association for computational linguistics: human language technologies, pp 1442–1451

  48. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: NIPS

  49. Peters M, Neumann M, Iyyer M, Gardner M, Clark C, Lee K, Zettlemoyer L (2018) Deep contextualized word representations. In: Proceedings of the 2018 conference of the North American chapter of the association for computational linguistics: human language technologies, vol 1 (Long Papers), pp 2227–2237

  50. Wang B, Kuo C-CJ (2020) Sbert-wk: a sentence embedding method by dissecting bert-based word models. IEEE/ACM Trans Audio Speech Lang Process 28:2146–2157. https://doi.org/10.1109/TASLP.2020.3008390

    Article  Google Scholar 

  51. Arora S, Liang Y, Ma T (2017) A simple but tough-to-beat baseline for sentence embeddings. In: 5th international conference on learning representations, ICLR 2017

  52. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780

    Article  Google Scholar 

  53. Liu X, Chen Q, Deng C, Zeng H, Chen J, Li D, Tang B (2018) Lcqmc: A large-scale chinese question matching corpus. In: Proceedings of the 27th international conference on computational linguistics, pp 1952–1962

  54. Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv:1412.6980

  55. Mitra B, Diaz F, Craswell N (2017) Learning to match using local and distributed representations of text for web search. In: Proceedings of the 26th international conference on World Wide Web, pp 1291–1299

  56. Agirre E, Banea C, Cardie C, Cer D, Diab M, Gonzalez-Agirre A, Guo W, Mihalcea R, Rigau G, Wiebe J (2014) Sem Eval-2014 task 10: Multilingual semantic textual similarity. In: Proceedings of the 8th international workshop on semantic evaluation (SemEval 2014). https://doi.org/10.3115/v1/S14-2010, https://www.aclweb.org/anthology/S14-2010. Association for Computational Linguistics, pp 81–91

Download references

Acknowledgements

This work is funded in part by the NSFC, China under Grant 61902309; in part by the Fundamental Research Funds for the Central Universities, China (xxj022019003, xzd012022006); in part by the China Postdoctoral Science Foundation (2020M683496); and in part by the National Postdoctoral Innovative Talents Support Program, China (BX20190273); in part by the Humanities and Social Sciences Foundation of Ministry of Education, China under Grant 16XJAZH003; and in part by the Science and Technology Program of Xi’an, China under Grant 21RGZN0017.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Guoshuai Zhao.

Ethics declarations

Conflict of Interests

The authors declare that they have no conflicts of interest.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Chen, Q., Zhao, G., Wu, Y. et al. Fine-grained semantic textual similarity measurement via a feature separation network. Appl Intell 53, 18205–18218 (2023). https://doi.org/10.1007/s10489-022-04448-6

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-022-04448-6

Keywords

Navigation