Skip to main content

Learning Lexical-Semantic Relations Using Intuitive Cognitive Links

  • Conference paper
  • First Online:
Advances in Information Retrieval (ECIR 2019)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 11437))

Included in the following conference series:

Abstract

Identifying the specific semantic relations between words is crucial for IR and NLP systems. Our goal in this paper is twofold. First, we want to understand whether learning a classifier for one semantic relation (e.g. hypernymy) can gain from concurrently learning another classifier for a cognitively-linked semantic relation (e.g. co-hyponymy). Second, we evaluate how these systems perform where only few labeled examples exist. To answer the first question, we rely on a multi-task neural network architecture, while for the second we use self-learning to evaluate whether semi-supervision improves performance. Our results on two popular datasets as well as a novel dataset proposed in this paper show that concurrent learning of semantic relations consistently benefits performance. On the other hand, we find that semi-supervised learning can be useful depending on the semantic relation. The code and the datasets are available at https://bit.ly/2Qitasd.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 109.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 139.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    This issue is out of the scope of this paper.

  2. 2.

    We are aware that this architecture can further be improved by additional task-specific inputs, but as a great deal of possible models can be proposed, which deserve intensive research, this issue remains out of the scope of this paper.

  3. 3.

    https://github.com/esantus/ROOT9.

  4. 4.

    Available at https://bit.ly/2Qitasd.

  5. 5.

    http://wordnetcode.princeton.edu/3.0/.

  6. 6.

    This value was set experimentally.

  7. 7.

    A large number of hypernym pairs contain the root synset ā€œentityā€, i.e. path length equals to 0.

  8. 8.

    All datasets are available at https://bit.ly/2Qitasd.

  9. 9.

    A multi-class model learns to separate between several classes and direct comparison with binary models is not fair. Nevertheless, we report its performance as it highlights the potential of multi-class learning for problems that are cognitively similar.

  10. 10.

    The code is available at https://bit.ly/2Qitasd.

  11. 11.

    Column 3 of TableĀ 4.

  12. 12.

    Note that due to the lexical split process, results can not directly be compared to the ones obtained over ROOT9 or RUMEN.

References

  1. Amini, M., Laviolette, F., Usunier, N.: A transductive bound for the voted classifier with an application to semi-supervised learning. In: 22nd Annual Conference on Neural Information Processing Systems (NIPS), pp. 65ā€“72 (2008)

    Google ScholarĀ 

  2. Amini, M.R., Usunier, N.: Learning with Partially Labeled and Interdependent Data. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-15726-9

    BookĀ  MATHĀ  Google ScholarĀ 

  3. Attia, M., Maharjan, S., Samih, Y., Kallmeyer, L., Solorio, T.: Cogalex-V shared task: GHHH - detecting semantic relations via word embeddings. In: Workshop on Cognitive Aspects of the Lexicon, pp. 86ā€“91 (2016)

    Google ScholarĀ 

  4. Baroni, M., Bernardi, R., Do, N.Q., Shan, C.-C: Entailment above the word level in distributional semantics. In: 13th Conference of the European Chapter of the Association for Computational Linguistics (EACL), pp. 23ā€“32 (2012)

    Google ScholarĀ 

  5. Baroni, M., Lenci, A.: How we blessed distributional semantic evaluation. In: Workshop on Geometrical Models of Natural Language Semantics (GEMS) Associated to Conference on Empirical Methods on Natural Language Processing (EMNLP), pp. 1ā€“10 (2011)

    Google ScholarĀ 

  6. Benotto, G.: Distributional Models for Semantic Relations: A Study on Hyponymy and Antonymy. Ph.D. thesis, University of Pisa (2015)

    Google ScholarĀ 

  7. Bingel, J., SĆøgaard, A.: Identifying beneficial task relations for multi-task learning in deep neural networks. arXiv preprint arXiv:1702.08303 (2017)

  8. Braud, C., Plank, B., SĆøgaard, A.: Multi-view and multi-task training of RST discourse parsers. In: 26th International Conference on Computational Linguistics (COLING), pp. 1903ā€“1913 (2016)

    Google ScholarĀ 

  9. Caruana, R.: Multitask learning. In: Thrun, S., Pratt, L. (eds.) Learning to Learn, pp. 95ā€“133. Springer, Boston (1998). https://doi.org/10.1007/978-1-4615-5529-2_5

    ChapterĀ  Google ScholarĀ 

  10. Chapelle, O., Scholkopf, B., Zien, A.: Semi-supervised learning. IEEE Trans. Neural Networks 20(3), 542 (2009)

    ArticleĀ  Google ScholarĀ 

  11. Chollet, F.: Keras. https://keras.io (2015)

  12. Dias, G., Moraliyski, R., Cordeiro, J., Doucet, A., Ahonen-Myka, H.: Automatic discovery of word semantic relations using paraphrase alignment and distributional lexical semantics analysis. Nat. Lang. Eng. 16(4), 439ā€“467 (2010)

    ArticleĀ  Google ScholarĀ 

  13. Dong, L., Mallinson, J., Reddy, S., Lapata, M.: Learning to paraphrase for question answering. In: Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 886ā€“897 (2017)

    Google ScholarĀ 

  14. Gambhir, M., Gupta, V.: Recent automatic text summarization techniques: a survey. Artif. Intell. Rev. 47(1), 1ā€“66 (2017)

    ArticleĀ  Google ScholarĀ 

  15. Glavas, G., Ponzetto, S.P.: Dual tensor model for detecting asymmetric lexico-semantic relations. In: Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1758ā€“1768 (2017)

    Google ScholarĀ 

  16. Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: 13th International Conference on Artificial Intelligence and Statistics, pp. 249ā€“256 (2010)

    Google ScholarĀ 

  17. Harris, Z.S.: Distributional structure. Word 10(2ā€“3), 146ā€“162 (1954)

    ArticleĀ  Google ScholarĀ 

  18. Hearst, M.: Automatic acquisition of hyponyms from large text corpora. In: 14th Conference on Computational Linguistics (COLING), pp. 539ā€“545 (1992)

    Google ScholarĀ 

  19. Kathuria, N., Mittal, K., Chhabra, A.: A comprehensive survey on query expansion techniques, their issues and challenges. Int. J. Comput. Appl. 168(12), (2017)

    ArticleĀ  Google ScholarĀ 

  20. Kiela, D., Hill, F., Clark, S.: Specializing word embeddings for similarity or relatedness. In: Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 2044ā€“2048 (2015)

    Google ScholarĀ 

  21. Kotlerman, L., Dagan, I., Szpektor, I., Zhitomirsky-Geffet, M.: Directional distributional similarity for lexical inference. Nat. Lang. Eng. 16(4), 359ā€“389 (2010)

    ArticleĀ  Google ScholarĀ 

  22. Levy, O., Remus, S., Biemann, C., Dagan, I.: Do supervised distributional methods really learn lexical inference relations? In: Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 970ā€“976 (2015)

    Google ScholarĀ 

  23. Liu, P., Qiu, X., Huang, X.: Adversarial multi-task learning for text classification. In: 55th Annual Meeting of the Association for Computational Linguistics (ACL) (2017)

    Google ScholarĀ 

  24. Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. CoRR abs/1301.3781 (2013)

    Google ScholarĀ 

  25. Miller, G.A., Beckwith, R., Fellbaum, C., Gross, D., Miller, K.J.: Introduction to wordnet: an on-line lexical database. Int. J. Lexicogr. 3(4), 235ā€“244 (1990)

    ArticleĀ  Google ScholarĀ 

  26. Nguyen, K.A., Kƶper, M., Schulte im Walde, S., Vu, N.T.: Hierarchical embeddings for hypernymy detection and directionality. In: Conference on Empirical Methods in Natural Language Processing, pp. 233ā€“243 (2017)

    Google ScholarĀ 

  27. Nguyen, K.A., Schulte im Walde, S., Vu, N.T.: Distinguishing antonyms and synonyms in a pattern-based neural network. In: 15th Conference of the European Chapter of the Association for Computational Linguistics (EACL), pp. 76ā€“85 (2017)

    Google ScholarĀ 

  28. Pedregosa, F., et al.: Scikit-learn: machine learning in python. J. Mach. Learn. Res. 12(Oct), 2825ā€“2830 (2011)

    MathSciNetĀ  MATHĀ  Google ScholarĀ 

  29. Pennington, J., Socher, R., Manning, C.D.: Glove: global vectors for word representation. In: Conference on Empirical Methods on Natural Language Processing (EMNLP), pp. 1532ā€“1543 (2014)

    Google ScholarĀ 

  30. Roller, S., Erk, K., Boleda, G.: Inclusive yet selective: Supervised distributional hypernymy detection. In: 25th International Conference on Computational Linguistics (COLING), pp. 1025ā€“1036 (2014)

    Google ScholarĀ 

  31. Ruder, S., Plank, B.: Strong baselines for neural semi-supervised learning under domain shift. In: 56th Annual Meeting of the Association for Computational Linguistics (ACL) (2018)

    Google ScholarĀ 

  32. Santus, E., Lenci, A., Chiu, T., Lu, Q., Huang, C.: Nine features in a random forest to learn taxonomical semantic relations. In: 10th International Conference on Language Resources and Evaluation, pp. 4557ā€“4564 (2016)

    Google ScholarĀ 

  33. Santus, E., Shwartz, V., Schlechtweg, D.: Hypernyms under siege: linguistically-motivated artillery for hypernymy detection. In: 15th Conference of the European Chapter of the Association for Computational Linguistics, pp. 65ā€“75 (2017)

    Google ScholarĀ 

  34. Santus, E., Yung, F., Lenci, A., Huang, C.R.: Evalution 1.0: an evolving semantic dataset for training and evaluation of distributional semantic models. In: 4th Workshop on Linked Data in Linguistics (LDL) Associated to Association for Computational Linguistics and Asian Federation of Natural Language Processing (ACL-IJCNLP), pp. 64ā€“69 (2015)

    Google ScholarĀ 

  35. Shwartz, V., Dagan, I.: Cogalex-V shared task: lexnet - integrated path-based and distributional method for the identification of semantic relations. CoRR abs/1610.08694 (2016)

    Google ScholarĀ 

  36. Shwartz, V., Goldberg, Y., Dagan, I.: Improving hypernymy detection with an integrated path-based and distributional method. In: 54th Annual Meeting of the Association for Computational Linguistics, pp. 2389ā€“2398 (2016)

    Google ScholarĀ 

  37. Snow, R., Jurafsky, D., Ng, A.Y.: Learning syntactic patterns for automatic hypernym discovery. In: 17th International Conference on Neural Information Processing Systems (NIPS), pp. 1297ā€“1304 (2004)

    Google ScholarĀ 

  38. Vylomova, E., Rimell, L., Cohn, T., Baldwin, T.: Take and took, gaggle and goose, book and read: evaluating the utility of vector differences for lexical relation learning. In: 54th Annual Meeting of the Association for Computational Linguistics, pp. 1671ā€“1682 (2016)

    Google ScholarĀ 

  39. Weeds, J., Clarke, D., Reffin, J., Weir, D.J., Keller, B.: Learning to distinguish hypernyms and co-hyponyms. In: 5th International Conference on Computational Linguistics (COLING), pp. 2249ā€“2259 (2014)

    Google ScholarĀ 

  40. Yu, Z., Wang, H., Lin, X., Wang, M.: Learning term embeddings for hypernymy identification. In: 24th International Joint Conference on Artificial Intelligence, pp. 1390ā€“1397 (2015)

    Google ScholarĀ 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Gaƫl Dias .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

Ā© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Balikas, G., Dias, G., Moraliyski, R., Akhmouch, H., Amini, MR. (2019). Learning Lexical-Semantic Relations Using Intuitive Cognitive Links. In: Azzopardi, L., Stein, B., Fuhr, N., Mayr, P., Hauff, C., Hiemstra, D. (eds) Advances in Information Retrieval. ECIR 2019. Lecture Notes in Computer Science(), vol 11437. Springer, Cham. https://doi.org/10.1007/978-3-030-15712-8_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-15712-8_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-15711-1

  • Online ISBN: 978-3-030-15712-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics