Skip to main content

Sentence Representations and Beyond

  • Chapter
  • First Online:
Neural Representations of Natural Language

Part of the book series: Studies in Computational Intelligence ((SCI,volume 783))

Abstract

This chapter discusses representations for larger structures in natural language. The primary focus is on the sentence level. However, many of the techniques also apply to sub-sentence structures (phrases), and super-sentence structures (documents). The three main types of representations discussed here are: unordered models, such as sum of word embeddings; sequential models, such as recurrent neural networks; and structured models, such as recursive autoencoders.

A sentence is a group of words expressing a complete thought

English Composition and Literature,

Webster,1923

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

eBook
USD 16.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 99.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://groups.google.com/forum/#!msg/word2vec-toolkit/Q49FIrNOQRo/DoRuBoVNFb0J.

References

  • Bird, Steven, Ewan Klein, and Edward Loper. 2009. Natural language processing with Python. O’Reilly Media, Inc.

    Google Scholar 

  • Blei, David M., Andrew Y. Ng, and Michael I. Jordan. 2003. Latent Dirichlet allocation. The Journal of Machine Learning Research 3: 993–1022.

    MATH  Google Scholar 

  • Bowman, Samuel R., Jon Gauthier, Abhinav Rastogi, Raghav Gupta, Christopher D. Manning, and Christopher Potts. 2016a. A fast unified model for parsing and sentence understanding. arXiv:1603.06021.

  • Bowman, Samuel R., Luke Vilnis, Oriol Vinyals, Andrew M. Dai, Rafal Jozefowicz, and Samy Bengio. 2016b. Generating sentences from a continuous space. In International conference on learning representations (ICLR) Workshop.

    Google Scholar 

  • Cho, Kyunghyun, Bart van Merrienboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using rnn encoder-decoder for statistical machine translation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), 1724–1734. Doha, Qatar: Association for Computational Linguistics.

    Google Scholar 

  • Dumais, Susan T., George W. Furnas, Thomas K. Landauer, Scott Deerwester, and Richard Harshman. 1988. Using latent semantic analysis to improve access to textual information. In Proceedings of the SIGCHI conference on Human factors in computing systems, 281–285. ACM.

    Google Scholar 

  • Goller, Christoph and Andreas Kuchler. 1996. Learning task-dependent distributed representations by back propagation through structure. In IEEE international conference on neural networks, 1996, vol. 1, 347–352. IEEE.

    Google Scholar 

  • Hofmann, Thomas. 2000. Learning the similarity of documents: An information geometric approach to document retrieval and categorization. In Advances in neural information processing systems, 914–920.

    Google Scholar 

  • Honnibal, Matthew and Mark Johnson. 2015. An improved non-monotonic transition system for dependency parsing. In Proceedings of the 2015 conference on empirical methods in natural language processing, 1373–1378. Lisbon, Portugal: Association for Computational Linguistics.

    Google Scholar 

  • Horvat, Matic and William Byrne. 2014. A graph-based approach to string regeneration. In EACL, 85–95.

    Google Scholar 

  • Iyyer, Mohit, Jordan Boyd-Graber, and Hal Daumé III. 2014a. Generating sentences from semantic vector space representations. In NIPS workshop on learning semantics.

    Google Scholar 

  • Iyyer, Mohit, Jordan Boyd-Graber, Leonardo Claudino, Richard Socher, and Hal Daumé III. 2014b. A neural network for factoid question answering over para-graphs. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), 633–644.

    Google Scholar 

  • Kingma, D. P. and M. Welling. 2014. Auto-encoding variational bayes. In The international conference on learning representations (ICLR). arXiv:1312.6114 [stat.ML].

  • Kiros, Ryan, Yukun Zhu, Ruslan Salakhutdinov, Richard S. Zemel, Antonio Torralba, Raquel Urtasun, and Sanja Fidler. 2015. Skip-thought vectors. In CoRR. arXiv:1506.06726.

  • Lau, Jey Han, and Timothy Baldwin. 2016. An empirical evaluation of doc2vec with practical insights into document embedding generation. In ACL, 78.

    Google Scholar 

  • Le, Quoc and Tomas Mikolov. 2014. Distributed representations of sentences and documents. In Proceedings of the 31st international conference on machine learning (ICML-14), 1188–1196.

    Google Scholar 

  • Li, Bofang, Tao Liu, Zhe Zhao, Puwei Wang, and Xiaoyong Du. 2017. Neural bag-of-ngrams. In AAAI, 3067–3074.

    Google Scholar 

  • Manning, C.D. and H. Schütze. 1999. Foundations of statistical natural language processing. Cambridge: MIT Press. ISBN: 9780262133609.

    Google Scholar 

  • Manning, Christopher D., Mihai Surdeanu, John Bauer, Jenny Finkel, Steven J. Bethard, and David McClosky. 2014. The Stanford CoreNLP natural language processing toolkit. In Association for computational linguistics (ACL) system demonstrations, 55–60.

    Google Scholar 

  • Mesnil, Grégoire, Tomas Mikolov, Marc’Aurelio Ranzato, and Yoshua Bengio. 2014. Ensemble of generative and discriminative techniques for sentiment analysis of movie reviews. arXiv:1412.5335.

  • Mitchell, Jeff and Mirella Lapata. 2008. Vector-based models of semantic composition. In ACL, 236–244.

    Google Scholar 

  • Pollack, Jordan B. 1990. Recursive distributed representations. Artificial Intelligence, 46 (1): 77–105. ISSN: 0004-3702. https://doi.org/10.1016/0004-3702(90)90005-K.

  • Rehůrek, Radim and Petr Sojka. 2010. Software framework for topic modelling with large corpora. English. In Proceedings of the LREC 2010 workshop on new challenges for NLP frameworks, 45–50. Valletta, Malta: ELRA. http://is.muni.cz/publication/884893/en

  • Ritter, Samuel, Cotie Long, Denis Paperno, Marco Baroni, Matthew Botvinick, and Adele Goldberg. 2015. Leveraging preposition ambiguity to assess compositional distributional models of semantics. In The fourth joint conference on lexical and computational semantics.

    Google Scholar 

  • Socher, Richard. 2014. Recursive deep learning for natural language processing and computer vision. Ph.D. thesis. Stanford University.

    Google Scholar 

  • Socher, Richard, Christopher D. Manning, and Andrew Y. Ng. 2010. Learning continuous phrase representations and syntactic parsing with recursive neural networks. In Proceedings of the NIPS-2010 deep learning and unsupervised feature learning workshop, 1–9.

    Google Scholar 

  • Socher, Richard, Eric H. Huang, Jeffrey Pennington, Andrew Y. Ng, and Christopher D. Manning. 2011a. Dynamic pooling and unfolding recursive autoencoders for paraphrase detection. In Advances in neural information processing systems, 24.

    Google Scholar 

  • Socher, Richard, Cliff C Lin, Chris Manning, and Andrew Y Ng. 2011b. Parsing natural scenes and natural language with recursive neural networks. In Proceedings of the 28th international conference on machine learning (ICML-11), 129–136.

    Google Scholar 

  • Socher, Richard, Jeffrey Pennington, Eric H. Huang, Andrew Y. Ng, and Christopher D. Manning. 2011c. Semi-supervised recursive autoencoders for predicting sentiment distributions. In Proceedings of the 2011 conference on empirical methods in natural language processing (EMNLP).

    Google Scholar 

  • Socher, Richard, Brody Huval, Christopher D. Manning, and Andrew Y. Ng. 2012. Semantic compositionality through recursive matrix-vector spaces. In Proceedings of the 2012 joint conference on empirical methods in natural language processing and computational natural language learning, 1201–1211. Association for Computational Linguistics.

    Google Scholar 

  • Socher, Richard, Andrej Karpathy, Quoc V. Le, Christopher D. Manning, and Y.Ng Andrew. 2014. Grounded compositional semantics for finding and describing images with sentences. Transactions of the Association for Computational Linguistics 2: 207–218.

    Google Scholar 

  • Stenetorp, Pontus. 2013. Transition-based dependency parsing using recursive neural networks. In Deep learning workshop at the, 2013. Conference on neural information processing systems (NIPS). Nevada, USA: Lake Tahoe.

    Google Scholar 

  • Wang, Sida and Christopher D. Manning. 2012. Baselines and bigrams: Simple, good sentiment and topic classification. In Proceedings of the 50th annual meeting of the association for computational linguistics: Short papers, vol. 2, 90–94. Association for Computational Linguistics.

    Google Scholar 

  • Wang, Rui, Wei Liu, and Chris McDonald. 2017. A matrix-vector recurrent unit model for capturing compositional semantics in phrase embeddings. In International conference on information and knowledge management.

    Google Scholar 

  • White, Lyndon, Roberto Togneri, Wei Liu, and Mohammed Bennamoun. 2015. How well sentence embeddings capture meaning. In Proceedings of the 20th Australasian document computing symposium. ADCS ’15, 9:1–9:8. Parramatta, NSW, Australia: ACM. ISBN: 978-1-4503-4040-3, https://doi.org/10.1145/2838931.2838932.

  • White, Lyndon, Roberto Togneri, Wei Liu, and Mohammed Bennamoun. 2016a. Generating bags of words from the sums of their word embeddings. In 17th international conference on intelligent text processing and computational linguistics (CICLing).

    Google Scholar 

  • White, Lyndon, Roberto Togneri, Wei Liu, and Mohammed Bennamoun. 2016b. Modelling sentence generation from sum of word embedding vectors as a mixed integer programming problem. In IEEE international conference on data mining: High dimensional data mining workshop (ICDM: HDM). https://doi.org/10.1109/ICDMW.2016.0113.

  • White, L., R. Togneri, W. Liu, and M. Bennamoun. 2017. Learning distributions of meant color. arXiv:1709.09360 [cs.CL].

  • Zhang, Jiajun, Shujie Liu, Mu Li, Ming Zhou, and Chengqing Zong. 2014. Bilingually constrained phrase embeddings for machine translation. In ACL.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Lyndon White .

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Singapore Pte Ltd.

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

White, L., Togneri, R., Liu, W., Bennamoun, M. (2019). Sentence Representations and Beyond. In: Neural Representations of Natural Language. Studies in Computational Intelligence, vol 783. Springer, Singapore. https://doi.org/10.1007/978-981-13-0062-2_5

Download citation

Publish with us

Policies and ethics