Latent Topic Models of Surface Syntactic Information

  • Roberto Basili
  • C. Giannone
  • Danilo Croce
  • C. Domeniconi
Conference paper

DOI: 10.1007/978-3-642-23954-0_22

Part of the Lecture Notes in Computer Science book series (LNCS, volume 6934)
Cite this paper as:
Basili R., Giannone C., Croce D., Domeniconi C. (2011) Latent Topic Models of Surface Syntactic Information. In: Pirrone R., Sorbello F. (eds) AI*IA 2011: Artificial Intelligence Around Man and Beyond. AI*IA 2011. Lecture Notes in Computer Science, vol 6934. Springer, Berlin, Heidelberg

Abstract

Topic Models like Latent Dirichlet Allocation have been widely used for their robustness in estimating text models through mixtures of latent topics. Although LDA has been mostly used as a strictly lexicalized approach, it can be effectively applicable to a much richer set of linguistic structures. A novel application of LDA is here presented that acquires suitable grammatical generalizations for semantic tasks tightly dependent on NL syntax. We show how the resulting topics represent suitable generalizations over syntactic structures and lexical information as well. The evaluation on two different classification tasks, such as predicate recognition and question classification, shows that state of the art results are obtained.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Copyright information

© Springer-Verlag Berlin Heidelberg 2011

Authors and Affiliations

  • Roberto Basili
    • 1
  • C. Giannone
    • 1
  • Danilo Croce
    • 1
  • C. Domeniconi
    • 2
  1. 1.Dept. of Enterprise EngineeringUniversity of Roma Tor VergataRomaItaly
  2. 2.Dept. of Computer ScienceGeorge Mason UniversityUSA

Personalised recommendations