Latent topic models of surface syntactic information

  • Authors:
  • Roberto Basili;C. Giannone;Danilo Croce;C. Domeniconi

  • Affiliations:
  • Dept. of Enterprise Engineering, University of Roma Tor Vergata, Roma, Italy;Dept. of Enterprise Engineering, University of Roma Tor Vergata, Roma, Italy;Dept. of Enterprise Engineering, University of Roma Tor Vergata, Roma, Italy;Dept. of Computer Science, George Mason University

  • Venue:
  • AI*IA'11 Proceedings of the 12th international conference on Artificial intelligence around man and beyond
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Topic Models like Latent Dirichlet Allocation have been widely used for their robustness in estimating text models through mixtures of latent topics. Although LDA has been mostly used as a strictly lexicalized approach, it can be effectively applicable to a much richer set of linguistic structures. A novel application of LDA is here presented that acquires suitable grammatical generalizations for semantic tasks tightly dependent on NL syntax. We show how the resulting topics represent suitable generalizations over syntactic structures and lexical information as well. The evaluation on two different classification tasks, such as predicate recognition and question classification, shows that state of the art results are obtained.