Syntactic N-grams as machine learning features for natural language processing

  • Authors:
  • Grigori Sidorov;Francisco Velasquez;Efstathios Stamatatos;Alexander Gelbukh;Liliana Chanona-Hernández

  • Affiliations:
  • Center for Computing Research (CIC), Instituto Politécnico Nacional (IPN), Mexico City, Mexico;Center for Computing Research (CIC), Instituto Politécnico Nacional (IPN), Mexico City, Mexico;University of the Aegean, Greece;Center for Computing Research (CIC), Instituto Politécnico Nacional (IPN), Mexico City, Mexico;ESIME, Instituto Politécnico Nacional (IPN), Mexico City, Mexico

  • Venue:
  • Expert Systems with Applications: An International Journal
  • Year:
  • 2014

Quantified Score

Hi-index 12.05

Visualization

Abstract

In this paper we introduce and discuss a concept of syntactic n-grams (sn-grams). Sn-grams differ from traditional n-grams in the manner how we construct them, i.e., what elements are considered neighbors. In case of sn-grams, the neighbors are taken by following syntactic relations in syntactic trees, and not by taking words as they appear in a text, i.e., sn-grams are constructed by following paths in syntactic trees. In this manner, sn-grams allow bringing syntactic knowledge into machine learning methods; still, previous parsing is necessary for their construction. Sn-grams can be applied in any natural language processing (NLP) task where traditional n-grams are used. We describe how sn-grams were applied to authorship attribution. We used as baseline traditional n-grams of words, part of speech (POS) tags and characters; three classifiers were applied: support vector machines (SVM), naive Bayes (NB), and tree classifier J48. Sn-grams give better results with SVM classifier.