Statistical modality tagging from rule-based annotations and crowdsourcing

  • Authors:
  • Vinodkumar Prabhakaran;Michael Bloodgood;Mona Diab;Bonnie Dorr;Lori Levin;Christine D. Piatko;Owen Rambow;Benjamin Van Durme

  • Affiliations:
  • Columbia University;University of Maryland;Columbia University;University of Maryland;Carnegie Mellon University;Johns Hopkins University;Columbia University;Johns Hopkins University

  • Venue:
  • ExProM '12 Proceedings of the Workshop on Extra-Propositional Aspects of Meaning in Computational Linguistics
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

We explore training an automatic modality tagger. Modality is the attitude that a speaker might have toward an event or state. One of the main hurdles for training a linguistic tagger is gathering training data. This is particularly problematic for training a tagger for modality because modality triggers are sparse for the overwhelming majority of sentences. We investigate an approach to automatically training a modality tagger where we first gathered sentences based on a high-recall simple rule-based modality tagger and then provided these sentences to Mechanical Turk annotators for further annotation. We used the resulting set of training data to train a precise modality tagger using a multi-class SVM that delivers good performance.