Skeletons in the parser: using a shallow parser to improve deep parsing
COLING '04 Proceedings of the 20th international conference on Computational Linguistics
Deep linguistic processing for spoken dialogue systems
DeepLP '07 Proceedings of the Workshop on Deep Linguistic Processing
Extracting a verb lexicon for deep parsing from FrameNet
DeepLP '07 Proceedings of the Workshop on Deep Linguistic Processing
Discourse annotation in the Monroe corpus
DiscAnnotation '04 Proceedings of the 2004 ACL Workshop on Discourse Annotation
Backbone extraction and pruning for speeding up a deep parser for dialogue systems
ScaNaLU '06 Proceedings of the Third Workshop on Scalable Natural Language Understanding
Increasing the coverage of a domain independent dialogue lexicon with VerbNet
ScaNaLU '06 Proceedings of the Third Workshop on Scalable Natural Language Understanding
Interpretation and generation in a knowledge-based tutorial system
KRAQ '06 Proceedings of the Workshop KRAQ'06 on Knowledge and Reasoning for Language Processing
TFlex: speeding up deep parsing with strategic pruning
Parsing '05 Proceedings of the Ninth International Workshop on Parsing Technology
Generic parsing for multi-domain semantic interpretation
Parsing '05 Proceedings of the Ninth International Workshop on Parsing Technology
Hi-index | 0.00 |
This thesis deals with the problem of building fast, accurate and portable parsers for natural language understanding. Our focus is a multi-domain dialogue system in which we need a deep linguistically-motivated parser to produce the representations of the input suitable for reasoning. In this dissertation, we are concerned with building parsers which have the wide coverage and portability offered by a general syntactic grammar without sacrificing parsing speed and accuracy. Our approach relies on a domain-independent deep parser and grammar which uses selectional restrictions to control parsing speed and accuracy. We develop a feature list representation as the basis for selectional restrictions, and a formal model for using selectional restrictions in a unification based framework. We then develop a lexicon design for multi-domain parsing and semantic interpretation. We show how the restrictions based on feature sets can be integrated with a traditional frame-based semantics, and extend our formal model to cover inheritance and defaults in the lexicon. We show that none of the existing large-scale ontologies and lexicons provide all the information necessary for parsing and semantic disambiguation, and we develop a parsing lexicon suitable for use with a wide-coverage grammar in multiple domains. Our domain-independent lexicon provides coverage and portability over four different application domains. To customize the representations produced by the parser for domain reasoning, we designed an architecture with mappings between our domain-independent ontology and a domain model. This architecture allows us to produce semantic representations optimally suited for different application domains. In addition, we use the mappings to specialize the lexicon for improved parsing speed and accuracy, and show that our specialization method significantly improves parsing performance. Finally, we develop a statistical model to learn selectional restrictions from corpora and show how it can be used to distinguish between acceptable and unacceptable verb-object pairs in data sets derived from our lexicon and from the World Street Journal corpus.