Using LSI for text classification in the presence of background text
Proceedings of the tenth international conference on Information and knowledge management
A decision tree of bigrams is an accurate predictor of word sense
NAACL '01 Proceedings of the second meeting of the North American Chapter of the Association for Computational Linguistics on Language technologies
An empirical study of the domain dependence of supervised word sense disambiguation systems
EMNLP '00 Proceedings of the 2000 Joint SIGDAT conference on Empirical methods in natural language processing and very large corpora: held in conjunction with the 38th Annual Meeting of the Association for Computational Linguistics - Volume 13
One sense per collocation and genre/topic variations
EMNLP '00 Proceedings of the 2000 Joint SIGDAT conference on Empirical methods in natural language processing and very large corpora: held in conjunction with the 38th Annual Meeting of the Association for Computational Linguistics - Volume 13
Domain kernels for word sense disambiguation
ACL '05 Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics
Domain-specific sense distributions and predominant sense acquisition
HLT '05 Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing
Applying alternating structure optimization to word sense disambiguation
CoNLL-X '06 Proceedings of the Tenth Conference on Computational Natural Language Learning
On robustness and domain adaptation using SVD for word sense disambiguation
COLING '08 Proceedings of the 22nd International Conference on Computational Linguistics - Volume 1
Domain adaptation with structural correspondence learning
EMNLP '06 Proceedings of the 2006 Conference on Empirical Methods in Natural Language Processing
Word sense disambiguation using OntoNotes: an empirical study
EMNLP '08 Proceedings of the Conference on Empirical Methods in Natural Language Processing
SemEval-2007 task 17: English lexical sample, SRL and all words
SemEval '07 Proceedings of the 4th International Workshop on Semantic Evaluations
UBC-ALM: combining k-NN with SVD for WSD
SemEval '07 Proceedings of the 4th International Workshop on Semantic Evaluations
Domain adaptation for statistical classifiers
Journal of Artificial Intelligence Research
SemEval-2010 task 17: all-words word sense disambiguation on a specific domain
DEW '09 Proceedings of the Workshop on Semantic Evaluations: Recent Achievements and Future Directions
Knowledge-based WSD on specific domains: performing better than generic supervised WSD
IJCAI'09 Proceedings of the 21st international jont conference on Artifical intelligence
All words domain adapted WSD: finding a middle ground between supervision and unsupervision
ACL '10 Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics
Negative training data can be harmful to text classification
EMNLP '10 Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing
Value for money: balancing annotation effort, lexicon building and accuracy for multilingual WSD
COLING '10 Proceedings of the 23rd International Conference on Computational Linguistics
Identification of domain-specific senses in a machine-readable dictionary
HLT '11 Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies: short papers - Volume 2
Proceedings of the 20th ACM international conference on Information and knowledge management
Word sense disambiguation as an integer linear programming problem
SETN'12 Proceedings of the 7th Hellenic conference on Artificial Intelligence: theories and applications
A new minimally-supervised framework for domain word sense disambiguation
EMNLP-CoNLL '12 Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning
Hi-index | 0.00 |
The lack of positive results on supervised domain adaptation for WSD have cast some doubts on the utility of hand-tagging general corpora and thus developing generic supervised WSD systems. In this paper we show for the first time that our WSD system trained on a general source corpus (Bnc) and the target corpus, obtains up to 22% error reduction when compared to a system trained on the target corpus alone. In addition, we show that as little as 40% of the target corpus (when supplemented with the source corpus) is sufficient to obtain the same results as training on the full target data. The key for success is the use of unlabeled data with svd, a combination of kernels and svm.