Language resources extracted from Wikipedia

  • Authors:
  • Denny Vrandecić;Philipp Sorg;Rudi Studer

  • Affiliations:
  • KIT and Wikimedia Deutschland, Karlsruhe and Berlin, Germany;KIT, Karlsruhe, Germany;KIT, Karlsruhe, Germany

  • Venue:
  • Proceedings of the sixth international conference on Knowledge capture
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Wikipedia provides an interesting amount of text for more than hundred languages. This also includes languages where no reference corpora or other linguistic resources are easily available. We have extracted background language models built from the content of Wikipedia in various languages. The models generated from Simple and English Wikipedia are compared to language models derived from other established corpora. The differences between the models in regard to term coverage, term distribution and correlation are described and discussed. We provide access to the full dataset and create visualizations of the language models that can be used exploratory. The paper describes the newly released dataset for 33 languages, and the services that we provide on top of them.