OWL reasoning with WebPIE: calculating the closure of 100 billion triples

  • Authors:
  • Jacopo Urbani;Spyros Kotoulas;Jason Maassen;Frank van Harmelen;Henri Bal

  • Affiliations:
  • Department of Computer Science, Vrije Universiteit, Amsterdam;Department of Computer Science, Vrije Universiteit, Amsterdam;Department of Computer Science, Vrije Universiteit, Amsterdam;Department of Computer Science, Vrije Universiteit, Amsterdam;Department of Computer Science, Vrije Universiteit, Amsterdam

  • Venue:
  • ESWC'10 Proceedings of the 7th international conference on The Semantic Web: research and Applications - Volume Part I
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

In previous work we have shown that the MapReduce framework for distributed computation can be deployed for highly scalable inference over RDF graphs under the RDF Schema semantics. Unfortunately, several key optimizations that enabled the scalable RDFS inference do not generalize to the richer OWL semantics. In this paper we analyze these problems, and we propose solutions to overcome them. Our solutions allow distributed computation of the closure of an RDF graph under the OWL Horst semantics. We demonstrate the WebPIE inference engine, built on top of the Hadoop platform and deployed on a compute cluster of 64 machines. We have evaluated our approach using some real-world datasets (UniProt and LDSR, about 0.9-1.5 billion triples) and a synthetic benchmark (LUBM, up to 100 billion triples). Results show that our implementation is scalable and vastly outperforms current systems when comparing supported language expressivity, maximum data size and inference speed.