Multiplying Concept Sources for Graph Modeling

  • Authors:
  • Loic Maisonnasse;Eric Gaussier;Jean Pierre Chevallet

  • Affiliations:
  • LIG-UJF, Grenoble Cedex 9, France 38041;LIG-UJF, Grenoble Cedex 9, France 38041;IPAL-I2R, Singapore 119613

  • Venue:
  • Advances in Multilingual and Multimodal Information Retrieval
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

The main idea in this paper is to incorporate medical knowledge in the language modeling approach to information retrieval (IR). Our model makes use of the textual part of ImageCLEFmed corpus and of the medical knowledge as found in the Unified Medical Language System (UMLS) knowledge sources. The use of UMLS allows us to create a conceptual representation of each sentence in the corpus. We use these representations to create a graph model for each document. As in the standard language modeling approach, we evaluate the probability that a document graph model generates the query graph. Graphs are created from medical texts and queries, and are built for different languages, with different methods. After developing the graph model, we present our tests, which involve mixing different concepts sources (i.e. languages and methods) for the matching of the query and text graphs. Results show that using language model on concepts provides good results in IR. Multiplying the concept sources further improves the results. Lastly, using relations between concepts (provided by the graphs under consideration) improves results when only few conceptual sources are used to analyze the query.