SWARM: a scientific workflow for supporting bayesian approaches to improve metabolic models

  • Authors:
  • Xinghua Shi;Rick Stevens

  • Affiliations:
  • University of Chicago, Chicago, IL, USA;Argonne National Laboratory, Argonne, IL, USA

  • Venue:
  • CLADE '08 Proceedings of the 6th international workshop on Challenges of large applications in distributed environments
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

With the exponential growth of complete genome sequences, the analysis of these sequences is becoming a powerful approach to build genome-scale metabolic models. These models can be used to study individual molecular components and their relationships, and eventually study cells as systems. However, constructing genome-scale metabolic models manually is time-consuming and labor-intensive. This property of manual model-building process causes the fact that much fewer genome-scale metabolic models are available comparing to hundreds of genome sequences available. To tackle this problem, we design SWARM, a scientific workflow that can be utilized to improve genome-scale metabolic models in high-throughput fashion. SWARM deals with a range of issues including the integration of data across distributed resources, data format conversions, data update, and data provenance. Putting altogether, SWARM streamlines the whole modeling process that includes extracting data from various resources, deriving training datasets to train a set of predictors and applying Bayesian techniques to assemble the predictors, inferring on the ensemble of predictors to insert missing data, and eventually improving draft metabolic networks automatically. By the enhancement of metabolic model construction, SWARM enables scientists to generate many genome-scale metabolic models within a short period of time and with less effort.