Conference Paper

Using Dynamic Condor-Based Services for Classifying Schizophrenia in Diffusion Tensor Images.

Cardiff Univ., Cardiff
DOI: 10.1109/CCGRID.2008.12 Conference: 8th IEEE International Symposium on Cluster Computing and the Grid (CCGrid 2008), 19-22 May 2008, Lyon, France
Source: DBLP

ABSTRACT Diffusion tensor imaging (DTI) provides insight into the white matter of the human brain, which is affected by schizophrenia. By comparing a patient group to a control group, the DTI-images are on average expected to be different for white matter regions. Principal component analysis (PCA) and linear discriminant analysis (LDA) are used to classify the groups. In this work, the number of principal components is optimised for obtaining the minimal classification error. A robust estimate of this error is computed in a cross-validation framework, using different compositions of the data into a training and a testing set Previously, sequential runs were performed in MATLAB, resulting in long execution times. In this paper we describe an experiment where this application was run on a grid with minimal modifications and user effort. We have adopted a service-based approach that autonomously launches image analysis services onto a campus-wide Condor pool comprising of volunteer resources. This allows high throughput analysis of our data in a dynamic resource pool. The challenge in adopting such an approach comes from the nature of the resources, which change randomly with time and thus require fault tolerance. Through this approach we have reduced the computation time of each dataset from 90 minutes to less than 10. A minimal classification error of 22% was obtained, using 15 principal components.

  • [Show abstract] [Hide abstract]
    ABSTRACT: Diffusion Tensor MRI (DTI) is a rather recent image acquisition modality that can help identify disease processes in nerve bundles in the brain. Due to the large and complex nature of such data, its analysis requires new and sophisticated pipelines that are more efficiently executed within a grid environment. We present our progress over the past four years in the development and porting of the DTI analysis pipeline to grids. Starting with simple jobs submitted from the command-line, we moved towards a workflow-based implementation and finally into a web service that can be accessed via web browsers by end-users. The analysis algorithms evolved from basic to state-of-the-art, currently enabling the automatic calculation of a population-specific `atlas' where even complex brain regions are described in an anatomically correct way. Performance statistics show a clear improvement over the years, representing a mutual benefit from both a technology push and application pull.
    10th IEEE/ACM International Conference on Cluster, Cloud and Grid Computing, CCGrid 2010, 17-20 May 2010, Melbourne, Victoria, Australia; 01/2010
  • [Show abstract] [Hide abstract]
    ABSTRACT: Many research institutions and Universities own computational capacity that is not effectively utilized, thereby providing an opportunity for such institutions to use such capacity to offer Cloud services (to both internal and external users). However, the unreliability and unpredictability of these resources mean that their use in the context of a Service Level Agreement (SLA) is high risk, leading to a reduction in reputation as well as economic penalties in case of SLA violation. We propose a methodology that addresses the issues of unreliability and unpredictability such that Cloud software services could be hosted upon volunteered resources. To enable the harnessing of these resources we rely on autonomic fault management techniques that allow such systems to independently adapt the resources they use based upon their perception of individual resource reliability. Using our approach we were able to scale out the backend infrastructure of the Cloud service elastically (min 30thinspaces per worker), opportunistically and autonomically. We address two key questions in this article: can a campus volunteer infrastructure be used in Cloud provisioning? What measures are necessary in order to ensure reliability at the resource level? Copyright © 2011 John Wiley & Sons, Ltd.
    Concurrency and Computation Practice and Experience 06/2012; 24(9):992 - 1014. DOI:10.1002/cpe.1715 · 0.78 Impact Factor
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: This paper presents the design, implementation, and usage of a virtual laboratory for medical image analysis. It is fully based on the Dutch grid, which is part of the Enabling Grids for E-sciencE (EGEE) production infrastructure and driven by the gLite middleware. The adopted service-oriented architecture enables decoupling the user-friendly clients running on the user's workstation from the complexity of the grid applications and infrastructure. Data are stored on grid resources and can be browsed/viewed interactively by the user with the Virtual Resource Browser (VBrowser). Data analysis pipelines are described as Scufl workflows and enacted on the grid infrastructure transparently using the MOTEUR workflow management system. VBrowser plug-ins allow for easy experiment monitoring and error detection. Because of the strict compliance to the grid authentication model, all operations are performed on behalf of the user, ensuring basic security and facilitating collaboration across organizations. The system has been operational and in daily use for eight months (December 2008), with six users, leading to the submission of 9000 jobs/month in average and the production of several terabytes of data.
    IEEE transactions on information technology in biomedicine: a publication of the IEEE Engineering in Medicine and Biology Society 01/2010; 14:979-985. DOI:10.1109/TITB.2010.2046742 · 2.07 Impact Factor

Full-text (2 Sources)

Available from
May 27, 2014