Show simple item record

Algorithmic Enhancements to Data Colocation Grid Frameworks for Big Data Medical Image Processing

dc.creatorBao, Shunxing
dc.date.accessioned2020-08-23T15:51:11Z
dc.date.available2018-11-27
dc.date.issued2018-11-27
dc.identifier.urihttps://etd.library.vanderbilt.edu/etd-11222018-050053
dc.identifier.urihttp://hdl.handle.net/1803/14739
dc.description.abstractLarge-scale medical imaging studies to date have predominantly leveraged in-house, laboratory-based or traditional grid computing resources for their computing needs, where the applications often use hierarchical data structures (e.g., Network file system file stores) or databases (e.g., COINS, XNAT) for storage and retrieval. The resulting performance for laboratory-based approaches reveal that performance is impeded by standard network switches since typical processing can saturate network bandwidth during transfer from storage to processing nodes for even moderate-sized studies. On the other hand, the grid may be costly to use due to the dedicated resources used to execute the tasks and lack of elasticity. With increasing availability of cloud-based big data frameworks, such as Apache Hadoop, cloud-based services for executing medical imaging studies have shown promise. Despite this promise, our studies have revealed that existing big data frameworks illustrate different performance limitations for medical imaging applications, which calls for new algorithms that optimize their performance and suitability for medical imaging. For instance, Apache HBase’s data distribution strategy of region split and merge is detrimental to the hierarchical organization of imaging data (e.g., project, subject, session, scan, slice). Big data medical image processing applications involving multi-stage analysis often exhibit significant variability in processing times ranging from a few seconds to several days. Due to the sequential nature of executing the analysis stages by traditional software technologies and platforms, any errors in the pipeline are only detected at the later stages despite the sources of errors predominantly being the highly compute-intensive first stage. This wastes precious computing resources and incurs prohibitively higher costs for re-executing the application. To address these challenges, this research propose a framework - Hadoop & HBase for Medical Image Processing HadoopBase-MIP) - which develops a range of performance optimization algorithms and employs a number of system behaviors modeling for data storage, data access, and data processing. We also introduce how to build up prototypes to help empirical system behaviors verification. Furthermore, we introduce a discovery with the development of HadoopBase-MIP about a new type of contrast for medical imaging deep brain structure enhancement. And finally, we show how to move forward the Hadoop based framework design into a commercialized big data / High performance computing cluster with a cheap, scalable and geographically distributed file system.
dc.format.mimetypeapplication/pdf
dc.subjectcloud computing
dc.subjectgrid computing
dc.subjectmedical image processing
dc.subjectApache Hadoop ecosystem
dc.subjectBig data infrastructure
dc.titleAlgorithmic Enhancements to Data Colocation Grid Frameworks for Big Data Medical Image Processing
dc.typedissertation
dc.contributor.committeeMemberDouglas C. Schmidt
dc.contributor.committeeMemberAlan Tackett
dc.contributor.committeeMemberHongyang Sun
dc.type.materialtext
thesis.degree.namePHD
thesis.degree.leveldissertation
thesis.degree.disciplineComputer Science
thesis.degree.grantorVanderbilt University
local.embargo.terms2018-11-27
local.embargo.lift2018-11-27
dc.contributor.committeeChairBennett A. Landman
dc.contributor.committeeChairAniruddha Gokhale


Files in this item

Icon

This item appears in the following Collection(s)

Show simple item record