Download Dialogue DataGrid Motivating Applications

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts
no text concepts found
Transcript
Dialogue DataGrid
Motivating Applications
Joel Saltz, MD, PhD
Chair Biomedical Informatics
College of Medicine
The Ohio State University
Dialogue DataGrid
• Relational databases, files, XML databases,
object stores
• Strongly typed
• Multi-tiered metadata management system
• Incorporates elements from OGSA-DAI,
Mobius, caGrid, STORM, DataCutter, GT4 …
• Scales to very large data, high end
platforms
Requirements
• Support or interoperate with caGrid,
eScience infrastructure
• Interoperate with or replace SRB
• Well defined relationship to Globus Alliance
• Services to support high end large scale
data applications
• Design should include semantic metadata
management
• Well thought out relationship to commercial
products (e.g. Information Integrator,
Oracle)
Motivating Application Class I: Phenotype
characterization
• Information heterogeneity,
data coordination and data
size
• Synthesize information from
many high throughput
information sources
• Sources can include multiple
types of high throughput
molecular data and multiple
imaging modalities.
• Coordinated efforts at multiple
sites
• Detailed understanding of
biomedical phenomena will
increasingly involve the need
to analyze very large high
resolution spatio-temporal
datasets
Structural Complexity
Example Questions (Phenotypes
associated with Rb knockouts)
1.
What are the
mechanisms of fetal
death in mutant mice?
2.
What structural
changes occur in the
placenta?
3.
How different are the
structural changes
between the wild and
mutant types?
4.
…
Rb+
Rb-
Dataset Size: Systems Biology
Future big science animal
experiments on cancer,
heart disease, pathogen
host response
Basic small mouse is 3 cm3
1 μ resolution – very roughly
1013 bytes/mouse
Molecular data (spatial
location) multiply by 102
Vary genetic composition,
environmental
manipulation, systematic
mechanisms for varying
genetic expression; multiply
by 103
Total: 1018 bytes per big
science animal
experiment
Now: Virtual Slides
(roughly 25TB/cm2 tissue)
Compare phenotypes of normal vs Rb
deficient mice
Slides/Slices
Alignment
Placenta
Visualization
Segmentation
Computational Phenotyping
Challenges
• Very large datasets
• Automated image analysis
• Three dimensional reconstruction
• Motion
• Integration of multiple data sources
• Data indexing and retrieval
Large Scale Data Middleware Requirements
• Spatio-temporal datasets
• Very large datasets
– Tens of gigabytes to 100+ TB data
• Lots of datasets
– Up to thousands of runs for a study are possible
• Data can be stored in distributed collection
of files
• Distributed datasets
– Data may be captured at multiple locations by multiple
groups
– Simulations are carried out at multiple sites
• Common operations: subsetting, filtering,
interpolations, projections, comparisons,
frequency counts
Very Large Dataset Hardware is Proliferating
LinTel boxes (PvFS/
Active Disk Archive) (20)






D V D

D V D
(2)
890 MB/s through
MetaData Servers
(2)

D V D

D V D
(2)
(2)
890 M
B/s Th
rough
put

D V D
(2)
)
(2
(2)
D V D
DVD
DVD
DVD
DVD





DVD
DVD
DVD
DVD
DVD





DVD
DVD
DVD
DVD
DVD





DVD
DVD
DVD
DVD
DVD
(40 - 2 per xSeries)
10 GB/s
)
(2

DVD
(40 - 2 per T600)
384 MB/s throughput
put
r)
Cisco Directors 9509
ve ut
er hp
r s oug
e
p thr
4
(4)
6 B/s MB/s throughput
(1 M772
0
(4)
89
(4)
772 MB/s throughput
FAStT600 Turbo (20)
Scratch / Archive Storage Pool (310/420 TB)
(4)
772 MB/s throughput
(4)
772 MB/s throughput
SAN Volume Controller
(4 servers)
FAStT900 (4)
Core Storage Pool (35/50 TB) with SAN.FS
Backup Storage
3584 Tape
1 L32 2 D32
Actual: 640 cartridges @ 200
GB for a total of 128 TB
4 drives
max drive data rate is 35 MB/s
• 50 TB of performance
storage
– home directories, project
storage space, and longterm frequently accessed
files.
• 420 TB of
performance/capacity
storage
– Active Disk Cache compute jobs that require
directly connected storage
– parallel file systems, and
scratch space.
– Large temporary holding
area
• 128 TB tape library
– Backups and long-term
"offline" storage
Our Example: Ohio Supercomputing Center Mass Storage
Testbed
STORM Services
• Query
• Meta-data
• Indexing
• Data Source
• Filtering
• Partition
Generation
• Data Mover
STORM Results
Seismic Datasets
10-25GB per file.
About 30-35TB of Data.
STORM I/O Performance
4500
4000
Bandwidth (MB/s)
3500
3000
2 Threads
2500
4 Threads
2000
Max
1500
1000
500
0
1
2
4
# XIO nodes
8
16
Motivating Application II: caBIG In vivo
Imaging Workspace
Testbed
• Study the effects of image acquisition and reconstruction
parameters (i.e. slice thickness, reconstruction kernel and
dose) on CAD and on human ROC.
– use multiple datasets and several CAD algorithms to
investigate the relationship between radiation dose and
nodule detection ROC.
• Cooperative Group Imaging Study Support
– Children’s Oncology Group: quantify whether perfusion
study results add any additional predictive value to the
current battery of histopathological and molecular studies
– CALGB: Grid based analysis of PET/CT data to support
phase I, II studies
– NTROI: Grid based OMIRAD -- registration, fusion and
analysis of MR and Diffusive Optical Tomography (DOT).
CAD Testbed Project RSNA 2005 (joint with
Eliot Siegel et al at Univ. Maryland)
• Expose algorithms and data management as Grid Services
• Remote execution of multiple CAD algorithms using
multiple image databases
• CAD algorithm validation with larger set of images
• Better research support — recruit from multiple institutions,
demographic relationships, outcome management etc.
• Remote CAD execution - reduced data transfer & avoid
need to transmit PHI
• CAD compute farms that reduce the turnover time
• Scalable and open source — caBIG standards
Architecture
Image Data Service
•Expose data in DICOM PACS with grid service
wrappers
•An open source DICOM server — Pixelmed
•XML based data transfer
5 Participating Data Services
3x Chicago
1x Columbus
1x Los Angeles
CAD Algorithm Service
• Grid services for algorithm invocation and image retrieval service
• caGRID middleware to wrap CAD applications with grid services
• Report results to a result storage service
caGrid Introduce
facilitates service
creation
GUMS/CAMS is used
to provide secure data
communication and
service invocation
CAD algorithms provided by iCAD and Siemens Medical Solutions.
Prototypes for investigational use only; not commercially available
Framework Support Services
• Result storage server — A distributed XML database for caching
CAD results
• GME — Manage communication and data protocols
User Interface
Available
data
services
DICOM
image viewer
17
5
14
12
Queried
results
18
15
16
Slice = 127
W/L = 2200/-500
Click to browse images, submit
CAD analysis, and view results
Motivating Application III: Integrative
Cancer Biology Center on Epigenetics (PI
Tim Huang, OSU)
• TGFβ/Smad4 targets are regulated via
epigenetic mechanisms. Promoter
hypermethylation is a hallmark of this
epigenetically mediated silencing. In
this study, we have combined both
chromatin immunoprecipitation
microarray (ChIP-chip) and differential
methylation hybridization (DMH) to
investigate this epigenetic mechanism
in ovarian cancer
Translating a goal into workflow
n9
n1
n8
K
Data Mining
(ex: clustering
Literature Survey
- Experimentally verified
TGF-B target genes
- Housekeeping Genes
Clinical data
Data Collection-Genome
-UCSC Genome
-BLAT alignment
Analytical service J
Data source K
n10
Construct KbTSMAD
(Knowledgebase of the
TGF-B/SMAD signaling pathway)
L
n7
Normilization
with statistical tools
Experimental Results
From Other Groups
A
Analytical service I
KbTSMAD
Data source L
Data source A
n2
n3
B
C
Description of
Experiment
Chip-chip Results
(Microarray data)
D
E
Description of
Experiment
DMH Results
(Microarray
Data)
DMH Experiment
Chip-on-chip Experiment
Data source B and C
Data source D and E
n5
n4
Chip design
G
Analytical service F
Custom chip design info
(e.g. from Agilent)
Data source G
n6
H
Candidate cutting Enzyme
Information
Data source H
ArrayAnnotator
Application of caGrid to the
workflow
• Application needs to support access to a diverse set of
databases and execution of different analysis methods
• Data services
–
–
–
–
–
–
KbSMAD
Chip information from chip company
Enzyme data
Clinical data
Experimental results
Experimental design
• Analytical services
– Designing a custom array
– Normalization
– Data mining (ex: clustering)
Example: Prototype of Clone
Annotation Analytical Service
• Analytical Service: ArrayAnnotator
 Goal: Provide a annotation for each clone to select a subset of clones
among 400,000 candidate clones to design a custom array for DMH
experiment
 Clone selection criteria




Clones within a promoter region
Clones with proper internal and external cuts
Clones within CpG island region and/or high CG contents
Clones with Transcription Factor binding sites
 Input: CloneType information
 extended sequence, enzyme info, genomic location, etc
 Functions
• Determine external cut locations around a clone region (e.g., cut-site by BfaI)
• Examine the internal cut around a clone region (e.g., cut-site by HapII, HinpII,
and MCrBc)
• Identify the location of clone in genome
• Show ether it is within promoter region or not
• Calculate CG content and overlapping with CpG islands
• Identify which Transcription Factor binding sites are overlapped with clones
Example caGrid Usage in P50 chip design
application
Query (geneId)
Result: List of clones
Clone Info
Data Services
1
2
Genome Sequence
Data Source
Result: extended genome sequence
4 of clone
3 Query
Annotation
Analytical Service
5
Request (cloneInfo)
6
Chip design application
Result: annotation (CpG, cutsite,
promoter region, etc)
ArrayAnnotator output (Hao Sun, Ramana
Davuluri)
Multiscale Laboratory Research Group
Ohio State University
Joel Saltz
Gagan Agrawal
Umit Catalyurek
Dan Cowden
Mike Gray
Tahsin Kurc
Shannon Hastings
Steve Langella
Scott Oster
Tony Pan
DK Panda
Srini Parthasarathy
P. Sadayappan
Sivaramakrishnan (K2)
Michael Zhang
The Ohio
Supercomputer Center
Stan Ahalt
Jason Bryan
Dennis Sessanna
Don Stredney
Pete Wycoff
Microscopy Image Analysis
• Biomedical Informatics
– Tony Pan
– Alexandra Gulacy
– Dr. Metin Gurcan
– Dr. Ashish Sharma
– Dr. Kun Huang
– Dr. Joel Saltz
• Computer Science and
Engineering
– Kishore Mosaliganti
– Randall Ridgway
– Richard Sharp
●
Pathology
–Dr.
Dan Cowden
Human Cancer
Genetics
●
–Pamela
Wenzel
–Dr.
Gustavo Leone
–Dr.
Alain deBruin
caGrid Team
• National Cancer
Institute
– Peter Covitz
– Krishnakant
Shanbhag
• Ohio State
University
– Shannon
Hastings
– Tahsin Kurc
– Stephen
Langella
– Scott Oster
– Joel Saltz
• SAIC
–
–
–
–
–
Tara Akhavan
Manav Kher
William Sanchez
Ruowei Wu
Jijin Yan
• Booze | Allen |
Hamilton
– Manisundaram
Arumani
• Panther
Informatics Inc.
– Nick Encina
– Brian Gilman
RSNA 2005 Team
Tony Pan, Stephen Langella, Shannon Hastings, Scott Oster,
Ashish Sharma, Metin Gurcan, Tahsin Kurc, Joel Saltz
Department of Biomedical Informatics
The Ohio State University Medical Center, Columbus OH
Eliot Siegel, Khan M. Siddiqui
University of Maryland School of Medicine, Baltimore, MD
Thanks to Siemens, ICAD for supplying CAD algorithms