Abstracts - BCatS - Stanford University

Transcription

Abstracts - BCatS - Stanford University
bcats
2005
biomedicalcomputationatstanford
abstractbook
bcats 2005 is brought to you by
stanford center for bio-medical computation • nih/nibib • bio-x
apple computer • genentech • alloy ventures • roche • agilent technologies
life sciences society • wilson sonsini goodrich & rosati • bionexus
stanford office of technology licensing • enterprise • stanford bookstore
biomedical • computation • at • stanford
Bio-X is a Stanford University program supporting
interdisciplinary work related to biology. The program
is a joint project of the Schools of Humanities and
Sciences, Engineering, Earth Sciences, and
Medicine. Physics, chemistry, engineering, and
computer science are combined in approaching
important basic and applied problems in
biology and medicine. Bio-X is designed to build
community and communication.
The Program is facilitated by a new Center
(The James H. Clark Center), which was completed
in Summer of 2003 thanks to the enormous
generosity of Jim Clark and Atlantic Philanthropies.
The Clark Center comprises the equipment,
resources and utilities required to conduct
breakthrough research at the cutting edge of
engineering, science and medicine.
For more information on Bio-X,
visit our website at:
http://biox.stanford.edu
2
october • fifteen • 2005
A Winning Formula
in Drug Discovery
At Roche Palo Alto, we are dedicated to discovering and developing new
medicines that improve the lives of people affected with serious illnesses such as
HIV/AIDS, Hepatitis C, Alzheimer’s disease and arthritis.
Our Global Research Informatics Group is enabling the search for new medicines
by providing information- and technology-based solutions to researchers
around the world. Our areas of expertise include Bioinformatics, Discovery IM,
Information Science, Infrastructure, NCD IM, and Scientific Computing.
Our park-like campus offers a wide array of professional and personal amenities
designed to promote a healthy work-life balance. Our employees enjoy all the
benefits of the San Francisco Bay Area, from access to leading academic and
research institutions to beautiful year-round weather and a diverse intellectual
community. We are pleased to provide a competitive compensation program,
with generous vacation and holiday schedules, a comprehensive relocation program,
and retirement and pension plans. To learn more about career opportunities at
Roche Palo Alto, please visit our website at http://paloalto.roche.com.
Roche is a proud sponsor of BCATS!
As an equal opportunity employer, we are committed to workforce diversity.
3
biomedical • computation • at • stanford
CHARTER Membership fees: ONLY US $60.00 (professionals, students, postdocs)
About the Life Sciences Society
The Life Sciences Society (LSS) was officially inaugurated on August 10, 2005 as an unaffiliated, non-profit
society. World-renowned scientists support the LSS in its mission and goals.
MISSION
To promote the integration of life sciences with computing, mathematics, and engineering and
to encourage new benefits through interdisciplinary advances involving biological and medical applications.
GOALS
The goal of the Life Sciences Society is to be the premier organization for scientists and
engineers in the life sciences by providing the venue, tools and opportunity to expand the reach and focus of
their individual disciplines into life sciences. The Life Sciences Society will provide tutorials, workshops,
conferences, publications and community participation.
Get involved and make a difference !
Join LSS today and help to establish LSS as the premier organization for the life sciences.
Membership Benefits
•
•
•
•
•
•
•
•
Discounts on all LSS activities and at the LSS BioShop
Complimentary mailbox at [email protected]
Complimentary subscriptions to The Scientist
Discounts on Morgan Kaufmann publications and Springer books (other discounts t.b.a.)
Discounts for LSS online tutorial streaming tapes and DVDs
LSS conference / workshop registration discounts
Free Life Sciences Review articles by leading scientists
Professional network, promoting your career, research and business goals
Join online:
http://LifeSciencesSociety.org/
Visit our table at the BCATS 2005 Symposium
4
For more information contact: Vicky Markstein, LSS President ([email protected])
october • fifteen • 2005
Welcome
to the sixth annual symposium on
Biomedical Computation at Stanford (BCATS).
This student-run one-day symposium provides an interdisciplinary forum
for Stanford students and post-docs to share their research in widelyvarying fields involvoing biomedical computation. Since its inception
in 1999, BCATS has seen growth and change in the field of biomedical
computation. This year’s schedule features diverse, cutting-edge research
from the largest applicant pool ever.
We thank our keynote speakers, student presenters,
judges, sponsors, and all 2005 attendees.
The BCATS 2005 organizing committee
Maureen Hillenmeyer, Biomedical Informatics
Lior Ron, Graduate School of Business
Lucinda Southworth, Biomedical Informatics
Ryan Spilker, Biomechanical Engineering
Beverly Tang, Biomechanical Engineering
5
biomedical • computation • at • stanford
BCATS 2005 Schedule
8.00 am
Onsite Registration and Breakfast
Sequoia Plaza
8.45
Opening Remarks
Hewlett 200
Scientific Talks Session I
Hewlett 200
9.00
Eran A. Mukamel
Network structure and the biological representation of
time in cerebellar neural circuits
9.15
R. Dana Carpenter
Computer simulation of bone cross-sectional
morphogenesis
9.30
(page 17)
Omar David Perez
Derivation of causal protein signaling networks from
multiparameter single-cell data
9.45
(page 16)
(page 18)
Tianfang Li
Radiation dose reduction in 4D computed tomography
(page 19)
10.00
Poster Session I (even-numbered posters)
Packard Lobby
11.00
Keynote Address
Hewlett 200
Jeffrey Weiss, PhD
(pages 8–9, 29–87)
(page 12)
Verification, Validation, and Sensitivity Studies in
Computational Biomechanics
11.45
Lunch
Sequoia Plaza
12.30p
Poster Session II (odd-numbered posters)
Packard Lobby
1.30
Keynote Address
Hewlett 200
Leroy Hood, MD, PhD
Computational Challenges in Systems Biology
6
(pages 6–9, 29–77)
(page 13)
october • fifteen • 2005
Scientific Talks Session II
2.15
Irene Elisabeth Vignon-Clementel
Impedance outflow boundary conditions for threedimensional patient specific modeling of blood flow
2.30
(page 20)
Serban Nacu
Foraging Regulation in Harvester Ants
2.45
Hewlett 200
(page 21)
Daniel Bruce Ennis
Visualization and Analysis of Diffusion Tensor Magnetic
(page 22)
Resonance Images using Orthogonal Tensor Invariants
3.00
Vishal Vaidyanathan
Modeling the aggregation of Alzheimers peptide Aβ21-43
(page 23)
using Molecular Dynamics and Distributed Computing
3.15
Break
Scientific Talks Session II
3.30
Pan-Chi Lee
(page 24)
Optimizing Dialysis
3.45
Rachel Weinstein
(page 25)
Pre-stabilization for Rigid Body Articulation with
Contact and Collision
4.00
Balaji Srinivasan
(page 26)
Integrated Protein Interaction Networks for 230
Microbes
4.15
David Rodriguez-Gomez
(page 27)
Efficient Free Energy Calculation of Peptide Insertion in
Membranes
4.30
Industry Keynote Address
Thomas Wu, MD, PhD
Hewlett 200
(page 14)
Bioinformatics at Genentech
5.00 pm
Awards and closing
Hewlett 200
7
biomedical • computation • at • stanford
BCATS 2005 Posters
poster
page
0
Abhay Sukumaran
Distributed cluster computing: a method for parallel simulation of large molecular
systems
30
1
Aditya Mandayam
tr0m
31
2
Andrew Holbrook
Segmentation of costal cartilage in lower abdomen CT data using watershed
markers
32
3
Alex Ten Eyck
Modeling biological tissue with nonconforming finite elements
33
4
Alain Laederach
Reconstructing RNA folding pathways from local probes of macromolecular
structure
34
5
Alex Pang
Prediction of venous obstruction following transvenous pacemaker placement: a
computer simulation approach
35
6
Alison L. Marsden
A computational study of energy efficiency and pressure losses in the total
cavopulmonary connection
36
7
Artit Wangperawong
Effects of size and hydrophobicity on single-walled carbon nanotubes
37
8
Andreas Sundquist
A strategy for whole-genome sequencing and assembly with high-throughput,
short-read technologies
38
9
Rebecca Maria Currano
Video-audio mapping of the physical environment for blind navigation
39
10
C. Alberto Figueroa
A new formulation to model blood flow and vessel motion in large, patient-specific
models of the cardiovascular system
40
11
Chase Garrett
Yarbrough
Application of a three dimensional image-based modeling technique to the circle of
41
Willis
12
Christopher B. Egner
SchemaTrans: automating the translation of arbitrarily formatted data to relational
databases
42
13
Christopher Enedah
Robotic perception of the skin’s mechanical properties - a validation study
43
14
Cesar A. Rodriguez
A theoretical model that complements an experimental system for quantitative
investigation of in vivo protein-protein interactions
44
15
Chuong B. Do
CONTRAlign: a discriminative framework for protein sequence alignment
45
16
Christopher Sewell
Metrics for a mastoidectomy simulator
46
17
Christopher D. Snow
Bringing the ribosome exit tunnel to life with distributed computing: a first look at
the thermodynamics of confined solvent and the nascent chain
47
18
Chand T. John
From biomedical images to geometric models
48
19
David Jing-Piao Lin
Algorithmic analysis of the quantitative distribution of neural stem cells in malignant
49
glioma
20
Diane Schroeder
Locating single-stranded regions of the human genome
50
21
Eduard Schreibmann
Image interpolation in 4D CT using a Bspline deformable registration model
51
22
Erich Konrad Elsen
GROMACS on the GPU
52
23
Tim A. Fonte
3-D simulations of hemodynamic factors in pulmonary hypertension
53
24
Gal Chechik
Filling missing components in yeast metabolic pathways using heterogeneous data
54
25
Gunnar F. Schroeder
Simulation of fluorescence anisotropy experiments: probing protein dynamics
55
26
Haidong Wang
Efficient prediction of protein-protein interactions using Markov networks
56
27
John Patrick Adams
Framework for identification of allosteric networks
57
8
october • fifteen • 2005
poster
page
28
Jessica Ebert
Unsupervised Identification of Functionally Relevant Residues in Protein Families
58
29
Jessie Dale
Tenenbaum
A modified Bayesian approach to determination of causation in differential gene
expression
59
30
Jason A. Hackney
Use of a naive Bayesian classifier to identify gene regulatory networks in the
protozoan parasite, Enatmoaba histolytica
60
31
John F. LaDisa, Jr.
Patient-specific modeling of altered hemodynamics and arterial deformation induced
61
by coarctation of the aorta and subsequent stent implantation
32
Joseph Minhow Chan
Preferential coplanarity of RNA helices and its impact on RNA activity
62
33
Magda Anna Jonikas
Using coarse-grained simulations of RNA to explore intermediate folding pathways
63
34
Justus Roos
Quantitative and qualitative assessment of a new knowledge-based algorithm for
centerline restoration through femoro-popliteal artery occlusions in peripheral CT
angiography (CTA)
64
35
Jacob Michael Zahn
A common signature for human aging
65
36
Hyun Jin Kim
Chronic spinal cord injury modifies aortoiliac biomechanical forces and morphology
and increases abdominal aortic aneurysm disease risk
66
37
Jan Lipfert
Novel approaches to simulating amyloid formation: Reaction path annealing and
single molecule pulling
67
38
Michael Yu Zhang
Genomic instability signatures of colorectal carcinoma subtypes by hierarchical
clustering of gene copy data
68
39
Nina Singhal
Automatic state decomposition for constructing Markov models for macromolecular
dynamics
69
40
Padma Sundaram
Fold removal in CT colonography
70
41
Patrick Timothy
McGrath
Analysis of the transcriptional regulation of a bacterial cell cycle
71
42
Qiyu Peng
2D ultrasound image processing in identifying displacement history of urogenital
structures to pelvic floor muscle activity
72
43
Qiyu Peng
A novel palpation system to evaluate internal organs of genitourinary system
73
44
Rashmi Raghu
Viscoelastic arterial wall model for one-dimensional modeling of blood flow and
pressure
74
45
Rebecca Meredith
Loew
Differential gene expression between healthy and melanoma patients for T cells, B
cells and NK cells
75
46
Ron Y. Kwon
Mechanics of deformation of bone cells
76
47
Ryan J Tibshirani
A novel software for plotting multidimensional FACS data
77
48
Shaohua Sun
Registration of lung nodules using a semi-rigid model
78
49
Anthony Sherbondy
A Metropolis sampling framework for inferring neural connectivity in Diffusion Tensor
79
Images
50
Shirley Wu
Modeling of phosphorylation sites in protein structures using the FEATURE system
80
51
Su-In Lee
Identifying regulatory mechanisms associated with genetic variations among
individuals
81
52
Samuel S. Gross
CONTRAST: de novo gene prediction using a semi-Markov conditional random field
82
53
Brian Thorndyke
Removal of respiratory motion artifacts and the construction of 4D PET
83
54
Vincent Bangping Chu Ion size effects and the Poisson-Boltzmann equation
84
55
Jing Xu
85
Three-dimensional correlation analysis of the macromolecules on synaptic vesicles
9
biomedical • computation • at • stanford
10
october • fifteen • 2005
The Stanford Center for Biomedical Computation
is pleased to sponsor BCATS.
Executive Committee:
Russ Altman — [email protected]
Scott Delp — [email protected]
David Paik — [email protected]
The Center for Bio-Medical Computation (CBMC) and its affiliated faculty group, Biomedical
Information Technology at Stanford (BITS), is an inter-connected, cross-disciplinary group of
researchers who develop, share, and utilize computer graphics, scientific computing, medical imaging, and biomechanical modeling applications in biology, bioengineering, and
medicine. Our mission is to establish a world-class biomedical computing and visualization center at Stanford that will support joint initiatives between the Schools of Engineering,
Medicine, and Humanities and Sciences.
Participating labs promote the efficient development of new courses, programs, computational models, and tools that can be used in classrooms, clinical practice, and the biomedical
research community. Our goal is to become be an international resource for partners in the
biotechnology, biomedical device, computing, medical imaging, and software industries.
The CMBC and BITS faculty support teaching and training in the biomedical computing sciences and the creation of interdisciplinary biocomputational courses at the undergraduate,
graduate, and post-graduate levels, both on-campus and at remote sites.
cbmc.stanford.edu
Genentech is a Proud Sponsor of the BCATS Symposium
Genentech
IN BUSINESS FOR LIFE
www.gene.com
11
biomedical • computation • at • stanford
BCATS Keynote Speaker
Jeffrey Weiss, PhD
University of Utah
Departments of Bioengineering and Orthopedics,
Scientific Computing and Imaging Institute,
University of Utah
http://hodad.sci.utah.edu/~weiss/mrl
Jeff Weiss is an Associate Professor of Bioengineering, an Adjunct Associate Professor of Orthopedics and a Faculty Member in the Scientific Computing and Imaging (SCI) Institute at the
University of Utah. He grew up in Southern California and received the B.S. (1989) and M.S.
(1990) degrees in Bioengineering from the University of California, San Diego. He completed
his Ph.D. in Bioengineering at the University of Utah in 1994. Dr. Weiss received postdoctoral
training in computational biomechanics as a finite element analyst working with the Applied
Mechanics and Methods Development Groups at Lawrence Livermore National Laboratory.
Research in the Weiss Laboratories has historically focused on the biomechanics and healing
of musculoskeletal soft tissues, in particular the ligaments of the knee. Over the past five years,
the research focus has expanded considerably to include hard tissue as well as cardiovascular
tissues including the heart, coronary arteries and smaller vessels involved in angiogenesis.
A central theme of the approach to research is the use of multidisciplinary techniques. A
primary strength continues to be the combination of both experimental and computational
techniques, especially in the area of biomechanics, to address research questions that would
be impossible to investigate using approaches from a single technical discipline. Major areas
of research focus include subject-specific computational modeling of hip, knee and shoulder
biomechanics, structure-function relationships in connective tissues, mechanical aspects of
angiogenesis, and the direct integration of image data into biomechanical analyses.
Professor Weiss routinely serves on study section panels for the NIH and NSF and is currently an
Associate Editor for the Journal of Biomechanical Engineering, the Journal of Applied Biomechanics and Computer Methods in Biomechanics and Biomedical Engineering. He is a regular
reviewer for 15 additional international journals. Dr. Weiss has received a number of highly
coveted awards, including the Taylor & Francis prize for “outstanding innovation in computer
methods in biomechanics & biomedical engineering”, the ASME YC Fung Young Investigator
Award in May 2002, a NSF CAREER Award in February 2002 and a Whitaker Foundation Research Grant in March 1995.
12
october • fifteen • 2005
BCATS Keynote Speaker
Leroy Hood, MD, PhD
Institute for Systems Biology
President, Institute for Systems Biology
M.D., Johns Hopkins School of Medicine, 1964
Ph.D., Biochemistry, California Institute of Technology, 1968
Dr. Hood’s research has focused on the study of molecular immunology, biotechnology, and
genomics. His professional career began at Caltech where he and his colleagues pioneered
four instruments—the DNA gene sequencer and synthesizer, and the protein synthesizer and
sequencer—which comprise the technological foundation for contemporary molecular biology. In particular, the DNA sequencer has revolutionized genomics by allowing the rapid automated sequencing of DNA, which played a crucial role in contributing to the successful mapping of the human genome during the 1990s. In 1992, Dr. Hood moved to the University of
Washington as founder and Chairman of the cross-disciplinary Department of Molecular Biotechnology. In 2000, he co-founded the Institute for Systems Biology in Seattle, Washington to
pioneer systems approaches to biology and medicine. Most recently, Dr. Hood’s lifelong contributions to biotechnology have earned him the prestigious 2004 Association for Molecular
Pathology (AMP) Award for Excellence in Molecular Diagnostics and the 2003 Lemelson–MIT
Prize for Innovation and Invention. He was also awarded the 2002 Kyoto Prize in Advanced
Technology and the 1987 Lasker Prize for his studies on the mechanism of immune diversity. He has published more than 600 peer-reviewed papers, received 14 patents, and has coauthored textbooks in biochemistry, immunology, molecular biology, and genetics, and is a
member of the National Academy of Sciences, the American Philosophical Society, the American Association of Arts and Sciences, and the Institute of Medicine. Dr. Hood has also played a
role in founding numerous biotechnology companies, including Amgen, Applied Biosystems,
Systemix, Darwin and Rosetta.
13
biomedical • computation • at • stanford
BCATS Industry Keynote Speaker
Thomas Wu, MD, PhD
Genentech
Dr. Wu is currently a Senior Scientist in the Department of Bioinformatics at Genentech, Inc.,
where he serves as group leader for microarray data analysis. His work also involves the development of bioinformatics tools for genomic sequence analysis, including the programs GMAP
and PMAP.
Prior to working at Genentech, he obtained his B.S. and M.S. degrees from Stanford University, M.D. from Harvard Medical School, and Ph.D. in computer science from the Massachusetts Institute of Technology. He completed his internship and residency in internal medicine
from Stanford University. He subsequently worked as a postdoctoral fellow in the Department
of Biochemistry at Stanford University, where he helped to develop the functional genomics
tools EMOTIF and EMATRIX, as well as methods for the superposition of multiple protein structures.
His awards include a Physician Postdoctoral Fellowship from the Howard Hughes Medical Institute; George M. Sprowls Award from M.I.T.; Martin Epstein Award from the American Medical Informatics Association, Frederick E. Terman Award from Stanford; Laureate Award from the
Tau Beta Pi engineering society; and election to Tau Beta Pi, Sigma Xi, and Phi Beta Kappa.
14
october • fifteen • 2005
BCATS Talk
Abstracts
15
biomedical • computation • at • stanford
session I
9:00am
Eran A.
Mukamel
Network structure and the
biological representation of
time in cerebellar neural
circuits
Purpose
Behaviors from language perception and production to learning conditioned Pavlovian reflexes require neural circuits that encode and retain information about the timing of stimuli. The cerebellum is critical for learning
timed behaviors. Timing information is likely generated within the cerebellum,
and may be an emergent property of the large network of granule (GR) and Golgi (GO)
cells in the input layer. However, neural mechanisms for the generation of behaviorally relevant timing information are not known. Proposals to date included sparse representation and coherent oscillation, both
of which rely on individual GR cells to provide well-timed signals. Here we propose a model for distributed
encoding of timing information in cerebellar GR cell layer.
Mark J. Schnitzer
Methods
The cerebellar cortex is a crystal-like circuit of four or five neuronal cell-types. The GR axons extend several mm
along the cerebellar folia. This quasi-1D structure determines the dynamics of the recurrent net of excitatory
GRs and inhibitory GOs.
We studied the dynamics of a firing rate net with cerebellar structure, as well as a similar control net with randomly oriented GR axons. We used analytic theory and simulation to study the normal modes of each type of
net, and derived the distribution of oscillation frequencies from structural factors such as the orientation and
length of GR axons.
Results
Our main finding is that the number of distinct modes is limited by the population of the minority cell-type,
the GO. Due to the ratio of 10,000 GRs to each GO, this result implies that the vast number of GRs is not responsible for the fine temporal representation of sensory information.
A second result is that the proposed distributed mechanism for representing temporal information provides a
biologically plausible alternative to the single-cell based mechanisms of sparse encoding and coherent oscillation. Such a distributed scheme is a natural emergent property of the crystal-like network geometry.
Conclusion
Our results suggest that the organization of GR axons parallel to each other may be critical to create a broad
spectrum of oscillatory modes in the cerebellar input layer, and hence for representing a wide range of time intervals. We predict that manipulations that disrupt the geometrical structure of the cerebellar network would
narrow the range of stimulus frequencies represented by network dynamics. This could lead to behavioral
deficits, such as mistimed responses following training in classical conditioning experiments.
References
1. R.B. Ivry and R. M. Spencer, Curr Op Neurobio 14, 225 (2004)
2. M.D. Mauk and D. V. Buonomano, Ann Rev Neurosci 27, 307 (2004)
3. B. Ermentrout, Rep Prog Phys 61, 353 (1998)
16
october • fifteen • 2005
Computer simulation of bone
cross-sectional morphogenesis
Purpose
Previous computational models of the development of bone cross-sections have used far-field loads such as joint reaction forces and muscle
forces as stimuli for bone growth [1,2]. However, numerous studies have
suggested that wrap-around tendons, adjacent musculature, and the fibrous layer of the periosteum can exert local bone surface pressures that
may also influence bone development [3-5]. In this study we incorporate
local periosteal surface pressures into a computational model of rat tibia morphogenesis.
session I
9:15am
R. Dana
Carpenter
Dennis R. Carter
Materials and Methods
The stimulus for new bone formation or bone resorption was calculated based on intracortical stresses
due to far-field bending and torsional loads, and the bone modeling response was modified based on
local periosteal surface loads imposed by adjacent muscles. Simulations began with a circular bone section representing the mid-shaft of the rat tibia at 17 days post-conception. Bending and torsional loads
were scaled with body mass data from growing rats [6]. Engineering beam theory was used to determine
bending stresses, and finite element software was used to determine shear stresses, principal moments
of inertia (properties proportional to bone bending stiffness), and the principal angle (θ1) corresponding
to the maximum moment of inertia (I1). A section with no surface loads and a section with surface loads
were grown to a simulated age of 17 months.
Results
The section that experienced no surface loading developed an elliptical geometry, and θ1 for this section
coincided with the direction of the applied bending moment. Application of periosteal surface pressures
resulted in a triangular geometry typical of the rat tibial diaphysis. The application of periosteal surface
pressures resulted in a 20% decrease in I1, and θ1 for this section did not coincide with the direction of
the applied bending moment.
Conclusions
The simulations performed in this study suggest that overall bone size is dictated by far-field loads and
that periosteal surface pressures can have large effects on bone cross-sectional morphology and the direction of a bone’s maximum bending stiffness. Including periosteal surface pressures imposed by adjacent muscles in simulations of bone morphogenesis presents a more realistic model of the in vivo loading
situation and subsequently produces a more realistic bone morphology.
References
[1] van der Meulen et al., 1993. Bone 14(4). [2] Levenston et al., 1998. Comput Methods Biomech Biomed
Engin 1(4). [3] Giori et al., 1993. J Orthop Res 11(4). [4] Teng and Herring, 1998. J Morphol 238(1). [5] Feik
et al., 1987. Br J Exp Pathol 68(6). [6] Donaldson, 1924. The Rat.
17
biomedical • computation • at • stanford
session I
9:30am
Derivation of causal protein
signaling networks from
multiparameter single-cell data
Purpose
Omar
To demonstrate the use of Bayesian networks for automated elucidation of
influences of protein signaling networks using simultaneous multiDavid Perez causal
variate measurements of phosphoproteins on single cells. To objectively
Karen Sachs
Dana Pe’er
Douglas Lauffenburger
Garry P. Nolan
derive inter-molecular influence connections, we applied a probabilistic
modeling algorithm to determine a graph of causal influences among
phospho-signaling molecules in human primary CD4+ T cells.
Materials and Methods
Simultaneosly measuring 11-phospho-epitopes by multiparameter flow
cytometry. Critical to the derivation of causal inference, was the use of
multiple independent perturbation events that aided inferring the direction of influence between the signaling components under study.
Results
The approach identified a majority of classically reported signaling relationships, and predicted several novel influence connections. A prediction of novel
inter-pathway crosstalk, causal influence of Erk on Akt, was confirmed experimentally.
Conclusion
The multivariate, simultaneous measures of biological states in cell populations allow for unprecedented
automatic construction of causal signaling network models in primary human cell subsets. Extension of
this concept will allow high throughput assembly of signaling network hierarchies in a variety of important biological systems.
References
Sachs K, Perez O, Pe’er D, Lauffenburger DA, Nolan GP. Causal protein-signaling networks derived from
multiparameter single-cell data.
Science. 2005 Apr 22;308(5721):523-9.
Perez OD, Nolan GP.
Simultaneous measurement of multiple active kinase states using polychromatic flow cytometry. Nat Biotechnol. 2002 Feb;20(2):155-62.
Perez OD, Krutzik PO, Nolan GP.
Flow cytometric analysis of kinase signaling cascades.
Methods Mol Biol. 2004;263:67-94.
18
october • fifteen • 2005
session I
9:45am
Radiation dose reduction in 4D
computed tomography
Purpose
Tianfang
Li
Four-dimensional (4D) CT scans, acquired synchronously with a respiratory signal, provide
not only the 3D spatial information, but also temporal changes of the anatomy as a function of the respiratory phase during the imaging, and can therefore be employed in 4D
treatment planning to explicitly account for the respiratory motion. However, it usually
delivers 10~15 times more radiation dose to the patient as compared to the standard
3D CT, since multiple scans at each couch position are required to obtain the temporal
information. In this work we propose a statistic method to obtain high quality 4D CT
with low tube current, hence reducing the radiation exposure of patients.
Eduard Schreibmann
Brian Thorndyke
Method/Materials
Gayle Tillman
When 4D CT data are acquired with low tube current, the quality of each individual
phase image is degraded because of the decrease in detected photon number. HowArthur Boyer
ever, the statistics can be dramatically improved if the time-resolved images of a patient are properly registered to a specific phase and superposed together. In our apAlbert Koong
proach, the registration of the patient’s image at different phases is accomplished by
using a finite element-based deformable registration model, which is capable of regis- Karyn Goodman
tering two arbitrary phases to within a spatial accuracy of 2mm. The improvement of the
signal-to-noise ratio (SNR) of the CT image at a given phase was then achieved by statistiL. Xing
cally stacking the registered images. In order to find the optimal trade-off between noise
and resolution, we developed a novel 4D penalized weighted least square (4D-PWLS) method
to smooth the data spatially and temporally. The method was validated by motion-phantom and
patient studies using a GE Discovery-ST PET/CT scanner. A Varian RPM respiratory gating system was used
to track the motion and to facilitate the phase binning of the 4D CT data. In both studies, the CT data were acquired in
cine mode, at 120 kVp. The x-ray tube angular velocity was set to 2 rotations/sec. The cine duration at each couch position was set 1 sec longer than the “respiration” period to 6.1 sec, and the cine interval between images was 0.45 sec. The
total acquisition time for the motion phantom was about 30 sec covered 10 cm in axial direction with 8x2.5 mm slice
thickness; for the patient, the total acquisition time was about 1.5 min, with axial coverage of 25 cm using 4x2.5 mm
slice thickness. The 4D scans were acquired twice for the phantom at current 10 mA and 100 mA, respectively, with all
other parameters kept the same. The SNRs of the two sets of images were compared. The clinical 4D-CT patient study
was performed once at 80 mA. The dose to the patient was 80.2 mGy. The cine images were sorted on a GE Advantage
4D workstation to generate phased images.
Results
We calculated the SNRs for both studies. The average SNR of 10 mA phantom images increased by more than three-fold
from 0.051 to 0.165 after the proposed 4D-PWLS processing, without noticeable resolution loss. The patient images
acquired at 80mA showed an increase from 2.204 to 4.558 for the end-inspiration phase, and from 1.741 to 3.862 for
the end-expiration phase, respectively. By examining the subtraction images before and after processing, good edge
preservation was also observed in the patient study.
Conclusions
We have developed a low-dose 4D CT protocol. Through deformable registration and statistical smoothing, the method
incorporates information of different phases and allows us to obtain high quality 4D-CT images with low x-ray tube
current. The protocol greatly reduces the patient radiation dose during 4D CT scan and may have important practical
implications.
19
biomedical • computation • at • stanford
Impedance outflow boundary
2:15pm
conditions for three-dimensional
patient specific modeling of
blood flow
Irene E.
VignonClementel
session II
Purpose
When modeling blood velocity and pressure fields, the assignment of appropriate boundary conditions is of critical importance. In three-dimensional problems,
constant pressure is typically assigned to outlet boundaries resulting in unrealistic flow splits and inaccurate pressure waveforms. This problem is especially acute
in fluid-solid interaction problems since the pressure pulse drives the vessel motion. To address this limitation, zero-dimensional, lumped parameter, models have
been utilized in recent years. However, the parameters are difficult to relate to patient specific data as they do not incorporate vasculature branching patterns or geometric data. Three-dimensional to one-dimensional iterative coupling methods have
been derived and used for simple geometries but this staggered approach may lead to
convergence issues for problems with multiple outlets. A third type of boundary condition
based on the impedance derived from a one-dimensional model of vascular trees using linear
wave propagation theory has been developed and applied to patient data but only as a boundary condition for the
one-dimensional equations of blood flow (see for example [1-2]).
C.A. Figueroa
K.E. Jansen
C.A. Taylor
Material and Methods
We have developed a general method to couple a three-dimensional numerical domain with various analytic models
of the downstream domains [3]. Here we apply this method to assign impedance outflow boundary conditions for
three-dimensional simulations of blood flow based on a downstream model with one-dimensional wave propagation
theory in distributed vascular networks.
Results
Patient-specific examples are presented where realistic flow splits and pressure fields are computed. Time-varying
pressure and velocity fields were quantified for a normal subject and a diseased patient, showing characteristics consistent with clinical observations that could not be modeled with previously described boundary conditions.
Conclusion
The use of one-dimensional impedance outflow boundary conditions with three dimensional numerical models enables the simulation of blood flow at physiologic pressure. This methodology is applicable to predicting changes in
blood flow in the large arteries under varying physiologic conditions affecting the modeled domain and downstream
vasculature. The main limitation of the concept of impedance resides in the assumption of a periodic solution. For
problems where transient effects need to be modeled, coupling between a three-dimensional domain and a nonlinear one-dimensional model of blood flow is a necessary extension of the current methodology.
References
[1] M.S. Olufsen et al., Numerical Simulation and Experimental Validation of Blood Flow in Arteries with StructuredTree Outflow Conditions, ABME, 28, 1281-1299, 2000.
[2] B.N. Steele et al., Fractal Network Model for Simulating Abdominal and Lower Extremity Blood Flow During Resting
and Exercise Conditions. Submitted to Comp. Meth. Biomech. Biomed. Eng., 2005.
[3] I.E. Vignon-Clementel et al., Outflow Boundary Conditions for Three-Dimensional Finite Element Modeling of Blood
Flow and Pressure in Arteries, Accepted in Comp. Meth. in App. Mech. and Eng., 2005.
20
october • fifteen • 2005
session II
Foraging regulation in
harvester ants
Purpose
An ant colony is an interesting example of a complex system. At the low level, a
colony consists of a large number of similar individuals, performing fairly simple
tasks, often with a random element. Yet at the high level, the colony is capable
of complex and robust behavior, achieved without central control. Information
is transmitted across the colony by interactions among ants, and the behavior
of individuals is strongly influenced by these interactions. Ants are also very
succesful creatures; it is estimated they make up at least 10% of the terrestrial
animal biomass. We have been studying a population of harvester ants living in
SE Arizona, focusing on trying to understand some of the laws that determine their
foraging behavior.
2:30pm
Serban
Nacu
Deborah Gordon
Susan Holmes
Materials and Methods
Harvester ants forage for seeds. On a typical morning there is a steady flow of foragers moving in and out of
the nest. We counted the number of ants in and out over ten second intervals. In one experiment, some of
the returning foragers were removed before reaching the nest. The colony was observed before and after the
removal, in order to estimate how it reacts to the removal. This was performed on multiple days and colonies.
Since some of the removed ants would have been otherwise seen again, the ant counts after the removal must
be adjusted upwards. The adjustment factors were computed from a related experiment, involving marking
individual ants in various colors and following them during the day. Computer simulations of the foraging
and the removal experiment were also performed.
Results
Foraging outflow decreased after the removal, even when adjusted for the removed foragers. There appears
to be a high level of randomness in the data: colony parameters do not seem to predict strongly how a colony
will react. One explanation for the flow decrease could be that ants base their decision to leave the nest on
encounters with returning ants. Preliminary results from computer simulations of this model appear to match
the results of the experiment.
Conclusion
Harvester ants respond to decreased forager inflow by sending fewer foragers outside. This may be an adaptive response to a potential external danger. In ongoing work, we are trying to understand in more detail the
low-level laws that cause this high-level behavior.
21
biomedical • computation • at • stanford
Visualization and analysis of
2:45pm
diffusion tensor magnetic
resonance images using
orthogonal tensor invariants
Daniel
Purpose
Diffusion tensor magnetic resonance imaging (DTMRI) is a technique that can
Bruce
map the three-dimensional orientation of tissue microstructure in the human
brain. The richness of spatially varying tensor data, however, requires specialEnnis
ized methods to visualize and quantify salient features. Diffusion tensors can be
session II
Gordon
Kindlmann
mathematically decomposed into orientation (eigenvectors) and shape (eigenvalues) information. The purpose of our work was to analytically derive a new
set of orthogonal tensor invariants (compact descriptors of tensor shape) and
use them to improve the visualization and analysis of DTMRI data. Orthogonality is a useful property of an invariant set, because it isolates the measurement of
variation in one physiological property from variations in another.
Materials and Methods
Theoretical extensions of the work by Criscione [1] in the field of strain energy functions resulted in the development of a spherical set of orthogonal tensor invariants. The invariant set is comprised of the tensor’s
isotropic norm, fractional anisotropy, and mode which respectively quantify the magnitude of isotropy, the
magnitude of anisotropy, and the mode of anisotropy. The inclusion of the tensor’s mode is useful for resolving an anisotropic region as linear-anisotropic, orthotropic, or planar-anisotropic.
Results
DTMRI data were obtained from a healthy volunteer who provided a signed statement of informed consent.
The theory was used to enhance the visualization of DTMRI data by implementing a colormap that encodes a
traditional fractional anisotropy map with the underlying tensor mode. Thus, areas of high fractional anisotropy (highly ordered structure) can be visually resolved as being linear- or planar-anisotropic, thereby distinguishing areas of purely fibrous structure from areas in which fibers cross or branch. Such maps also improve
the delineation of adjacent fiber tracks such as the corpus callosum and the cingulum bundle. Improved directionally encoded colormaps – maps that encode the direction of the primary eigenvector, often associated
with the local fiber direction – can also be generated. By incorporating tensor mode, areas of fibrous structure
are easily identified by enhanced color encoding of highly linear structures. Similarly, areas of planar anisotropy are de-emphasized as the primary eigenvector direction is indeterminant within a plane of rotation.
Supplemental material is available at www.stanford.edu/~dbe/BCATS_2005.
Conclusion
The theoretical development of orthogonal tensor invariant sets provides an improvement in the visualization and analysis of DTMRI data.
Reference
Criscione JC, et al. J. Mech. Phys. Solids 2000;48(12).
22
october • fifteen • 2005
Modeling the aggregation of session II
3:00pm
Alzheimers peptide Aβ21-43
using molecular dynamics and
distributed computing
Vishal
Purpose
Alzheimer’s Disease(AD) has been associated with the formation of fibril- Vaidyanathan
lar aggregates of the 42 residue long beta amyloid peptide (Aβ) in the
brain[1]. Recent studies have shown that the active neurotoxic agents
are in fact low molecular weight oligomeric species rather than fibrillar
forms[2]. Knowledge of the structure of these oligomeric species and the
mechanism by which they form could prove to be vital in the search for
cures for AD. However, despite significant research focus, the Aβ peptides
have proved to be extremely challenging to study experimentally and relatively little is known about their structure or aggregation mechanisms. In this
work we use computational methods to elucidate the mechanism of aggregation of Aβ21-43 and present insights from simulation on the possible structures of the
oligomers.
Nicholas William
Kelley
Vijay Pande
Materials and Methods:
Since the process of aggregation takes secods to hours at experimental concentrations(~10µM) and one can
only simulate nanoseconds per day by traditional methods, simulating aggregation would require 10 billion
years to complete. Clearly new methods are needed and consequently, we have developed a means to use
Markov models to extend the information obtained from short simulations to experimentally relevant time
scales. We build this model by combining analytic theory and large-scale distributed computing of the aggregation starting from the encounter complex. The encounter complex aggregation of four chains of Aβ21-43
was studied by running molecular dynamics simulations in explicit water at a concentration of 14mM. The
Folding@Home distributed computing project[4] was used to obtain nearly 100µs of aggregate simulation
time. By the combination of these techniques, we can now create a kinetic model of this process.
Results and Conclusions
Using our methods, we have constructed a kinetic model of Aβ aggregation, predicting the long timescale
behavior of oligomerization. In agreement with experimental data, we find that Aβ21-43 forms heterogenous
oligomers with a hydrophobic core comprising of C-terminal residues of the chains and relatively unstructured
solvent exposed N-terminal chains[3]. However, our results allows us to go beyond experiment to predict a
more detailed structure of the oligomeric state. This is the first computational prediction of the structure and
mechanism of the Aβ oligomerization from a kinetic model corresponding to experimental conditions. We
conclude with a brief discussion of how our model could be used to test the impact of small molecule drugs
on the predicted oligomeric state.
References
[1] Annu. Rev. Neurosci. 24, 2001, pp 1121-1159.
[2] Proc. Natl. Acad. Sci. U.S.A. 100(18), 2003
[3] Biochemistry 42(44), 2003, pp 12749 - 12760;
[4] Science 290(5498), 2000, pp 1903-1904.
23
biomedical • computation • at • stanford
session III
3:30pm
Pan-Chi
Lee
Stefanos A.
Zenios
Optimizing dialysis
Purpose
The dialysis population in the US (currently 400,000 patients) has seen unusually
high mortality (>20% annually) and a rapidly rising cost of care ($23 billion per
year). The population is also expected to grow six-fold over the next three decades. Yet owing to a lack of thorough analysis, practitioners and policy-makers
continue to disagree over the best approach to dialysis. As a result, patients are
dialyzed ineffectively and poor outcomes are observed.
Methods
We constructed a richly-detailed simulation model that generates realistic dialysis
outcomes (i.e., patient demographics, co-existing medical conditions, disease progression, hospitalization, dialysis treatment, transplantation, quality-of-life, treatment expenses, etc.). The
model can simulate over 200,000 possible patient profiles from incidence of disease through death and was
calibrated to several databases, some of which were not available until recently. We then posed an optimization problem (a Markov decision process) with dynamics driven by the simulation model and then solved it
using Machine Learning. Through this elaborate setup, we were able to identify the most cost-effective approach to dialysis.
Results
This optimized approach to dialysis entails varying dosage depending on the characteristics of each individual patient. When we simulated the entire American dialysis population under this optimized approach,
we found that cost savings on the order of $400 millions per year are possible compared to guidelines currently recommended by the National Kidney Foundation. Furthermore, we identified several bottlenecks that
handicap the existing approach to dialysis. Sensitivity analysis revealed the findings are not adversely affected
by sampling error in the simulation parameters.
Conclusion
We have shown that it is possible to analyze dialysis in its full complexity using simulation and computationally-intensive methods. More generally, the problem investigated here is the optimal initiation of an expensive
therapy. This problem is systemic in health care and important for policy setting and clinic management. Yet
robust methodologies to develop such strategies are not widely available, inhibiting the practical implementation. This research develops such methodologies utilizing the tools of mathematical modeling and computation.
References
Pastan, Stephen, James Bailey. 1998. Dialysis therapy. The New England Journal of Medicine, Vol. 338 1428–
1437.
Sutton, Richard S., Andrew G. Barto. 1998. Reinforcement learning: An introduction. MIT Press, Cambridge,
MA.
24
october • fifteen • 2005
Pre-stabilization for rigid body session III
articulation with contact 3:45pm
and collision
Purpose
Rachel
Modeling human motion accurately requires a robust method for dynamically
simulating joints. We developed a method for dynamic simulation of articulated
Weinstein
rigid bodies (ARBs) with unpredictable contact and collision. Many practitioners
solve these types of problems using reduced coordinate formulations. However,
unpredictable contact and collision can pose serious difficulties, and closed loops,
such as those occurring between the legs and ground during human motion, often require adding a nonlocal constraint to ensure loops remain closed. Furthermore, true dynamic simulation must be able to deal with unpredictable obstacles
requiring contact and collision handling. We found it easier to design algorithms that
treat closed loops and frequent contact and collision in maximal coordinates. Our approach
works with any method for specifying valid joint constraints, and requires no special considerations for closed
loops and/or branching.
Joseph Teran
Ron Fedkiw
Methods
Our pre-stabilization method relies on the ability to target any desired joint state. We temporarily evolve a
pair of rigid bodies forward in time to obtain a predicted joint state, and then use a black box joint model to
determine the closest allowable or desired joint state as input to our solver. This entails solving a 6n degree
of freedom nonlinear equation where n is the number of joints in the ARB. The situation is compounded by
contact and collision. We take an iterative approach and apply impulses one joint at a time. We iteratively
solve the 6 degree of freedom nonlinear system for each joint using Newton iteration on smaller systems of 3
equations by first finding and applying a linear impulse to satisfy the positional constraint, and then finding
and applying an angular impulse to satisfy the orientation constraint.
Results
Our method readily handles complicated contact, collision and articulation scenarios such as twenty skeletons dropping into a pile (over 20 million triangles) or a tank with gears driving treads. No special treatment
is required for branching, closed loops, contact, collision, friction, stacking, etc. The algorithm is efficient, effectively linear in the number of bodies and auxiliary constraints (comprising about 5% of the total computation time).
Conclusion
Currently, we are integrating control into our joint framework. Once that step is complete we will combine
this work with our existing muscle simulation and motion capture to develop an accurate model of dynamic
human motion allowing for contact and collision with an unpredictable environment.
References
Guendelman, E., Bridson, R., and Fedkiw, R. 2003. Nonconvex rigid bodies with stacking. ACM TOG (SIGGRAPH
Proc.) 22, 3, 871-878.
25
biomedical • computation • at • stanford
session III
4:00pm
Integrated protein interaction
networks for 230 microbes
Balaji
Srinivasan
Purpose
We sought to systematically combine diverse data types to create an integrated picture of protein interaction in 230 microbes.
Antal F. Novak
Jason A. Flannick
Serafim Batzoglou
Harley H. McAdams
Materials and Methods
We calculated several different types of functional linkage between
proteins and assessed their ability to predict whether two proteins
were likely to share the same COG, GO, or KEGG functional category.
For example, protein pairs which are coinherited in multiple organisms are more likely to be functionally linked than pairs which show no
such pattern, and pairs which are both coexpressed *and* coinherited
are even more likely to be functionally linked.
Given this intuition, we framed the problem of network integration as a high
dimensional binary classifier, where we use a vector of interaction predictors to classify protein pairs as functionally linked or unlinked. We require no simplifying assumptions about statistical dependence and no complicated parametric inference steps. Instead we directly compute the high
dimensional conditional posterior P(L|E), which gives the probability that two proteins are in the same
functional category given a vector of evidences. The evaluation of this posterior over millions of protein
pairs is made feasible by our use of the Gray-Moore dual tree algorithm, which makes our method arbitrarily scalable in terms of both the number of evidence types and the number of protein pairs considered. The output of our algorithm is the probability of functional linkage for every pair of proteins in a
given genome.
Results
We find that a plurality of the predictions in each integrated network hinge on moderate but consistent
evidence from multiple sources rather than strong evidence from a single source, yielding hidden biology
which would be missed if a single data source such as coexpression or coinheritance were used in isolation. We show that these subtle interactions have implications for fundamental cell biological processes
in several microbial species. For example, we find new proteins linked to cell division, DNA uptake, and
the cell cycle which would not have been found if our data sources had been used in isolation.
Conclusion
By systematically integrating several different data types across 230 species, we have constructed the
largest collection of probabilistic protein interaction networks compiled to date. Our methods are scalable and can be applied to any sequenced organism and any kind of experimental or computational
technique which produces pairwise measures of protein interaction.
References
Gray and Moore, Rapid Evaluation of Multiple Density Models, AI & Statistics 2003
26
october • fifteen • 2005
Efficient free energy calculation of
peptide insertion in membranes
Purpose
As membrane-bound proteins are essential to cellular life, we are interested
in the possible roles of small peptides (in some sense the precursors of modern proteins). We focus on the insertion of small alpha-helical peptides into
membranes, examining the stability of the inserted peptide relative to the
interfacial configuration and its role in the association of individual peptides into larger multimeric structures, much like modern cellular channels.
session III
4:15pm
David
RodriguezGomez
Materials and Methods
An ideal candidate for studies of peptide insertion is the synthetic peptide
(LSLLLSL)3. At the water-membrane interface, this peptide folds into an amphipathic alpha-helix. It was shown experimentally [1] that, in the presence
of an electric field, the orientation changes from parallel to the membrane to
perpendicular and the location of the center-of-mass (COM) changes from the membrane surface to the center of the lipid bilayer. Furthermore, it associates into tetrameric ion
channels. Experimental results, however, provide no information about stability of individual helices in the
transmembrane orientation.
E. Darve
A. Pohorille
The so-called Adaptive Biasing Force (ABF) method for computation of free energy (FE) was developed in [2],
where it was shown that ABF is much more efficient than previous methods. This method allows for efficient
evaluation of the FE changes associated with the evolution of the system along a reaction coordinate. An
external force, opposite to the mean force, is added to the system and updated adaptively as the simulation
proceeds. Asymptotically, sampling becomes uniform along the reaction coordinate.
Results
We present the FE surface of insertion of (LSLLLSL)3 as a function of two coordinates: the distance of the COM
of the peptide to the center of the membrane (Z) and the orientation of the long axis of the protein relative to
the membrane surface (Φ). Also, we give information on the evolution of the peptide secondary structure. We
show how uniformity of sampling using ABF is accomplished. Finally, we give estimates of statistical errors.
Conclusion
The performed simulations point to a global minimum corresponding to the parallel orientation at the water-membrane interface. The transmembrane arrangement of a single peptide is only metastable, i.e. it corresponds to a local FE minimum. Our multidimensional version of ABF proves capable of sampling slow degrees
of freedom orthogonal to Z. In particular, no hysteresis is observed due to the penetration of water molecules
during insertion. Also, the helical structure of the backbone is preserved thanks to a careful definition of Φ
using solid-body mechanics.
References
[1] Lear et al, Science 240, 1177 (1988)
[2] Rodriguez-Gomez et al, J. Chem. Phys. 120, 3563 (2004).
27
biomedical • computation • at • stanford
We’re not Thor
We just make his hammer
Developing a device that connects the world. Testing
the water to make sure it stays clean. Discovering
a cure that keeps the world safe. They all require the
same thing: the right tools.
With Agilent we make sure you have them. Our
experience in the fields of electronics, communications,
life science and chemical analysis gives us a unique
perspective shared by no other company in the world.
And we build that expertise into every product
we make, from integrated test and measurement
solutions to advanced technologies and breakthrough
components.
So whether you’re a titan of industry or on the verge of
becoming one, trust the high-tech toolmakers at Agilent.
www.agilent.com
© Agilent Technologies, Inc. 2004
28
We’ll help make you stronger.
october • fifteen • 2005
BCATS Poster
Abstracts
29
poster
0
biomedical • computation • at • stanford
Distributed cluster computing: a
method for parallel simulation
of large molecular systems
Abhay
Sukumaran
Purpose
Protein simulations on large timescales are very computationally expensive. We describe a method of handling otherwise intractable
problems using distributed clusters. Such clusters, with rapid internal
communication, and together comprising a larger network, facilitate
a hybrid approach between supercomputers and traditional heterogeneous distributed systems.
Guha Jayachandran
Vijay S. Pande
Materials and Methods
MPI (message passing interface) is a specification for a library that supports parallel computing in a portable and scalable manner. It is designed
to transparently support low-overhead communication within a collection
of heterogeneous processors. We use MPI-enabled software, a modified version
of GROMACS, to carry out molecular dynamics simulations on clusters. Each cluster has a master
node which supervises execution of the simulation on individual computers, and communicates with a
central server to receive simulation parameters and return results. The clusters might range from SMP
machines to heterogeneous distributed systems. Each cluster works independently, and the central servers coordinate distribution of work. Distribution of the MPI-enabled computational core and work units is
via the existing Folding@Home infrastructure, a project that enables individuals worldwide to contribute
processor time to protein simulations.
Results
The system enables new categories of applications to access the computing power afforded by clusters.
Cluster configurations vary widely, but on a standard benchmark DPCC membrane system on SMP machines, typical scaling factors approach 90% and 75% for 4 and 8 processors, respectively.
Conclusion
This system will make possible simulation of molecular systems larger than a single computer can handle
or to longer timescales. At the same time, it can generate large numbers of trajectories, more than can
be obtained with a traditional supercomputer. Such ensembles of trajectories are necessary for good
statistics. Immediate targets of interest for the system include simulation of the ribosome and of vesicle
fusion.
References
1. Charles Peck, Joshua Hursey, Joshua McCoy, and John Schaefer. Folding@Clusters. Super Computing
2004, Purdue University Exhibit
2. Stefan M. Larson, Christopher D. Snow, Michael Shirts, and Vijay S. Pande. Folding@Home and
Genome@Home: Using distributed computing to tackle previously intractable problems in computational biology. Modern Methods in Computational Biology, Horizon Press (2003)
3. Message Passing Interface (MPI) specification http://www.mpi-forum.org/docs/
30
october • fifteen • 2005
tr0m poster
1
tr0m, or the transcendentrecursiveobjectmodel, is an attempt to recreate
linguistic patterns with transforms between existing, mutually disparate
models of representation and behavior.
the hypothesis remains transcendence, or the emergence of patterns
and properties invisible at the present. tr0m aims at creating bidirectional transforms on non-collocated object models with an underlying
set of base data.
Aditya
Mandayam
Amar Das
agilent laboratories’s alfa and cytoscape platforms serve as testbeds.
we attempt to create a set of interdependent models with the intention of a
single transform into the t-domain, eliminating the need for a conventional three
point conversion with dual transforms. with a focus solely on svo ordered languages, we have managed to categorise the behavior of a hundred kilobytes of raw text into one of twelve artificial bins.
these patterns, ‘grounds’, are modeled with units such as whitespace and punctuation.
isotropicity is added to by the introduction of models analogous to network topologies; the sentence
‘the network in a sentence’ suffices. raw textual data from scientific literature, experimental data, and
diagrammatic models define a common hierarchical hyper-graph data structure, based on agilent
cytoscape.
31
biomedical • computation • at • stanford
poster Segmentation of costal cartilage
2
in lower abdomen CT data using
watershed markers
Andrew
Holbrook
Purpose
Focused ultrasound ablation is a promising non-invasive technique that heats
a specific tumor region to fatal levels while minimizing cell death in nearby
healthy areas. Since high attenuation and reflection regions like bone and air
can affect the focus (and thus effectiveness) of ultrasound, treatment planning is necessary. With respect to the liver, the rib cage surrounds it, allowing ultrasonic transducer placement only in limited locations: the windows
between the ribs and the front and center abdominal region below the
sternum. Knowing the rib cage location with respect to the tumor is crucial.
However, the ribs are not solely comprised of bone; the ribs eventually transition to cartilage, which has a similar image intensity range in CT as organs like
the liver. This makes visualization difficult. In order to determine the available
treatment windows for a tumor, the ribs need to be fully extracted. Simple thresholding and a region-growing method (1) are inadequate, since neither address the problem
of similar intensity regions touching each other. We developed an algorithm to overcome this.
Kim Butts Pauly
Materials and Methods
An abdominal CT dataset was removed of its bright skin pixels through a distance based algorithm developed by Zhou et al (1). A threshold, roughly removing fatty and muscular regions, segmented the remaining region. Starting from the superior slice, rib regions were determined based on their size. Should a
region become larger than expected (i.e. a region was connected to the liver), a marker based watershed
transformation would separate the two and continue slice by slice. Finally, 3D labeling was used to remove any item smaller than a final threshold, and the remaining pixels were relabeled as bone. Algorithm
development was performed in MATLAB, and visualizations were performed in OsiriX.
Results
Preliminary results indicated that this segmentation method worked fairly well, successfully making the
ribs distinguishable from nearby organs. One small non-rib region seems to be misclassified: a lone vessel
below the sternum. Also, the ribs, compared with the original data may be slightly enlarged, but this is
often difficult to perceive and does not detract from the visualization of the dataset.
Conclusion
The algorithm does a good job segmenting the ribs from the surrounding tissues. In a 3D software package the data can be easily visualized and proper planning can now occur. The algorithm will be further
tested, accelerated, and automated with other datasets.
References
1. X. Zhou, T. Hara, H. Fujita, R. Yokoyama, T. Kiryu, and H. Hoshi, “Automated segmentations of skin, softtissue, and skeleton from torso CT images,” Proceedings of SPIE, vol. 5370, 1634-1639, 2004.
32
october • fifteen • 2005
Modeling biological tissue with poster
nonconforming finite elements
3
Purpose
The ongoing pursuit in understanding the complex behavior of biological materials such as muscles, blood vessels, and the cornea as well as the materials
used in biomedical devices such as the shape memory alloys used in stents,
and the polymers used in intraocular lenses has lead us to take advantage
of the fast and robust tools of computational mechanics. Unfortunately, the
tools presently available are challenged by the large computational costs
of modeling real three dimensional materials and complex kinematic constraints such as incompressibility. Our goal is to develop a numerical method
for the study of biological materials that exploits the advantages of nonconforming finite elements, in an attempt to yield high accuracy with relatively low
computational cost.
Alex Ten
Eyck
Adrian Lew
Methods
Discontinuous Galerkin (DG) methods are a class of numerical methods that use nonconforming elements to model the behavior of physical phenomena. Nonconforming finite elements allow the property
of interest to be discontinuities across element faces. DG has long been used in the study of phenomena
where discontinuities are present such as shock waves. However, recent studies have shown that DG can
be equally successful in analyzing phenomena where the property of interest is inherently smooth across
the domain, such as the stresses in a blood vessel wall. Using the DG methodology, we developed and
implemented a class of numerical methods for quasistatic nonlinear elasticity. This method can be used
to model any material with a known strain energy density.
Results
The use of DG for nonlinear elasticity has shown tremendous advantages over traditional conforming
finite elements. In studies of neohookean materials (similar to isotropic biological tissues), DG can yield
better accuracy than conforming finite elements for the same computational cost. Unlike conforming
elements, DG does not lock with low order elements when modeling nearly incompressible materials.
Also, the spurious solutions caused by nonphysical kinematic constraints found in conforming elements
we not seen with DG.
Conclusions
The numerical evidence from 2D simulations and a 3D simulation of an idealized stent angioplasty shows
that DG is presently well suited for analyzing the behavior of biological materials. As this behavior is better understood, and multiscale models of tissue are developed, DG will become an even more attractive
candidate. The main advantages of using discontinuous finite element spaces have yet to be exploited
by researchers. The present method is hoped to be a preliminary agent in the development of robust and
parallel methods with the ability to completely model complex multiscale material behavior.
33
poster
biomedical • computation • at • stanford
4
Alain
Laederach
Reconstructing RNA folding
pathways from local probes of
macromolecular structure
A defining characteristic of most large RNA molecules is the presence
of multiple, long-lived intermediates along their folding pathway.
Inna Shcherbakova The ability to quantitatively identify and structurally characterize
these intermediates is at the heart of the RNA folding problem. Local
Mike Liang
probes of conformational change (e.g. the time-resolved changes in
Michael Brenowitz solvent accessibility of individual nucleotides reported by hydroxyl
radical (•OH) footprinting) can help define these intermediates and
Russ B. Altman
provide insight into their structures. However, reconstructing folding
pathways from a multitude of individual accessibility changes is daunting
because it creates a combinatorial explosion of possible kinetic models. We
show that clustering of time progress curves sufficiently reduces the dimensionality of
the data so as to make the problem of reconstructing the folding pathways computationally tractable.
Exhaustive enumeration of all possible kinetic models on a super-computer grid identifies the most
likely intermediates. A quantitative analysis of the flux through the best-fit kinetic model identifies
the major folding pathways of the intron. We used this combined experimental and computational
approach to reconstruct the folding pathways of the Tetrahymena thermophila group I intron under
two very different solution conditions. The flux during Mg2+-mediated folding is di-vided among
many parallel pathways. In contrast, the flux courses predominantly through three pathways during
the Na+-mediated reaction. These differences emphasize the critical role played by counterions and
highlight the unique role of Mg2+ in RNA folding. This work also establishes a general computational
framework for the kinetic analysis of local probes of macromolecular structure and illustrates the
potential of high performance computing as a tool for the analysis and interpretation of complex
processes like RNA folding.
34
october • fifteen • 2005
Prediction of venous obstruction poster
following transvenous
pacemaker placement: a
computer simulation approach Alex Pang
5
Alex Pang
Anne M. Dubin
Nathan Wilson
Jeffrey A. Feinstein
Charles Taylor
Purpose
Venous obstruction is a major problem following transvenous pacemaker
implantation in children with a prevalence of up to 26%. Obstruction is most
commonly observed at the lead-entry site and the junction of the subclavian (SC) and internal jugular (IJ) veins. Early studies have suggested that
the ratio of lead to vessel size can act as a predictor of venous obstruction
though recent studies have disputed this finding. Understanding the effects
of pacemaker lead implantation on blood flow characteristics may provide a
more accurate predictor of obstruction. We present a computational fluid dynamics model which can predict areas of flow stagnation that may lead to vessel
obstruction.
Materials and Methods
We used custom software to create a computer model of the central veins based on angiographically-obtained patient data. Assuming resistance outlet boundary conditions, steady parabolic flow at the inlets, and
rigid vessel walls, steady-state flow simulations were performed. Two venous model sets, one patient-specific
and one idealized, were constructed. Each set included simulations before lead addition and after lead addition at several locations relative to the vessel lumen.
Results
Both model sets showed a three-fold increase in pressure differential [±10%] with lead addition regardless of
relative lead location. Velocity maps for both model sets show areas of stagnant flow in the innominate vein
and at the junction of the left SC and IJ veins induced by lead addition. Variations in placement significantly
altered blood flow characteristics in the innominate vein.
Conclusions
Consistent increases in pressure differential across all models imply that addition of leads cause a consistent
increase in vascular resistance regardless of position. Velocity maps confirm that placement of leads will significantly alter vessel hemodynamics, inducing areas of stagnant flow specifically at the common sites of occlusion. Future methods that quantify stagnant flow to rank lead positions could yield an improved clinical
strategy for lead placement.
35
poster
biomedical • computation • at • stanford
6
Alison L.
Marsden
A computational study of energy
efficiency and pressure losses
in the total cavopulmonary
connection
Irene E. VignonClementel
Jeffrey A. Feinstein
Charles A. Taylor
Purpose
The total cavopulmonary connection (TCPC) is an operation performed to treat single ventricle congenital heart defects. The superior and inferior vena cavae are connected to the pulmonary
arteries in a t-shaped junction, separating the systemic and pulmonary circulations. Previous studies of the total cavopulmonary
connection (TCPC) have focused on rest conditions and simplified
geometry. In this work, we hypothesize that the effects of respiration and exercise cause significant hemodynamic disturbances and
energy loss.
Methods
Time-dependent, 3-D blood flow simulations were performed using a
custom finite element solver. A patient specific TCPC model, including LPA
and RPA branching, was constructed from CT data. Blood flow features, pressure,
and energy losses were analyzed at rest and exercise, and with respiratory variation (based on MRI literature). Exercise IVC flow rates range between 2 and 5 times resting flow. Resistance boundary conditions
are enforced at the LPA and RPA outlets.
Results
Energy efficiency is high at rest but is found to drop substantially with maximal exercise. Flow vortices
increase in intensity with respiration and exercise, explaining higher energy dissipation when compared
to rest. Pressure drop and energy loss in the TCPC are small at rest but increase to significant levels, even
at moderate exercise. Simulation results at rest are well correlated with pressures obtained by catheterization (13 mmHg).
Conclusions
Pressure drop and energy loss in the TCPC are small at rest but increase to significant levels, even at moderate exercise. We conclude that the effects of respiration and exercise should be incorporated in models
to provide realistic evaluations of TCPC performance, and for future work in optimizing TCPC geometry.
36
october • fifteen • 2005
Effects of size and
hydrophobicity on singlewalled carbon nanotubes
poster
7
Artit
Wangperawong
Purpose
Carbon nanotubes have recently gained much attention from various
bioscience societies due to their affinity with ion channels in biological entities. In addition to potential applications of these materials
as artificial ion channels, the physics behind the confinement effect
on water and ions is of fundamental interest. In order to understand
these phenomena, modern computational tools such as molecular
dynamics are more useful than physical experiments since they can
provide more detailed analyses. Previous studies showed the existence
of helical water structures inside nanotubes and seemingly abnormal
water transport phenomena through the nanotubes. The water transport
is especially indebted to the interplay of lost hydrogen bonds and hydrophobicity. Therefore, we hypothesize that different hydrophobicities would result in great changes in water transport. For computational simplicity we only used
carbon nanotubes, but we varied their hydrophobicities with water so that we can infer similar effects of other
nanotube types that actually exhibit the particular hydrophobicity with water.
Daejoong Kim
Materials and Methods
The physical model includes water and nanotubes of varying sizes and different hydrophobicities. We used a
GROMACS molecular dynamics package. The forcefield was Lennard-Jones 12-6 potentials (different C6 parameters results in different hydrophobicities) with electrostatic potentials.
Results
We analyzed the static and dynamic properties of water both inside and outside the nanotubes. Structural
analyses include visualization of water transport, density profiles in radial and axial directions and probability
density profiles of water dipole angles. Dynamic analyses include mean-squared displacement of water (selfdiffusion of water) and water occupancy within the nanotubes. Of special interest are angular distributions of
water as it is indicative of the network of hydrogen bonding. As the C6 value increases, the HOH plane vector
of the water within the tube tend to take positions such that the HOH plane vector is perpendicular to the
wall-normal vector. In other words, rotation in this dimension is more restrictive with a more hydrophilic tube.
In contrast, the HOH plane vector of the water within the tube remains perpendicular to the tube-axis vector
despite varying hydrophobicities.
Conclusion
Using molecular dynamics, we computed static and dynamic properties of water both inside and outside
nanotubes of varying sizes and different hydrophobicities.
References
(1) Hummer, G., Rasaiah, J. C., Noworyta, J. P. (2001) Nature, 414, 188-190.
(2) Koga, K., Gao, G. T., Tanaka, H., Zeng, X. C. (2001) Nature, 412, 802-805.
37
poster
biomedical • computation • at • stanford
A strategy for whole-genome
sequencing and assembly with
high-throughput, short-read
technologies
Andreas
Purpose
Sundquist
DNA sequencing is a central tool in the biological sciences that is now
8
Mostafa Ronaghi
Haixu Tang
Pavel Pevzner
Serafim Batzoglou
being used to obtain the entire genetic code of organisms. The predominant technology, Sanger sequencing, has improved tremendously and
can now read snippets of DNA up to 1000 bp long at a cost of 0.1 cents/
base, but it is still prohibitively expensive and time-consuming for large
genomes. New technologies, such as Pyrosequencing, will lower this cost
by orders of magnitude at the expense of shorter read lengths. In order to
obtain the full sequence of a genome, we randomly sample reads from the
genome to high coverage and then assemble the reads together based on
their overlaps. For complex genomes this process was traditionally infeasible
for short reads due to smaller and less unique overlaps. We have developed a sequencing protocol and an assembly program for sequencing large, mammalian genomes
with short reads.
Methods
Our strategy is a variant on a technique known as hierarchical sequencing that is more amenable to highthroughput methods. First, we take the genome and fragment it into BAC-sized pieces, randomly select fragments to high depth of coverage, and clone them individually. Then, we sample short reads from each clone
with relatively low coverage, keeping track of which clone each read came from.
Our assembly program uses the reads to determine which sets of clones overlap contiguously in the genome
and in which order, effectively breaking up and linearizing the assembly problem. Since each clone is only
sampled with low coverage, we pool reads from multiple, overlapping clones and assemble in three stages. In
the first stage we compute groups of reads that fall within small regions formed by the boundaries of clones
and merges reads within each group into larger, contiguous sequence (contigs). The next two stages combine
the contigs in successively larger regions for assembly into the final assembled sequence.
Results
We tested our methodology on S. cerevisiae, D. melanogaster, and human chromosomes 1, 11, and 21 by
simulating reads from their known sequence using different depths of coverage and read length. Our assembly program manages to deconvolute the random clones into contiguous, independent regions, allowing us
to scale linearly with genome size. The program assembles the reads into large contigs, which we compared
against the original genome for errors. We found that with short read lengths of 200 bp and only 11-fold coverage we were able to achieve contig sizes and misassembly rates comparable to those in draft mammalian
sequencing projects.
We conclude that our strategy will enable future high-throughput sequencing technologies to produce cheap,
high-quality assemblies of complex genomes.
38
october • fifteen • 2005
Video-audio mapping of the
physical environment for blind
navigation
Purpose
The purpose of this research is to employ a new computer-based navigation
aid in developing a new methodology, focusing on perception and cognition, to govern the design of assistive technology for the blind.
Blindness or significant visual impairment eliminates or significantly decreases the effectiveness of a primary sensory channel. Detecting and
avoiding hazards is typically passed on to other primary or secondary
channels of lower bandwidth than vision. New devices are often designed
on this premise, with a high accompanying risk of sensory and cognitive
overload.
poster
9
Rebecca
Maria Currano
David Grossman
Materials and Methods
The research involves development and utilization of a portable interface that performs a
video-to-audio mapping of the physical environment to enable the blind and visually impaired to safely and independently navigate the world around them. It captures the environmental scene through 2 video cameras, and
applies geometrically and color-related artificial intelligence algorithms to decode the scene and assign names
and categories to objects within the scene. It then presents specific objects as sounds in 3D audio space.
We hypothesize that the likelihood of overloading the hearing channel and/or the cognitive process can be reduced through smart control of sound variables such as volume, repetition rate, accompanying tactile alerts, and
the classes of sounds chosen to represent objects. Cues should not unduly distract the user, or mask important
natural sounds such as speech, flow of traffic, or rear-approaching hazards.
Testing
Since cognitive load cannot be measured directly, tests must be created to measure other variables instead, such
as information retention from conversations, and rate of unintended encounters with objects and obstacles presented by the system, as well as those discernable by other natural environmental cues. Ear-tracking and gesture
study may also provide a means to study cognitive load through attention and discovery-related tasks.
Results
Currently the system recognizes red and green traffic lights and outputs a concatenated stereo audio stream that
identifies their presence, color, and azimuthal direction to the user.
Conclusion
The first stage of the project is complete. Next steps include further development of the object recognition capabilities and testing in various urban and campus settings.
Reference
Miele, JA, 9th International Conference on Audio Display, Boston, MA, July 2003.
Adcock, AB, Effects of Cognitive Load on Processing and Performance.
King, WJ, & Weghorst, SJ, Ear-Tracking: Visualizing Auditory Localization Strategies, CHI 95 proceedings, ACM.
Begley, S, Gestures Help Us Think, Wall Street Journal, Nov 14, 2003.
39
poster
biomedical • computation • at • stanford
A new formulation to model blood
flow and vessel motion in large,
patient-specific models of the
cardiovascular system
C. Alberto
Purpose
Figueroa
Blood velocity and pressure fields in large arteries are greatly influenced
10
by the deformability of the vessel. However, computational methods
for simulating blood flow in three-dimensional models of arteries have
either considered a rigid wall assumption for the vessel or significantly
simplified or reduced geometries. Computing blood flow in deformable
domains using standard techniques like the ALE method remains a formidable problem for large, realistic anatomic and physiologic models of the
cardiovascular system.
Irene E.
Vignon-Clementel
Kenneth E. Jansen
Thomas J.R. Hughes
Charles A. Taylor Materials and Methods
We have developed a new method to simulate blood flow in three-dimensional
deformable models of arteries called the Coupled Momentum Method for FluidSolid Interaction problems (CMM-FSI). This formulation starts from a conventional stabilized finite element formulation for the Navier-Stokes equations in a rigid domain and modifies it in such a way
that the deformability of the wall domain surrounding the fluid is taken into account. We consider a strong
coupling of the degrees-of-freedom of the fluid and the solid domains. The effect of the vessel wall boundary is
therefore added in a monolithic way to the fluid equations, resulting in a remarkably robust scheme.
Results
We present the application of the method to four different problems. The first problem considers an idealized
model of the common carotid artery and studies the impact of different outflow boundary conditions on the
simulation. The second example considers a subject-specific model of a healthy abdominal aorta, and compares
the solutions obtained with rigid and deformable wall theories. The third example illustrates an application to
disease research by comparing solutions obtained in an Abdominal Aortic Aneurysm model (AAA) under rest
and simulated exercise conditions. Lastly, the fourth geometry represents a subject-specific model of a severe
thoracic aortic coarctation and a modified post-operative model where the stenosis has been removed by placing a stent. This last problem exemplifies the applicability of the CMM-FSI to surgical planning.
Conclusion
We have derived, implemented and applied a new formulation to compute blood flow in deformable models of
arteries. The method has proven to be very robust and efficient, representing a minor increase in computational
effort beyond that of rigid wall simulations.
References
A Coupled Momentum Method for Modeling Blood Flow in Three-Dimensional Deformable Arteries. Alberto
Figueroa, Irene Vignon-Clementel, Kenneth E. Jansen, Thomas J.R. Hughes, Charles A. Taylor. Accepted for publication in Computer Methods in Applied Mechanics and Engineering, 2005.
40
october • fifteen • 2005
Application of a three poster
dimensional image-based
modeling technique to the
circle of Willis
Chase Garrett
Purpose
Image-based modeling methods have been applied to quantify biomechanical factors in cardiovascular disease progression [1] and predict outcomes
Yarbrough
of medical interventions [2]. One approach that has been broadly applied
11
to create complex vascular models consists of creating several two dimensional segmentations along the center line of the blood vessel and lofting a
three dimensional NURB surface over these two dimensional segmentations
[3,4]. A boolean union operation applied to the collection of
2-D vessel segments is then used to create solid models. The principal shortcomings of this 2-D approach are that it is time-consuming and user-intensive
to create models and that the resulting models are inaccurate near bifurcations and
branches. Recently, Bekkers and Taylor introduced a new approach based on direct 3D
segmentation, a multiscale retriangulation and NURB surface fitting [5]. In the present study, we
implemented this approach into our image-based modeling framework and applied it to model the cerebral vascular system in a patient with a cerebral aneurysm.
Erik Bekkers
Nathan Wilson
Charles Taylor
Method
This technique uses the fast marching and level set methods for image segmentation and then a novel NURB surface fitting method for model representation. The only user interaction required is picking seed points and setting
parameters for the level set equations.
Results
Initial results indicate a 20-fold time reduction over two dimensional methods, and a more accurate representation
of the dataset, especially where vessel junctions or irregular shapes occur. One challenge in applying this direct-3D
approach has been in separating adjacent vessels.
Conclusions: The 3-D image-based modeling technique presents dramatic benefits in efficiency and noticeable increases in accuracy. However, since 2-D methods still have some advantages with data that is difficult to segment,
future work should include creating an environment in which it is possible to combine the two techniques.
References
[1] C.A. Taylor, T.J.R. Hughes, and C.K. Zarins, “Computational Investigations in Vascular Disease.” Computers in Physics, Vol. 10, No. 3, pp. 224-232, 1996.
[2] C.A. Taylor, M.T. Draney, J. P. Ku, D. Parker, B. N. Steele, K. Wang, and C.K. Zarins, “Predictive Medicine: Computational Techniques in Therapeutic Decision-Making.” Computer Aided Surgery. Vol. 4, No. 5, pp. 231-247, 1999.
[3] K.C. Wang, R.W. Dutton, C.A. Taylor, “Level Sets for Vascular Model Construction in Computational Hemodynamics.” IEEE Engineering in Medicine and Biology. Vol. 18, No. 6, pp. 33-39, 1999.
[4] N. Wilson, K. Wang, R. Dutton, C.A. Taylor, “A Software Framework for Creating Patient Specific Geometric Models
from Medical Imaging Data for Simulation Based Medical Planning of Vascular Surgery.” Lecture Notes in Computer
Science, Vol. 2208, pp. 449-456, 2001.
[5] E.J. Bekkers, C.A. Taylor, “Multi-scale Smooth Vascular Surface Model Generation from Medical Imaging Data using Hierarchical Features.” Submitted to IEEE Transactions on Medical Imaging”, 2005.
41
poster
biomedical • computation • at • stanford
SchemaTrans: automating the
translation of arbitrarily
formatted data to relational
databases
Christopher B. While quantity of data in the bioinformatics world is increasing at
a dramatic rate, the data remain relatively difficult to access proEgner
grammatically. Much of this difficulty is due to disparate data mod-
12
els expressed in different lexical formats not well suited to rapid,
random access. SchemaTrans is an attempt to mitigate these isJudith D. Cohn
sues by providing a pipeline to move data and format information
into a relational database in the most automated manner possible
while still providing the user with the ability to tailor the database to
his needs. Once the data are in the database, significant gains in application development and data access times can be realized. Under continuing development, SchemaTrans is a tool for creating the framework to hold data
in a relational database and then transcribing the data into that framework. Currently, SchemaTrans
can read XML, a standard format, as well as the HSSP and Homstrad custom, text-based formats, and
can write its output either to SQL scripts files or directly to an Oracle database. Input and output
mechanisms have been generalized and abstracted to facilitate the addition of new formats at each
end of the pipeline. A graphical user interface has been developed, in addition to the command line
batch mode, to allow the user to easily customize the input, schema translation, data transcription,
and output stages of the pipeline. The project binary is currently available from the Protein Function
Inference Group website at Los Alamos National Laboratory and the transition to open-source licensing is in progress.
42
october • fifteen • 2005
Robotic perception of the skin’s poster
mechanical properties - a
validation study
Christopher
Current teledermatology consultations are limited to transmission of visual
images of the patient’s skin to the dermatologist via real-time video links
Enedah
or using the store-and-forward procedure. While visual information is
13
adequate in some cases, the absence of tactile information means the
dermatologist has to make a diagnosis without familiar tools. Studies
have found consultations relying only on visual images to be less accurate than face-to-face consultations1. Tactile images convey information
about the skin’s mechanical properties, such as skin roughness and tissue
stiffness, which are not conveyed visually. Such information is often relied
on by dermatologists and for good reason. For instance, cancerous tissue is
often stiffer and its surface typically feels rougher than normal skin.
Kenneth J. Waldron
Hayes B. Gladstone
The goal of the teledermatology project is to deliver in real time, tactile and visual images
of the human skin to a dermatologist at a remote location. This requires use of sensing devices such as the
texture and stiffness perception device described in [2]. Data from the sensors must be interpreted and used
to create a representation of the surface. This process is depicted below.
Sensing and Data Acquisition -> Data Analysis -> Digital Representation -> Transmission -> Tactile Display
(Phases in Teledermatology)
The sensing device used is the texture and stiffness perception device which is described in [2]. It consists of
an array of piezoelectric sensors and a pair of accelerometers. As the probe is dragged across the skin surface, the skin texture results in contact force variations, which are detected by the piezoelectric sensors, and
vibration of the probe body which is detected by the accelerometers. Data from both sensing mechanisms
is subsequently analyzed in time and frequency domains taking into account additional information such as
position and motion data of the probe. The end result is a digital representation of the skin surface. A digital
representation of the skin’s underlying tissue is generated from stress data which is acquired from the piezoelectric sensors when the probe is pushed a known distance into the tissue.
Experiments are being carried out to compare robotic results with the expert knowledge of dermatologists.
Preliminary clinical data comparing the robotic perception of skin pathologies with that of a dermatologist
would be presented.
References
1. Eedy DJ, Wootton R. Teledermatology: A Review. In The British Journal of Dermatology, 2001 Apr;144(4):696707.
2. Waldron KJ, Enedah C, Gladstone HB. Stiffness and Texture Perception for Teledermatology. In Studies in
Health Technology and Informatics, 2005;111:579-85
3. Howe RD, Cutkosky MR. Dynamic Tactile Sensing: Perception of Fine Surface Features with Stress Rate Sensing. In IEEE Transactions on Robotics and Automation, Vol. 9, No. 2, April 1993.
43
poster
biomedical • computation • at • stanford
A theoretical model that
complements an experimental
system for quantitative
investigation of in vivo proteinprotein interactions
Cesar A.
14
Rodriguez
Andrew Lipkin
Sanjiv S. Gambhir
Purpose
The goal of this project is to provide formal qualitative/quantitative models of the in
vivo interactions between the chaperone proteins hsp90, p23, and bioluminescent versions of the two proteins (hsp90-rluc, and p23-rluc). The models are the explicit expression of a theoretical framework for organizing and interpreting the data gathered using the protein reporter complementation assay which is an experimental system under
development for investigating in vivo protein-protein interactions. The theoretical and
experimental aspects of the overall project are intended to be complementary.
Methods
The following activities are required to construct the models; seed specie and component selection,
process selection, rule declaration, network generation, equation generation, analysis (overall behavior, steady
state, mass balance, parameter sensitivity), hypothesis generation, and iterative development.
Materials
The software tools used are BioNetGen 1.1.5, JDesigner 1.9, Visio 10, Cellerator 1.4.4, and Mathematica 5.1.
Results
Interactions between hsp90 and p23 generate a system with 6 species, 12 processes, and 6 rate constants (endogenous system). Interactions between hsp90, p23, hsp90-rluc, and p23-rluc generate a system with 29 species, 98 processes, and 6 rate
constants (experimental system).
The overall behavior as time elapses of the endogenous and experimental systems is composed of two phases; increase in the
concentration of complexes followed by a steady state.
In the endogenous and experimental systems, as the initial concentration of the monomers approaches 1000 µM or the ratio
ki/k-i approaches 1000, the percentage of monomers found in a complex at steady state approaches 100%.
In the endogenous system, as the ratio of k1/k-1 approaches 1000 (hsp90 homodimerization), the percentage of monomers
in the hsp90-p23 dimer configuration at steady state approaches 0.
In the endogenous and experimental systems, less p23 than hsp90 is present in the complexes at steady state.
In the endogenous system, as the initial concentration of the monomers approaches 1000 µM, the percentage of the monomers found in the tetramer configuration (hsp90-hsp90-p23-p23) at steady state approaches 100%.
As the initial concentration of the experimental monomers (hsp90-rluc, p23-rluc) approaches the concentration of the endogenous monomers (hsp90, p23), the steady state concentration of the endogenous complexes (X3-X6) decreases proportionally.
Conclusion
A set of quantitative and testable hypotheses can be formulated from the theoretical results reported above.
44
october • fifteen • 2005
CONTRAlign: A discriminative poster
framework for protein sequence
alignment
Purpose
Chuong B.
In recent years, most alignment methods that have claimed significant improvements in alignment accuracy have done so not by proposing substanDo
tially new algorithms for alignment but rather by incorporating additional
15
sources of information, such as extending sequences via profiles or using
predicted secondary structures. While these methods have demonstrated
the utility of external information, the scoring models used are often ad
hoc, difficult to train, and prone to overfitting.
Sam S. Gross
Robert C. Edgar
Serafim Batzoglou
Materials and Methods
We present CONTRAlign (CONditionally TRained Alignment), an extensible and
fully automatic framework for robust parameter selection and protein pairwise
sequence alignment based on a probabilistic model known as a conditional random
field (CRF). In the CONTRAlign methodology, the user first defines an appropriate model topology
for pairwise alignment. Unlike ad hoc algorithms in which model complexity (and hence the risk of overfitting) corresponds roughly with the number of free parameters in the model, the effective complexity of a
CONTRAlign CRF-based model is controlled by a set of regularization parameters, “knobs” which determine
the balance between model expressivity and overfitting. Given a set of gold standard partially labelled alignments, CONTRAlign uses gradient-based optimization and holdout cross validation to automatically determine optimal regularization constants and a set of alignment parameters which will perform well on both the
training set and future alignment problems.
Results
We show that even under stringent cross-validation conditions, CONTRAlign can learn both substitution and
gap parameters which generalize well to previously unseen sequences using as few as 20 alignments; in contrast, traditional methods of substitution parameter estimation typically require large databases of thousands
of aligned blocks. Augmenting the aligner to account for known secondary structures and positional solvent
accessibilities is done seamlessly in the CONTRAlign model and yields accuracy improvements of 5-10% in
total correctly aligned positions. Finally, an extension of CONTRAlign to multiple alignment gives cross-validated accuracies matching or exceeding the current best methods.
Conclusion
As an aligner, CONTRAlign yields alignment accuracies which are equal or superior to current state-of-theart protein alignment models while requiring orders of magnitude smaller training sets. As a framework,
CONTRAlign frees aligner developers from issues of parameter estimation and model overfitting, allowing
designers to quickly try out a variety of model topologies and automatically identifying relevant features for
alignment without the need for manual parameter tuning.
45
poster
16
biomedical • computation • at • stanford
Metrics for a mastoidectomy
simulator
Christopher
Sewell
Dan Morris
Nikolas Blevins
Federico Barbagli
Kenneth Salisbury
Purpose
Simulation-based training can be a safe, cost-effective, and easily accessible tool for gaining experience in surgery. Another promising,
though not yet largely exploited, advantage of simulations is their
potential to objectively evaluate user performance and provide detailed constructive criticisms.
Materials and Methods
We have developed a simulator for ear surgery (specifically, a mastoidectomy). We are now creating an interface for this simulator that
allows instructors or the trainee to replay a run of the simulator, objectively score the performance according to various criteria, and visually
highlight problem areas. Two views are shown side-by-side, the left showing the procedure as originally seen by the user, and the right showing special
effects to highlight various metrics. As bone is drilled away in the left view, dots can appear in
the right view, green if the removed bone was within the line of sight and red if not, since drilling bone
that cannot be seen can cause visual cues identifying vulnerable underlying structures to be missed. Alternatively, dots can appear green or red depending on whether the bone voxel was also removed in a
model run of the simulator by an expert surgeon in order to highlight exactly where the trainee should be
drilling. Cut-away views allow the locations of these dots to be seen in relation to critical structures of the
ear. Interactive graphs can be selected to show plots of drill forces and velocities as functions of distances
from vulnerable structures and compared to force and velocity profiles obtained from expert surgeons.
Collision forces with anatomical structures can be shown graphically and with markers at the contact
points in the right-side view. The areas of these structures that have and have not been properly exposed
(by sufficiently thinning the surrounding bone so as to be able to establish their positions through visual
cues) can also be appropriately shaded in the right-side view. Quantitative measures, such as percent of
bone removed that was visible, percent of correct voxels removed, number and severity of collisions with
vulnerable structures, and percent of structures properly exposed, are also reported.
Results and Conclusions
Our metrics have been developed in consultation with experienced surgeons. We are planning a user
study in which our metrics will be evaluated by determining correlations between medical students’ simulator scores and experts’ assessments of their performance on real plastic temporal bone models.
46
october • fifteen • 2005
Bringing the ribosome exit poster
tunnel to life with distributed
computing: a first look at the
thermodynamics of confined
solvent and the nascent chain
Christopher
Purpose
During translation the ribosome synthesizes proteins, extruding the new
D. Snow
polypeptides through a ~100 Å exit tunnel (1). To fully understand cotrans-
17
lational folding and the translation process, we must study the formative
environment of the nascent chain. The ribosome tunnel composition is heterogeneous, including multiple protein chains, rRNA, and attendant counterions. The tunnel varies in diameter between 10-20 Å. The diversity of protein sequence space translated by the ribosome is vast, and it is tempting to
imagine the exit tunnel playing a generic role as a passive container. However,
recent work on specific nascent chains has demonstrated that the ribosome is not a
universal translator (2). Certain nascent polypeptide chains (and the macrolide antibiotics) bind to
the exit tunnel and stall translation (3). Also, the tunnel may actively recognize and respond to an entire class
of nascent chain, namely transmembrane sequences (4). To date, no crystal structure has revealed the details
of nascent chain interactions with the ribosome, and no experiment has inventoried the complement of water
molecules and counterions associated with the exit tunnel. Thus, there is an opportunity for atomic simulation
to provide fresh insight.
Colin Aitken
Vijay S. Pande
Materials and Methods
We have built a large model of the solvated ribosome exit tunnel in atomic detail (~80,000 atoms). We relied
upon the Folding@Home distributed computing project to compute several thousand long molecular dynamics simulations. Essentially, we leveraged volunteer processors to focus an enormous amount of computational power (over 4,000 CPU years over the last 9 months). We then employed a novel array of algorithms
to analyze water structure, build Markovian models of long timescale kinetics, and visualize the interplay
between protein, solvent and rRNA.
Results and Conclusion
The scope of the computation allows us to simultaneously determine the relative free energy of water binding at all locations throughout the model with unprecedented precision. Surprisingly, we observed fine water
structure throughout the ribosome exit tunnel with implications for antibiotic binding affinity and interactions with nascent peptide chains. The organized hydration density provides a view of the exit tunnel that is
strikingly different from the prevailing notion of a passive, nonstick container. In this new view, water becomes
a dominant player for nascent chain thermodynamics.
References
1. Klein DJ, Moore PB, Steitz TA. J Mol Biol. 2004 Jun 25;340(1):141-77.
2. Nakatogawa H, Murakami A, Ito K. Curr Opin Microbiol. 2004 Apr;7(2):145-50.
3. Tu D, Blaha G, Moore PB, Steitz TA. Cell. 2005 Apr 22;121(2):257-70.
4. Johnson AE. FEBS Lett. 2005 Feb 7;579(4):916-20.
47
biomedical • computation • at • stanford
poster From biomedical images to
18
Chand T.
John
Scott L. Delp
geometric models
Purpose
Algorithms for constructing 3D models of structures from biomedical images are
usually context-specific. Hence, software developed to facilitate general biomedical image analysis is usually limited in scope. The 3D Slicer, however, is a
freely available, open-source, extensible and customizable image analysis software package currently under development. The objective of this project was
to develop modeling tools within Slicer that can be applied broadly while still
allowing the context-specific flexibility that clinical researchers need.
Materials and Methods
Development for Slicer 2.6 was done using Tcl 8.4.5/Tk 8.4.5 and Microsoft Visual
C++ .NET 2003 in Windows XP. Many tools were developed to facilitate manual
segmentation, four of which required designing nontrivial algorithms:
1. Inserting intermediate points in user-defined polygons
2. Selecting and moving points by automatic proximity detection
3. Real-time cardinal spline interpolation of a user-defined polygon
4. Automatic sampling of points from a cardinal spline for model construction
Algorithms 1 and 2 were developed to automatically detect the proximity of the user’s mouse pointer location
to different parts of a control polygon being modified in real time. Algorithm 3 was developed to compute
derivatives at the ends of a control polygon based on limited information and to render a cardinal spline
through an arbitrary set of control points based on the de Casteljau algorithm. Algorithm 4 was developed
using simple affine interpolation of parameter values. Preliminary testing was done to debug these algorithms in Tcl and Visual C++. Further testing is currently underway.
Results
Algorithms 1 and 4 work well when user-defined polygons have only small changes in curvature between adjacent vertices. Algorithm 2 works well under normal circumstances. Algorithm 3 performs well overall, but
does have precision problems and curvature problems at spline endpoints that are common to many types
of parametric curves.
Conclusions
The manual segmentation tools and the four algorithms developed above have helped make Slicer into a
more practical tool for image analysis in the Neuromuscular Biomechanics Lab at Stanford and Brigham and
Women’s Hospital at Harvard.
References
A. S. Arnold, S. Salinas, D. J. Asakawa, S. L. Delp. Accuracy of Muscle Moment Arms Estimated from MRI-Based
Musculoskeletal Models of the Lower Extremity. Computer Aided Surgery, 5: 108-119, 2000.
G. Farin. Curves and Surfaces for CAGD, 5th ed., Academic Press, 2002.
http://slicer.org
48
october • fifteen • 2005
Algorithmic analysis of the poster
quantitative distribution of
neural stem cells in malignant
glioma David JingPurpose
One of the largest obstacles in traditional brain tumor treatments such as surgery
Piao Lin
and radiation is the degree to which glioma cells migrate away from the main
19
tumor mass. As an emerging alternative therapy, neural stem cells (NSCs) hold
great promise for glioma therapy because of their inherent ability to migrate
toward the main tumor mass as well as target metastatic tumor sites. NSCs can
thus potentially be used as therapeutic delivery vehicles for targeting tumor
cells. Despite several studies supporting the ability of NSCs to track disseminating glioma, there have been limited studies characterizing the distribution of NSCs.
The purpose of our study was to quantitatively investigate the distribution
of human NSCs in human gliomas in a mouse model. Analyzing the quantitative distribution of NSCs is important for assessing the application of in
vitro chemoattractant studies to in vivo distribution and determining the
comprehensiveness of potential NSC drug therapy for gliomas.
Methods
We used confocal microscopy, medical imaging software, and mathematical
modeling to quantify the distribution of human NSCs in experimental human
glioma after caudal lateral implantation in adult nude mice.
Josip Najbauer
Paul Salvaterra
Adam Mamelak
Michael Barish
Elizabeth Garcia
Marianne Metz
Steven Kendall
Maggie Johnson
Karen Aboody
Results:
Our results confirm that injected human NSCs achieve a broad distribution throughout masses
of human glioma and target metastatic glioma islands. NSCs seemed to follow the tumor gradient, showing higher
density at the centroid of the tumor mass. We estimated that NSCs carrying a representative tumor therapeutic with
a “killing radius” of fifty microns would eliminate over 56% of the main tumor mass.
Conclusions:
We conclude that NSC-mediated delivery of therapeutic enzymes may likely be effective due to the broad area of
tumor covered by NSCs and their demonstrated ability to track metastatic islands of tumors. We have demonstrated
novel methods and algorithms for analyzing the distribution of neural stem cells within intracranial glioma. Our
study should serve as a quantitative model for future studies that seek to (1) understand the chemical and biological
factors involved in the stem-cell tumor tropism and (2) achieve the most efficient conditions that produce optimal
therapeutic effects with NSCs in glioma.
References
Aboody, K.S., et al. 2000. Neural stem cells display extensive tropism for pathology in adult brain: Evidence from intracranial gliomas. Proc. Natl. Acad. Sci. U.S.A. 97: 12846-12851.
Westphal M, Black P. Perspectives of cellular and molecular neurosurgery. Journ. of Neuro-Onc. 2004; 70: 255-269.
Weissleder R. Scaling down imaging: molecular mapping of cancer in mice. Nature Reviews Cancer. 2002; 2: 1-8.
49
biomedical • computation • at • stanford
poster Locating single-stranded regions of
20
the human genome
Diane
Schroeder
Richard Myers
50
DNA becomes single-stranded in a variety of important biological situations, including the transcription initiation bubble and DNA replication. In
addition, there is evidence that some genes, such as c-myc, might be regulated by nearby denaturation-prone DNA sequences and the proteins
that bind to those denatured sequences. I am developing a genomewide assay to find denatured DNA regions in vivo. I will use this data to
1) see how often DNA denaturation is involved in gene regulation and 2)
create an algorithm that predicts new genomic regions likely to regulate
genes via DNA denaturation.
october • fifteen • 2005
Image interpolation in 4D CT
using a Bspline deformable
registration model
Purpose
To study organ deformation during respiratory motion and develop a
method for deriving the phase-binned 3D image sets through the interpolation of the images at some known phase points.
poster
21
Eduard
Schreibmann
Brian Thorndyke
Lei Xing
Methods and materials
4D CT datasets for 3 patients were acquired. For each patient, the correlation between inhale and exhale phases was studied and the deformation of the anatomy between the two phases was quantified using
a BSpline deformable model. In this method, the deformations field between the inhale and exhale phases was defined on a set of nodes overlaid on the image and the deformation coefficients were determined through
iteratively modifying the node displacements until the optimal transformation was
obtained. Once the deformation field between the two phases is determined, the images at an arbitrary
phase can be deduced by an interpolation of the transformation coefficients. The accuracy of the proposed scheme was assessed by comparing the interpolated images with the actual image sets acquired
at the corresponding phases. The difference between the two sets of images was quantified by (a) monitoring the positions of implanted surface fiducial markers (b) checkerboard display and (c) analysis of the
subtraction of the interpolated and the actual images.
Results
It is shown that the images at the intermediate phases could be derived by an interpolation of the deformation field. An analysis of movements of implanted and surface fiducial markers in the interpolated
and actual 4D CT images indicated that 3 mm accuracy is achievable by the interpolation. The geometric
accuracy was confirmed by the evaluations of the checkerboard and subtraction images of the interpolated and actual 4D CT images. The subtraction of the two images indicated a similar level of success. The
fractional voxels having a difference higher than 20 HU was only 0.1% in all the subtraction images. The
proposed technique was also used to automatically map the organ contours in a known phase to other
phases. The mapping provides an effective way for designing patient specific margins in the presence of
respiratory motion. Finally, the technique lead to a 90% reduction in the acquired data because in the
BSpline approach a lattice of only a few thousand values is sufficient to describe a CT dataset of 25 million
pixels.
Conclusions
This study suggests that the organ deformation during the breathing process can be well modeled by using a BSpline deformable algorithm and it is possible to obtain 4D images by interpolating the 3D images
acquired at some distinct phases. The proposed technique offers useful means for radiation dose reduction, binning artifacts removal, and disk storage improvement in 4D imaging.
51
poster
22
biomedical • computation • at • stanford
GROMACS on the GPU
Erich Konrad
Elsen
Purpose
The purpose of our project is to harness the computational power of
GPUs (commercial graphics cards) to speed up molecular dynamics
simulations and reduce the runtime of simulations.
Vishal Vaidyanathan
Materials and methods
Eric Darve
We are re-implementing parts of the GROMACS molecular dynamics package on the GPU using the BrookGPU language developed
Pat Hanrahan
here at Stanford. All major parts of the code except the neighbor
Vijay Pande
list search have been mapped to the streaming computing model
and ported to the GPU.
Ian Buck
Results
For the force calculations, ATI’s X800XT PE achieves almost 5x the GFLOPS of a
Pentium IV 3.0Ghz. NVIDIA achieves only 1.5x the number of GFLOPS of the CPU. On ATI, the overall
runtime is reduced to 70% of the original runtime.
Conclusion
As GPUs increase in power faster than CPUs the performance gap will only widen. For a wide variety
of applications streaming computing will be the future of scientific computing. The new Cell processor from IBM shows promise for being extremely useful and powerful (256 GFLOPS peak in single
precision) as a streaming processor for scientific computation.
References
Buck, I., Foley, T., Horn, D., Sugerman, J., Fatahalian, K., Houston, M., Hanrahan, P. “Brook for GPUs:
Stream Computing on Graphics Hardware.” In Proceedings of Supercomputing 2005. http://graphics.
stanford.edu/papers/brookgpu/brookgpu.pdf.
Lindhal, E., Hess, B., van der Spoel, D., “Gromacs 3.0: A package for molecular simulation and trajectory
analysis.” Journal of Molecular Modeling. 7:306-317, 2001.
Erez, M., Ahn, J., Garg, A., Dally, W., Darve, E. “Analysis and Performance Results of a Molecular Modeling Application on Merrimac,” Proceedings of the 2004 ACM/IEEE conference on Supercomputing, p.
42, 2004.
52
october • fifteen • 2005
3-D simulations of hemodynamic poster
factors in pulmonary
hypertension
Tim A. Fonte
Purpose
23
Primary pulmonary hypertension is a deadly disease with undetermined
origins. Simulations of blood flow in healthy and hypertensive pulmonary
arteries (PAs) provide detailed information about flow characteristics and
forces on vessel walls, which can lead to a new understanding of how vessels remodel and the disease progresses.
Irene E.
Vignon-Clementel
C. Alberto Figueroa
Jeffrey A. Feinstein
Charles A. Taylor
Materials and Methods
Models of PAs for healthy and hypertensive subjects were created from
MRA [1] to the maximum detail of each data set (50-70 vessels per model). Blood flow was measured with PC-MRI at the main, left, and right PAs.
Arterial pressure was measured for the hypertension subjects. The models
were meshed with 3-4 million elements. A custom finite element method
was used to solve the incompressible Navier-Stokes equations over multiple
cardiac cycles. Resistance boundary conditions were set at the outlets to obtain
patient-specific flow split and pressure [2], and a Womersley velocity profile was set at
the inlet. Simulations were performed with both rigid walls and a deformable fluid-structure model [3].
Additional simulations utilized adaptive meshing based on local error [4].
Results
The mean wall shear stress (mwss) in the left and right PAs is 16.8 dynes/cm2 (healthy) and 3.5 dynes/cm2
(hypertension). Flow recirculation was observed in both healthy and hypertensive subjects; however there
is more complex flow in the healthy arteries. There are large regions of low, stagnant flow in the patient with
pulmonary hypertension. An adaptively refined anisotropic mesh of 1 million elements was used to obtain
comparable results to a 3 million element isotropic mesh (mwss difference of 6% in the left and right PAs).
Conclusions
There are dramatic differences in blood flow characteristics between healthy PAs and PAs with pulmonary
hypertension, especially the lower wall shear stress seen with pulmonary hypertension. Adaptive meshing
allowed for similar results at a fraction of the computational cost.
References
[1] N. Wilson et. al., A Software Framework for Creating Patient Specific Geometric Models from Medical Imaging Data for Simulation Based Medical Planning of Vascular Surgery. Lect. in Comp. Sci., 2001; 2208; 449-456.
[2] I. Vignon et. al., Outflow Boundary Conditions for Three-Dimensional Finite Element Modeling of Blood
Flow and Pressure in Arteries. Accepted in Comp. Meth. in App. Mech. and Eng., 2005.
[3] C.A. Figueroa et. al., A Coupled Momentum Method For Modeling Blood Flow In Three-Dimensional Deformable Arteries. Accepted in Comp. Meth. in App. Mech. and Eng., 2005.
[4] J. Müller et al., Anisotropic Adaptive Finite Element Method for Modeling Blood Flow, Accepted in Comp.
Meth. Biomech. Biomed. Engr., 2005.
53
poster
biomedical • computation • at • stanford
Filling missing components in
yeast metabolic pathways using
heterogeneous data
Gal Chechik The set of cellular metabolic reactions forms a complex network of
24
interactions, but even in well studied organisms the resulting pathways contain many unidentified enzymes. We study how ‘neighborAviv Regev
hood relations’ between enzymes in metabolic pathway are maniDaphne Koller
fested in functional properties of the corresponding genes and
their products, including mRNA expression, protein domain content and cellular localizations. We develop compact and interpretable probabilistic models for representing protein-domain co-occurrences and gene expression time courses. The former can provide
predictions of domains and genes functions and the later reveals the
relation between the activation time of yeast genes and the usage of their
protein products in the pathways. These models are then combined for completing the unidentified enzymes in the pathways, achieving accuracy that is significantly superior to
existing state-of-the-art approaches.
54
october • fifteen • 2005
Simulation of fluorescence poster
anisotropy experiments: probing
protein dynamics
Purpose
Gunnar F.
Time-resolved fluorescence anisotropy decay experiments on a protein-attached dye can probe local protein dynamics and steric restrictions, but
Schroeder
are difficult to interpret at the structural level [1]. Aiming at an atomistic
25
description, we have carried out molecular dynamics simulations of such
experiments.
Ulrike Alexiev
Helmut Grubmuller
Results
Our simulations describe an Alexa488 fluorescent dye maleimide derivative covalently attached via a single cysteine to the AB-loop of bacteriorhodopsin. Fluorescence anisotropy decay curves obtained from the
simulations agree well with the measured ones. Three anisotropy decay components were resolved and assigned 1) to the fast dynamics of the attached dye
on the picosecond timescale, 2) to the slower dynamics of the loop at the one nanosecond timescale and 3) to the overall tumbling of the molecule. For the biologically relevant one nanosecond component we identified two processes from simulations, the motion of the flexible loop as well as
slow conformational dynamics of the dye. These two processes are not separable by experiment alone.
Furthermore, analysis of the correlation between the dye and the protein motion revealed which part
and which motion of the protein is actually probed by the experiment. Finally, our simulations allowed to
test the usual and inevitable assumption underlying these types of spectroscopic measurements that the
attached dye probe does not severely perturb the protein dynamics. For the case at hand, by comparison
with a simulation of the dye-free protein, the perturbation was quantified and found to be small.
References
[1] U. Alexiev, I. Rimke, and T. Pohlmann. Elucidation of the nature of the conformational changes of the
EF-interhelical loop in bacteriorhodopsin and of the helix VIII on the cytoplasmic surface of bovine rhodopsin: A time-resolved fluorescence depolarization study. J. Mol. Biol. 328:705-719 (2003)
[2] G. F. Schroeder, U. Alexiev, and H. Grubmuller. Simulation of fluorescence anisotropy experiments:
Probing protein dynamics. Biophys. J., accepted.
55
biomedical • computation • at • stanford
poster Efficient prediction of protein-
26
Haidong
Wang*
protein interactions using Markov
networks
* These authors contributed equally
Purpose
Mapping the network of protein-protein interactions is a key step towards
understanding many cellular processes. Current assays for measuring interactions are noisy and partial. Recent work [1] has shown that Markov
networks can effectively integrate heterogeneous data and deal with
noise to make better predictions. However, standard inference methods
in these models are slow and do not scale to large, dense networks. We
propose a new algorithm that speeds up the computation without reducing prediction accuracy, opening the door towards a genome-wide analysis,
and allowing the use of richer models involving multiple interaction types such
as regulatory interactions.
Matt Schefer*
Gal Elidan
Daphne Koller
Methods
Our model utilizes multiple interaction and cellular localization assays, treating them as noisy sensors
for the underlying truth. It encodes a preference for transitivity in the interaction graph, and the fact
that interacting proteins tend to be localized in the same compartment. Unlike previous work, we
restrict to models that satisfy certain regularity conditions. While these conditions do not necessarily
hold in biological systems, they allow the use of more efficient inference algorithms, using min-cut or
linear programming rather than iterative methods such as loopy belief propagation.
Results
We train our model on a set of 2000 reliable protein-protein interactions and 2000 non-interactions,
and compare our models with standard max-likelihood learning [1] using 4-fold cross-validation. In
weak models, which do not account for transitivity, our method is worse than [1], because the approximation forced by the regularity condition introduces significant error. In the richer models that
include transitivity, we are collectively classifying all of the interactions together, which helps correct
these errors. Here, we achieve improvements in both accuracy and running time. The true positive,
false positive rate, and running time are: our method – TP=85.6%, FP=1%, time=280 sec vs. Jaimovich
et al [1] – TP=84%, FP=2%, time=452 sec. When we further account for the possibility of noisy localization results, we obtain further improvements: TP=87.2%, FP=1.2%, time=680 sec vs. TP=85.5%, FP=2%,
time=788 sec. Our method also scales to data sets with many more interactions, where the existing
method fails. With 6000 (less reliable) interactions, we obtain an accuracy of TP=86.4%, FP=2.7%.
References
1. Jaimovich A, Elidan G, Margalit H, and Friedman N. Towards an Integrated Protein-Protein Interaction Network. RECOMB 2005
56
october • fifteen • 2005
Framework for identification of
allosteric networks
Purpose
Allosteric regulation is a key feature of enzyme function. Lockless et al. proposed an algorithm that uses sequence data to search for potential allosteric
networks in proteins using statistical analysis of multiple sequence alignments. We have created user-friendly
software to perform this algorithm. We are applying the software to the
chaperonin proteins.
poster
27
John
Patrick
Adams
Erik J. Miller
Christoph Spiess
Materials and Methods
A protein of interest’s sequences are collected and aligned using the ClustalW algorithm and Blosum62 scoring matrix. We read the alignment into an
application that analyzes the alignment to identify statistically interesting pairs
of sites within the protein that appear to be evolutionarily coupled. The site-site coupling energies, a statistical measure of the strength with which the sites influence each other,
can then be clustered, mapped as a network, or colored on a three-dimensional PDB structure for analysis.
Identified coupled sites can be confirmed by assaying the protein containing designed point mutations at
those sites for loss of allosteric behavior.
Results
Chaperonins are large protein complexes that couple the energy of ATP hydrolysis to the folding of newly synthesized proteins. They are essential proteins composed of two rings of 7-9 subunits, each subunit consisting
of 3 domains. There are several modes of allostery within the chaperonins, providing an ideal family to test
our software. Cooperativity includes negative inter-ring and positive intra-ring cooperativity for ATP hydrolysis, and allosteric coupling of ATP hydrolysis within a subunit to substrate remodeling. We are undertaking
a statistical analysis of large-scale sequence alignments to generate hypotheses about the mechanisms of
chaperonin activity.
Conclusion
We have developed an extensible cross-platform java application that makes the analysis of protein alignments for site-site coupling simple for any user. We hope that this will promote the use of this new method
to understand communication within the domains of a protein. Compared to experimental approaches such
as mutagenesis, it drastically reduces the time needed for analysis and allows the study of proteins that are
difficult to manipulate experimentally. We hope that it will serve as a foundation for the development of more
advanced algorithms to be able to analyze more complex problems in protein science such as binding of receptors and ligands.
References
Lockless, S.W. et al., Evolutionarily conserved pathways of energetic connectivity in protein families. Science,
1999. 286(5438): p. 295-9.
Suel, G.M., et al., Evolutionarily conserved networks of residues mediate allosteric communication in proteins.
Nat Struct Biol, 2003: p. Advanced Online Publication.
57
poster
biomedical • computation • at • stanford
28
Jessica
Ebert
Unsupervised identification of
functionally relevant residues in
protein families
Purpose
Both the impending increase of solved structures for proteins of unknown function and the uncertainties involved in sequence based functional annotation methods point to the need for a greater understanding of the relationship between structure and function. Because global
structural similarity does not imply functional homology, methods are
required for the identification of functionally relevant regions of protein
structures upon which models can be built to recognize similar sites in
other proteins. Identification of examples of a functional site used in the
construction of such a model, however, often represents a bottleneck in functional annotation pipelines. Furthermore, a residue that is not directly involved in
a protein’s chemical activity may nonetheless contribute to an active site’s environment in a
manner that might make it an interesting target for protein engineering and drug design applications.
We propose to identify positions in a protein family that are directly or indirectly involved in the family’s
functional activity by locating residues whose physicochemical environments are both conserved and
unique. While core residues will likely yield models representing generic hydrophobic positions, the environments of functional residues may have unique characteristics that will give rise to highly sensitive and
specific models.
Russ Altman
Methods
We first build a multiple structure alignment in order to find residue correspondences across the family,
and then use FEATURE (Wei and Altman, 1998) to model the environment around each of the alignment’s
columns. FEATURE identifies attributes of the environments around a column’s residues that distinguish
them from negative training examples selected to match the column’s atom density and sequence composition. We build several models for a column, each of which excludes one residue.
To determine whether a column gives rise to a model that is both sensitive and specific, we scan each
residue in the family against the model for the column that was built without using information from the
protein to which the residue belongs. The area under the ROC curve for a column and the average score
separation between the column’s residues and other residues then reflect the ability of the column’s models to score residues from the column higher than residues from other columns.
Results
We have identified eight positions in a family of serine proteases whose environments are both conserved
and unique. The three positions of the catalytic triad are included in this set of residues, and the other five
residues are in close proximity to the family’s active site.
References
(1) Wei L, Altman RB. Pac Symp Biocomput. 1998: 497-508.
58
october • fifteen • 2005
A modified Bayesian approach to poster
determination of causation in
differential gene expression
Purpose
Jessie Dale
DNA microarrays have given researchers vast amounts of new data
regarding differences in gene expression between normal and disTenenbaum
eased tissue types. Still missing, however, is the ability to use this
29
information about differential gene expression to gain insight into
upstream causal mechanisms of disease. We have developed a modified Bayesian approach to predict the most likely ligand responsible
for changes in gene expression given a list of genes observed to be
significantly up- or down-regulated and a set of possible ligands.
Paul J. Utz
Atul J. Butte
Materials and Methods
Our model was trained on a large DNA microarray data set made available by
the Alliance for Cell Signaling in conjunction with Nature Publishing Group. In this
data set, B cells were stimulated with 32 different ligands and gene expression data was observed at
four time points. We have applied multiple methods for representing this time course data representation (extreme value z-scores, area under the curve, q-value, etc.) to develop a probabilistic model
which allows us to make predictions regarding the relative likelihood of each ligand in our training
data given a set of genes observed to be significantly affected in an independent test data set.
Results
Using the various methods described above for representation of time course data and different algorithms for determination of significance, we tested our prediction method on independent test
data sets which we downloaded from the Gene Expression Omnibus. Different methods above gave
varying success rates in predicting the ligand that was actually used in the experiment that generated
the test data. Strikingly, in some cases our method was able to make a correct prediction even in cell
types that differ from the type used to train the model (macrophages vs. B cells).
Conclusions
While our method has some obvious limitations, for example the finite set of ligands in the training
data, as well as a sub-optimal true positive rate of 67%, we believe it shows promise for future use in
disease mechanism elucidation. We continue to refine the method to improve accuracy.
References
1.
http://www.signaling-gateway.org/
2.
http://www.ncbi.nlm.nih.gov/geo/
59
biomedical • computation • at • stanford
poster Use of a naive Bayesian classifier
30
Jason A.
Hackney
to identify gene regulatory
networks in the protozoan
parasite, Enatmoaba histolytica
Gretchen M.
Ehrenkaufer
Upinder Singh
Purpose
Entamoeba histolytica is a protozoan parasite responsible for amebic colitis, a disease that kills over 100,000 per year. Little is known about gene
regulation in protozoan parasites, in general, and E. histolytica in particular. We wished to identify gene regulatory networks within E. histoyltica
trophozoites, in order to better understand the biology of this important
parasite.
Materials and Methods
We hybridized RNA isolated from mid-log cultures of E. histolytica HM-1:
IMSS trophozoites to a short oligonucleotide array. This array contains probes
designed against 7,770 of the ~10,000 amebic genes. We identified genes with
either high or low signal on the microarray. The promoters for these genes were retrieved from a promoter database, and motif elucidation using the MEME program was carried out. We
identified thirty motifs in each set, and filtered for motifs that occurred in both sets. These motifs were
used to create a naive Bayesian classifier, in which motif presence and position in the promoter were used
as predictive variables for gene expression. Results from our microarray hybridization were randomly partitioned into a training set and cross-validation set.
Results
The Bayesian classifier was able to accurately predict the expression of 70% of amebic genes. By determining which promoter motifs are used to classify genes, we were able to find a set of three motifs that
are predictive of low gene expression in amebic trophozoites. While any single motif was unable to predict gene expression, genes with more than one motif were highly likely to have low expression. The
three motifs were found in a stereotypical spacing and distance from the transcription start site. A large
number of genes with this regulatory pattern belong to the same gene family. This family is known to
have median to low expression in trophozoites, with increased expression during stress response. We find
that genes of this family with two motifs show this expression pattern, while a family member with none
of these motifs was equally expressed in normal and stress response conditions.
Conclusions
We have been able to identify a regulatory module that affects gene expression of a large family of amebic genes. The network identified in this work is highly complex, and understanding this network and the
genes it regulates should lead to a better understanding of both genetic regulation in ameba, and evolutionary dynamics of regulatory networks. Importantly, this work can be extended easily to novel analyses,
including comparisons of amebic strains that differ in virulence, or the stress response pathways.
60
october • fifteen • 2005
Patient-specific modeling of poster
altered hemodynamics and
arterial deformation induced
by coarctation of the aorta
John F.
and subsequent stent
implantation
LaDisa, Jr.
31
Purpose
Stents are playing an increasing role in the treatment of aortic coarctation
(CoA)(1). However, morbidity (hypertension, aneurysms, stroke) persists(2) and
may be attributed to abnormal hemodynamics in the aorta and its branches(3).
We hypothesized that computational fluid dynamics (CFD) modeling can be
used to quantify alterations in blood pressure (BP), flow and vascular deformation introduced by CoA, and stent implantation.
C. Alberto Figueroa
Irene E.
Vignon-Clementel
Frandics P. Chan
Charles A. Taylor
Jeffrey A. Feinstein
Materials and Methods
Magnetic resonance angiography was performed(4) before and after stent implantation in a 13 year old patient with severe CoA. Computational representations of the aorta and cervical arteries were created using custom software that facilitates vascular model construction(5). Time-resolved, 2D phase-contrast magnetic
resonance (PCMR) imaging was performed in the ascending and descending aorta and
cervical arteries. The ascending aorta PCMR waveform was mapped to the aortic inlet. Resistance boundary conditions were imposed at the descending aorta, innominate, left common
carotid (LCCA) and subclavian arteries (LSA) based on cuff BP and PCMR data. CFD simulations were performed using
a validated stabilized finite-element method(6). Systolic BP, diastolic BP and pulse pressure were quantified before and
after stenting and compared with clinical observations. Aortic deformation proximal and distal to the coarctation was
compared to that obtained from PCMR data of spatially equivalent regions.
Results
CoA caused hypertension (134-142 mmHg) and elevated pulse pressure (66-73 mmHg) in the cervical arteries consistent with BP measurements. Pronounced deformation (8%) occurred in the LCCA and LSA with concomitant increases
in blood flow through a collateral artery from the LSA to the descending aorta. SBP decreased after stenting (111-112
mmHg) in arteries proximal to the stenosis and collateral flow was attenuated by 56%. A 17 mmHg gradient across the
coarctation was also reduced to 4 mmHg after stenting, similar to clinical observations. Aortic displacement was consistent with the luminal behavior revealed by PCMR.
Conclusion
These findings demonstrate that CFD modeling can reproduce hemodynamic waveforms and clinical observations in
patients with CoA. The results may ultimately provide insight into sources of morbidity and offer quantitative assessment of treatment options such as stent implantation.
References
1.
Johnston, T. A. et al., Catheter Cardiovasc Interv 62, 499 (2004).
2.
Ou, P. et al., Eur Heart J 25, 1853 (2004).
3.
O’Rourke, M. F. et al., Circulation 44, 281 (1971).
4.
Alley, M. T. et al., Radiographics 18, 273–285 (1998).
5.
Ku, J. P. et al., Ann Biomed Eng 30, 743–752 (2002).
6.
Taylor, C. A. et al., Ann Biomed Eng 26, 1 (1998).
61
biomedical • computation • at • stanford
poster Preferential coplanarity of RNA helices
32
Joseph
Minhow
Chan
and its impact on RNA activity
Purpose
RNA molecules, which are transcribed from DNA and are used by the ribosome to synthesize proteins can also catalyze complex chemical reactions in the cell. As the structure of RNA dictates
function in the cell, it is essential to understand the basic principles that govern RNA structure.
The hypothesis to be tested is that RNA structure can be described through vector analysis. That
is, vectors normal to the RNA bases, and more specifically the RNA helices, are preferentially
oriented bipolarly in the same plane. Although this is a natural tendency for all helices, the
fact that multiple helices of many RNA molecules will often align in a plane despite turns and
loops is striking.
Methodology
Alain Laederach The
distribution of normal vectors to the bases of a given RNA molecule can be described by
using Principal Component Analysis (a statistical method of describing the spread of a given
Armin Schwartz- set of vectors) and fitting the Dimroth-Watson distribution, an antipodally symmetric probability density functions where f(x) = f(-x) [1]. This distribution can fit two general cases onto
man
a dataset: the girdle case (which has a minimum density along the third principal component
and describes a plane) and the bipolar case (which has a maximum density along the first principal
Russ Altman component
and describes a pole). One measurement of how well the model fits the data is dispersion
s of a dataset, the 3rd eigenvalue for the girdle case and the 1st eigenvalue for the bipolar case. The angle
dispersion, a more tangible measurement, can be calculated with the formula: arcsin(sqrt(s)). By fitting a density
function to RNA bases, several applications can be accomplished. First, RNA viewing can be optimized by orienting the viewer’s
vision along the third principal component of the normal vectors to the bases. Second, given the highly planar and bipolar orientation of helices, one can determine how significant helices are to overall molecular planarity and bipolarity, as well as to function. To
do this, one can perform a random helices test in which, for a given molecule, 100,000 random helical orientations of that molecule
are tested for higher planarity or bipolarity against the wild-type conformation. Finally, subdomains of RNA—important regions of
the RNA that, when cleaved from the rest of the molecule, maintain the same conformation—can be identified by local areas of
coplanarity or with specific, multiple bipolar orientations. Tests were performed on t-RNA (PDB ID 1EHZ), the P4-P6 RNA subdomain
of the protozoan Tetrahymena known for its catalytic capabilities (1GID), its supradomain (1GRZ), and rRNA (1FFK). This method
will later be performed on a whole database of RNA molecules. Other models for fitting datasets will be analyzed as well, such as
one that takes into account multiple bipolar concentrations. Such a model would use a k-means clustering approach with clusters
concentrated in distinct bipolar regions.
Results
The angle dispersions for 1EHZ, 1GID, 1GRZ, and 1FFK in the girdle case were respectively 16.63, 18.35, 25.10, and 30.90 degrees
with a maximum value of 35.26 degrees. In the bipolar case, the angle dispersions were respectively 45.66, 29.86, 42.03, and 51.36
degrees with a max value of 54.74 degrees. The results for the random helices test for planarity for these same RNA were as follows
respectively: 0.779%, 2.29%, 4.43%, and 34.7% of the random helical orientations scored a higher planarity than the wild-type. The
results for bipolarity were as follows: 97.9%, 0%, 6.58%, and 56.4% of random helical orientations scored higher bipolarity than the
wild-type. Locating the P4-P6 subdomain of 1GRZ was a success by finding the most planar group of 11 helices in 1GRZ.
Conclusion
The angle dispersions for the girdle case suggest that all the molecules are fairly planar except for 1FFK, the ribosomal subunit. Observation of this molecule confirms this measurement; 1FFK is a very nonplanar molecule. The angle dispersions for bipolarity were
poorer overall. Rather than fitting one bipolar concentration to the data, a more apt model would probably involve multiple bipolar
concentrations. This conclusion is most evident with the result of the random helices test for bipolarity for 1EHZ, which scored
97.9%, a very bad rating. However, a closer look of the t-RNA molecule reveals that it has two distinct, polar helical regions. Overall,
these results largely justify the Dimroth-Watson models as an accurate means of describing the planarity or bipolarity of any set of
vectors. Calculation of p-values is pending. An approach similar to the random helices test might be used to output a computational
p-value. The identification of RNA subdomains using these vector models is very promising, as the P4-P6 subdomain of 1GRZ was
a success.
Resources: 1. Geoffrey S. Watson. Equatorial distributions on a sphere. Biometrika, 52(1/2):193–201, 1965.
62
october • fifteen • 2005
poster
Using coarse-grained simulations
of RNA to explore intermediate
folding pathways
Magda Anna
The structure of large catalytic RNAs (e.g. group I and II introns) is of
interest because of their auto-catalytic self-splicing activity, which
is contingent upon the formation of specific tertiary contacts. We
Jonikas
are exploring the folding of the Tetrahymena thermophila group
33
I intron by running coarse-grained simulations of RNA using the
Alain Laederach
molecular dynamics (MD) package NAST (Nucleic Acid SimulaRuss Altman
tion Tool). Results from time-resolved hydroxyl radical footprinting define specific three-dimensional constraints (in the form of
contact order formation) on the structures of the folding intermediates. By carrying out constrained MD simulations, it is possible to
determine the effect of these experimental constraints on the relative
mobility of the rest of the molecule. Multiple replica exchange MD is used
to improve sampling by regularly exchanging the structures of simulations run at different temperatures on a large super-computer grid. Exchanges are attempted randomly between
runs and accepted or rejected using a metropolis criterion in order to ensure the detailed balance
condition in this Monte Carlo process. Using this approach we make structural predictions for 3 folding intermediates along the folding pathways of the T. thermophila group I intron.
63
poster
34
Justus
Roos
biomedical • computation • at • stanford
Quantitative and qualitative
assessment of a new knowledgebased algorithm for centerline
restoration through femoropopliteal artery occlusions in
peripheral CT angiography
(CTA)
Tejas Rakshe
Matus Straka
Sandy Napel
Dominik
Fleischmann
Purpose
To assess the accuracy of a new knowledge-based algorithm for automated centerline tracking trough femoro-popliteal occlusions in
CTA, a currently unresolved and labor intensive problem.
Material and Methods
CTA was performed in 20 patients (mean age 69y, range 44-88y) with a
total of 26 femoro-popliteal arterial occlusions (mean length132mm,
range 16-400mm),.Occlusions were categorized into <75mm (n=8), 75150mm (n=11), and >150mm (n=7). Two radiologists used a semi-automated centerline extraction technique and bridged occluded segments by manually setting subsequent control points in the center of the vessel. A new knowledge-based
centerline algorithm was used to track the same occluded segments. Accuracy of the algorithm was
quantified by calculating the maximum deviation from the average of the two manually derived centerlines. Diagnostic image quality was compared between curved planar reformations (CPR) generated with both methods.
Results
The knowledge-based algorithm successfully interpolated all femoro-popliteal occlusions in 20 patients. The average of the maximum deviation (in mm [for occlusion length] between the expert
driven and the automated centerline interpolation was 0.1mm [for <75mm], 3.7mm [for 75-150mm],
and 16.7mm [for >150mm]. CPRs based on automated centerlines were indistinguishable from those
based on manual vessel tracking for occlusions shorter than 75mm.
Conclusion
The proposed knowledge-based algorithm allows an accurate automated interpolation of femoropopliteal occlusions up to a length of 75 mm. This may substantially accelerate the creation of CPRs in
patients with peripheral arterial occlusive disease
64
october • fifteen • 2005
A common signature for poster
human aging
35
Purpose
Aging is marked by the gradual decline of a multitude of physiological functions
leading to an increasing probability of death. At the molecular level, we are just
beginning to assemble protein and gene expression changes that can be used
as biomarkers for aging. Rather than search for molecular biomarkers by focusing on only one gene, an attractive approach is to screen all genetic pathways
in parallel for age-related changes by using DNA microarrays to search for
gene expression changes in the elderly.
Materials and Methods
To profile transcriptional changes with age, we used Affymetrix Hu133 2.0
Plus DNA chips, which consist of approximately 55,000 probe sets corresponding to 25,000 unique genes. To determine age-regulated genes, we
implemented a multiple regression model including terms for age, sex, and
tissue type. We chose genes which were significant to a p-value threshold
of 0.001 or better. In our treatment of genetic pathways, we used a modified
Gene Set Enrichment Analysis(GSEA)(Mootha et al. 2003) consisting of a Van
der Waerden non-parametric test to choose genetic pathways that are regulated with age.
Jacob
Michael
Zahn
Rebecca Sonu
Emily Crane
Ralph Rabkin
Ronald W. Davis
Kevin G. Becker
Art B. Owen
Hannes Vogel
Stuart K. Kim
Results
We have identified a molecular profile for aging consisting of 250 age-regulated
genes in human muscle. Of these 250 genes, 92 genes can be shown to be physiological biomarkers of aging in the muscle. We compared the transcriptional profile of muscle aging to previous transcriptional profiles of aging in kidney and brain, and found a common signature for aging
in these diverse human tissues. The common aging signature consists of three genetic pathways: genes in
the extracellular matrix increase expression with age, genes encoding components of the cytosolic ribosome
increase expression with age, and genes encoding subunits of the mitochondrial electron transport chain decrease with age. We also compared transcriptional profiles of aging in human to that of the mouse, and found
that the electron transport chain pathway decreases expression with age in both mammals.
Conclusions
In this study, we have generated a high-resolution profile of aging in the human muscle. Our results provide
the first evidence for a common signature of changes of gene expression in different human tissues. We have
also found the first evidence for a public mammalian signature for aging in the decreased expression of the
electron transport chain.
References
1.) Mootha VK, Lindgren CM, Eriksson KF, Subramanian A, Sihag S et al. (2003) PGC-1alpha-responsive genes
involved in oxidative phosphorylation are coordinately downregulated in human diabetes. Nat Genet 34(3):
267-273.
65
poster
biomedical • computation • at • stanford
Chronic spinal cord injury modifies
aortoiliac biomechanical forces
and morphology and increases
abdominal aortic aneurysm
Hyun Jin Kim disease risk
36
Purpose
Spinal cord injury (SCI) patients have a greater risk of abdominal aortic aneurysms (AAA) compared to age and risk-factor matched ambulatory control
subjects1. We propose that reduced lower-limb blood flow and resistive hemodynamic conditions observed in SCI patients upregulate pro-inflammatory infrarenal aortic gene expression, potentially promoting AAA disease.
We used magnetic resonance (MR) imaging and computational flow modeling to determine the relationship between hemodynamic conditions in SCI
subjects and AAA risk.
Janice J. Yeung
C. Alberto Figueroa
Irene E.
Vignon-Clementel
Mary T. Draney
Material and Methods
Robert J. Herfkens We constructed 3-D computational models from the MRI data of five SCI subjects and six age and risk-factor matched control subjects using custom softRonald L. Dalman ware2.
We solved the 3-D equations of blood flow in deformable models of arteries using a novel finite element method3. Inflow boundary conditions at the
Charles A. Taylor supraceliac
level were specified using PC-MRI data, and impedance boundary conditions were specified at the outlets4. The impedance waveforms were generated using
Womersley’s elastic tube theory and an asymmetric binary fractal tree model of the distal vessels,
using vascular branching laws from the size of the outlet branch to that of the capillaries5.
Results
MRA of the abdominal aorta in all SCI subjects revealed bumpy, ectatic infrarenal aortas and very narrow iliac arteries. PC-MRI data at the level of the infrarenal aortas showed biphasic flow waveforms with little oscillatory flow in
the SCI subjects in comparison to triphasic flow waveforms in control subjects. Computed pressure waveforms of
SCI subjects showed slower decay in diastole. The computed daily averaged wall shear stress of the infrarenal aortic
regions was lower in SCI subjects as compared to control subjects.
Conclusion
Computational methods were used to identify the unique biomechanical forces acting in the abdominal aorta of patients with SCI that may contribute to aortic degeneration, enlargement and increased prevalence of AAA disease.
References
1. Gordon, I.L. et al. Am Surg. 1996;62(3):249-52.
2. Wilson, N.M. et al. Lecture Notes in Computer Science. 2001;2208:449-56.
3. Figueroa, C.A. et al., A coupled momentum method for modeling blood flow in three-dimensional deformable
arteries. Comp. Meth. Appl. Mech. Engng. Accepted for publication in 2005.
4. Vignon-Clementel, I.E. et al., Outflow boundary conditions for three-dimensional finite element modeling of blood
flow and pressure in arteries. Comp. Meth. Appl. Mech. Engng. Accepted for publication in 2005.
5. Steele, B.N. et al., Fractal network model for simulating abdominal and lower extremity blood flow during resting
and exercise conditions. Comput Methods Biomech Biomed Engin. Submitted in 2005.
66
october • fifteen • 2005
Novel approaches to simulating poster
amyloid formation: reaction path
annealing and single molecule
pulling
Jan Lipfert
Purpose
37
Insoluble, misfolded protein aggregates termed amyloid are involved in several
neurodegenerative diseases, including Alzheimer’s and Parkinson’s, and the prion protein’s transition from the native protein fold to amyloid fibrils is believed
to be the underlying mechanism of prion diseases, such as BSE in cattle and CJD
in humans. We employ two completely different computational strategies to investigate the structure, energetics and dynamics of amyloid formation in atomic
detail: “Reaction path annealing”[1] and “single molecule pulling”[2] simulations.
The simulations allow us to compute free energies of binding and asses the energetic barriers for the elementary step in amyloid formation and to highlight the molecular details that contribute to amyloid stability.
Sebastian
Doniach
Methods
We employ the heptapeptide GNNQQNY from yeast prion protein Sup35 as our model system. Eisenberg and
coworkers showed that GNNQQNY has amyloid properties similar to full length Sup35 (as judged by cooperative kinetics of aggregation, fibril formation and cross-beta structure) [3].
The “reaction path annealing” algorithm is developed by our group and uses a
massively parallel implementation to study trajectories connecting the native and amyloid conformations based
on the Onsager-Machlup action. Our “pulling” simulations mimic AFM or optical tweezer single molecule experiments, studying the stability of an amyloid fibril under external mechanical force.
Results
Our two simulation methods allow us to study the transition of a single GNNQQNY strand into the amyloid
conformation and the unbinding of a single strand from an amyloid fibril under external force, respectively. The
computed free energy of binding of ~14 kcal/mol for both methods are in good agreement.
The simulated trajectories highlight the importance of side chain hydrogen bonding for amyloid stability, in
agreement with Max Perutz’ “polar zipper” hypothesis and recent findings by Eisenberg and coworkers [4]. Our
results are consistent with an autocatalytic mechanism of prion formation, whereby the presence of misfolded
protein triggers the conversion from the native to a misfolded state.
Conclusion
Reaction path annealing and single molecule pulling simulations were successfully applied to the problem of
amyloid formation. The results are in agreement with existing experimental findings and allow to study the elementary process of amyloid formation at a level of detail unavailable to experiments.
References
[1] J. Lipfert, J. Franklin, F. Wu and S. Doniach, JMB (2005) 349:648-658
[2] J. Lipfert and S. Doniach, Biophys. J. (2005) 88:401A-401A
[3] M. Balbirnie, R. Grothe and D. S. Eisenberg PNAS (2001) 98:2375-2380
[4] R. Nelson, et. al. Nature (2005) 435:773-778
67
poster
biomedical • computation • at • stanford
Genomic instability signatures of
colorectal carcinoma subtypes
by hierarchical clustering of
Michael Yu gene copy data
Purpose
Zhang
Colorectal carcinoma (CRC) is the second most commonly diagnosed can-
38
Keyan Salari
Kyle Farnam
Jochen Kumm
James Ford
Ron Davis
Hanlee Ji
cer in the United States. The most widely-used indicator of CRC prognosis,
clinical staging, is unfortunately a poor predictor of tumor recurrence. We
developed a novel genomic technology that sheds light on the genomic
instability underlying neoplasia in CRC by looking at gene copy number
alterations to discover clinically relevant biomarkers. These biomarkers
can be used to form signatures which identify subtypes of CRC, providing
a more accurate and complete prognosis.
Materials and Methods
We used the high-resolution, barcode array-based molecular inversion probe
(MIP) [1] technology as a quantitative assay to detect genome-wide gene copy
number alterations. Our probe set targets a large number of candidate cancer
genes throughout the human genome along with a high density of markers along
chromosome 18q, a region containing established CRC prognostic indicators. We analyzed
a panel of 20 CRC cell lines with known genomic instability events, chromosomal (CIN) or microsatellite (MSI),
and a cohort of 33 clinical CRC samples with known staging, stage II or III. Log2 ratios for this data set were calculated against scale normalized control data from a set of 32 normal cell lines. We implemented the statistical
method clustering along chromosomes (CLAC) [2] to make amplification/deletion significance calls with a false
discovery rate of 0.01. Probes called significant received a score based on their log2 ratios. Scores of zero were
assigned to all other probes. Finally, we performed agglomerative hierarchical clustering analyses on the CRC
cell lines and clinical samples using their CLAC significance scores.
Results
Hierarchical clustering of the CRC cell lines yielded two distinct clusters corresponding to the two types of
genomic instability. CIN was characterized by concentrated deletions along chromosome 18q, while MSI was
characterized by sparser, more heterogeneous gene copy alterations. The type of genomic instability was a
statistically significant predictor of cluster group (p=0.0093). Clustering of the clinical CRC samples yielded two
major clusters; however, they did not correspond significantly to staging category (p=0.22).
Conclusion
The stark distinction of the two clusters within the clinical samples and their non-correspondence to tissue
stage affirms the insufficiency of staging for prognosis and reveals potential subtypes of colon cancer as delineated by genomic instability signatures. The ability to differentiate between these subtypes hold promise in
next generation clinical prognostics.
References
1. Hardenbol, P. et al. (2003). Nat Biotechnol 21(6): 673-8.
2. Wang, P. et al. (2005). Biostatistics 6(1): 45-58.
68
october • fifteen • 2005
Automatic state decomposition poster
for constructing Markov models
for macromolecular dynamics
Purpose
Nina Singhal
Understanding the kinetics of protein folding is an important biologi-
39
cal problem. Supercomputing, such as the Folding@Home project, can
generate large amounts of molecular dynamics simulation data. In
previous studies, we developed efficient models for analyzing these
large data sets that divided the conformation space into discrete
states and calculated the transition probabilities between the states,
assuming Markovian, or history independent, transitions. In this work,
we develop algorithms for the automatic decomposition of the protein
or other macromolecular conformation space into regions which are kinetically related and transitions between them Markovian.
John D. Chodera
Vijay S. Pande
William C. Swope
Methods
We first split conformations which are significantly dissimilar by some distance metric. In this work,
we examined torsion angle space and root mean square deviations. Following this initial decomposition,
we build a transition matrix and examine the eigenvalues and eigenvectors of this matrix to group similar
states. These states are iteratively refined by repeating this procedure.
Results
We apply these techniques to alanine dipeptide and compare the resulting states with a manual decomposition of the state space. We find our results in good agreement with the manual decomposition over
our distance metrics.
Conclusion
We have successfully implemented a new algorithm for finding kinetically relevant states from molecular
dynamics simulation data. This is an important first step in building a model of the folding process which
can describe the long timescale behavior. We plan to apply these algorithms to more complicated molecular systems in the future.
69
poster
40
biomedical • computation • at • stanford
Fold removal in CT colonography
Padma
Sundaram
Eftychios Sifakis
David Paik
Christopher
Beaulieu
Sandy Napel
Purpose
Normal haustral folds complicate both, the visual and computer-aided
detection (CAD) of polyps in CTC data. We present a method to straighten and flatten the colon by simulating stretching of the colonic surface
using a quasistatic Finite Element Model (FEM). The output of our algorithm is a 3D fold-free surface that can be displayed as a single image
and/or subjected to a CAD algorithm.
Materials and methods
We created several mathematical phantoms using MATLAB 7.0.1, with
folds and polyps modeled as half sine functions and hemispheres, respectively. The phantoms had folds, polyps on flat regions and polyps on
folds. We also tested our algorithm on a subvolume of actual patient data,
acquired under our IRB during a research CTC scan at our institution, containing a 6.9 mm polyp. Our algorithm stretches the surface in a direction normal to the
direction of the folds, resulting in flattening the folds, with minimal distortion to polyps. A spatially invariant response to pulling from both ends is essential, otherwise structures closer to the edges being
pulled will be more distorted than those farther away. Thus, we implemented a quasistatic assumption in our FEM so as to neglect inertial effects. To evaluate our methods we measured the curvature
and size of the polyps (diameters) and folds (heights) before and after simulated stretching.
Results
On phantom and patient data, our method shows reductions in fold height and curvature ranging
from 54.4% to 70.3%, and 36.3% to 86.1% respectively. Polyp size and curvature were reduced by
only 0% to 16%, and 0% to 20%, respectively. The output of our algorithm is a 3d surface that can be
viewed as a single image.
Conclusions
Our physics-based method to flatten the colonic surface using a quasistatic finite element model
suppresses folds while leaving polyps relatively undistorted, and has the potential to improve the efficiency of interpretation by reducing visual clutter and sources of false positive detections by CAD.
Continued development is warranted.
70
october • fifteen • 2005
Analysis of the transcriptional
regulation of a bacterial cell
poster
41
Patrick
Timothy
McGrath
We have created an Affymetrix chip for Caulobacter crescentus that is
both an expression array and an intergenic tiling array. This array can
be used for analysis of gene expression, for high-resolution chromatin
immunoprecipitation experiments, for identification of expression in
intergenic regions such as non-coding RNAs (ncRNAs), and for identification of transcriptional start sites. We report here the creation of a
a novel algorithm to identify transcriptional start sites by analyzing
Hong-Lok Lee
the cross-correlation between all probes found between 200 base
Antonio Iniesta
pairs upstream of the predicted start codon and the stop codon of
each gene. We have identified transcriptional start sites of ~600 opAlison Hottes
erons to an accuracy of 5 base pairs. Motif searching of the DNA sequence upstream of the predicted +1 sites revealed the existence of
Lucy Shapiro
several novel motifs. We have also extended this approach to identify
genes with multiple plus one sites. This approach has identified over Harley McAdams
half (4 of 7) of previously identified genes known to have multiple plus
one sites. We further predict 50 new genes to be transcribed from multiple
plus one sites.
71
poster
biomedical • computation • at • stanford
2D ultrasound image processing
in identifying displacement
history of urogenital
structures to pelvic floor
Qiyu Peng
muscle activity
42
Ruth C. Jones
Kirk Robert McConnell
Chris E. Constantinou
Purpose
The measurement of the PFM’ ability to contract voluntarily is of value in the assessment of women with urinary or fecal incontinence. Ultrasound imaging provides
a considerable amount of dynamic data that cannot be visually assimilated by the
observer, particularly during fast events such as coughing. The purpose of this study
is to gain greater understanding of the biomechanical parameters of the PFM, especially during fast, stress events by using 2-D ultrasound imaging.
Methods
1. Experimental protocols
Consenting healthy and SUI female volunteers, age range 21-54 were examined in the supine
lithotomy position by ultrasound using a ultrasound scanner with a 3.5 MHz high definition curved linear array probe with a footprint of 10cm. With the probe in place, volunteers were asked to perform 3 maneuvers: a slow maximum
effort of voluntary PFM contraction with a 5 second hold then release; a cough; and a Valsalva maneuver.
2. Image processing
The procedures of image processing and segmentation are shown in Fig. 1.
3. Establishment of the coordinate system
After the segmentation, a coordinate system fixed on the posterior inferior margin of the symphysis pubis, assumed to be ellipse
in shape, was established (Fig. 2).
4. Measurement of the movement
Three typical positions (P1, P2 and P3 in Fig. 2 were chosen to quantitatively describe the movement of each structure.
Results
1. Movement of the edges during activities
Fig. 3 shows the movement history of the anterior and posterior edges of the UVJ and the ano-rectal junction during cough, contraction and Valsalva.
2. Direction of the movement of tissues
The directions of the movement of the three positions during cough, contraction and Valsalva were derived by linear regression
and shown in Fig. 4.
3. Trajectories of the movement of tissues
Fig. 5 shows trajectories of the three positions on the anterior and posterior edges of the UVJ and on the ano-rectal junction during cough, contraction and Valsalva.
4. Kinematics analysis of the movement
Fig.6-8 shows displacement, velocity and acceleration during cough, contraction and Valsalva. The positions on anterior and posterior edges of UVJ and ano-rectal junction are shown in solid, dotted and dash lines respectively.
Conclusion
The movement of anterior vaginal wall during cough was also measured by a novel palpation probe designed to evaluate function
of pelvic floor muscle. Fig. 9 compares the displacements measured by ultrasound on anterior UVJ and by the probe on anterior
vaginal wall.
The presented methods are able to visualize, quantitatively measure, analyze and describe the outlines of the urethra, bladder and
rectal interfaces and
define new parameters of PFM function.
References
1.
Tunn R, Petri E. Introital and trans-vaginal ultrasound as the main tool in the assessment of urogenital and pelvic floor dysfunction: an imaging panel and practical approach. Ultrasound Obstet Gynecol. 2003; 22(2):205-13.
2.
Reddy A, Delancey JOL, Zwica LM, Ashton-Miller J. On screen-vector based ultrasound assessment of vesical neck movement. Am J Obst Gynecol. 2001; 65-70.
72
october • fifteen • 2005
A novel palpation system to
evaluate internal organs of
genitourinary system
Introduction
Palpation provides information relating the mechanical properties of the tissue such as its stiffness. In clinics, palpation is used to evaluate in accessible
internal organs of the genitourinary system such as the prostate, vagina and
uterus. We developed a novel palpation system suitable for the evaluation
of the static and dynamic properties of internal organs.
poster
43
Qiyu Peng
Ruth C. Jones
Kirk Robert
McConnell
Sadao Omata
Chris E.
Constantinou
Prototype of the System
The prototype of the system is made of a probe with four force and displacement sensors and a six degrees-of-freedom measuring device that can track
the position (x, y, z) and orientation (azimuth, elevation and roll angles) of the
probe (Flock of Birds, FOB) (Fig. 1).
The probe is made of an applicator with retractable low inertia cantilevered force and
displacement transducers in four directions (anterior, posterior, left and right). The transducers
are closed until they are inserted into the internal organ and touch the organ wall (Fig.2).
The sensor of FOB is fixed tightly on the handle of the probe. The coordinate of the point O’ (x0, y0, z0 ) and the orientation of the axes x’-y’-z’ (θa, θe, θr; namely, azimuth, elevation and roll angles) on the global coordinate system (Fig. 1)
were measured in real time by FOB.
Phantom Experiments
1. Static measurement
Phantom experiments were performed to test the probe. The internal cross sections of the phantoms, which are assumed to be elliptic, were derived from the displacements measured by the displacement sensors and the position and
orientation of the probe measured by FOB (Fig. 3).
Three phantoms (Fig. 4 a, b, c), which simulate three typical geometric and mechanical situations of the internal organ
of the genitourinary system, are designed to evaluate the static performances of the palpation system. 3D reconstruction of the phantoms are shown in Fig. 4 (d), (e), (f ). In (e) and (f ), the force signals in four directions are color-coded.
2. Dynamic measurements
A mechanical vibrator was used to simulate the activities of organ by perturbing the sensors on the applicator. The
results show the probe has ideal frequency response to low-frequency activities of organs.
Clinical Experiments
The palpation system was used to evaluate the functions of Pelvic Floor Muscles (PFM). Thirteen healthy volunteers and
six patients with Urinary Incontinence were tested. The typical force and displacement signals recorded during the static
measurements (profile of vaginal wall) and dynamic (cough) measurements are shown in Fig. 5. Detailed analysis of the
pilot data is ongoing.
Conclusions
The novel palpation system can be used to evaluate the static and dynamic properties of internal organ of Genitourinary system.
Acknowledgments
NIH-NIBIB grant 1 R21 EB001654
Biodesign, Clark center, Stanford University
73
poster
44
Rashmi
Raghu
biomedical • computation • at • stanford
Viscoelastic arterial wall model for
one-dimensional modeling of
blood flow and pressure
Purpose
Irene E.
Vignon-Clementel
C. Alberto Figueroa
Charles A. Taylor
The one-dimensional equations of blood flow consist of conservation of
mass, conservation of momentum and a constitutive relationship that describes the wall properties. The variables at a given point in the system are
the arterial flow rate, cross-sectional area and pressure. To date, only nonlinear elastic constitutive models have been used in these types of simulations. However, arterial tissue is known to have viscoelastic properties.
The purpose of this work is to derive and incorporate a viscoelastic constitutive model for the arterial wall in the existing one-dimensional equations of blood flow and to implement it in the corresponding finite element
simulation environment.
Materials and Methods
The viscoelastic arterial model used in the current work is based on Holenstein et al (1980).
The relevant viscoelastic relationship was derived from that work along with Fung (1993) and used to relate
pressure and cross-sectional area. This relationship consists of two parts: the instantaneous elastic response
of pressure and the convolution of the viscoelastic relaxation function with the history of instantaneous elastic
responses. The appropriate relaxation function was derived from a creep function based on Holenstein et al
(1980). The Olufsen (1999) non-linear elastic model is used as the instantaneous elastic response in the viscoelastic model.
Results
Preliminary results show that viscoelasticity has an impact on flow and pressure waves – viscoelasticity causes
internal energy dissipation. In a carotid model, with prescribed inlet flow and a resistance boundary condition
at the outlet, peak pressure and flow at the outlet were lower in the viscoelastic case than in the elastic case.
Plotting pressure against cross-sectional area also shows some hysteresis in the viscoelastic case unlike the
elastic case.
Conclusion
A viscoelastic arterial wall model has been derived for one-dimensional blood flow problems. The implementation of the model in a finite element code has allowed the investigation of differences between the non-linear
elastic and viscoelastic models. Further comparison studies of the elastic and viscoelastic constitutive models will be conducted with networks of vessels for a more thorough analysis of the effects of the viscoelastic
model.
References
Fung, Y.C. (1993) Biomechanics: Mechanical Properties of Living Tissues, 2nd edition, Springer-Verlag, New
York.
Holenstein, R., Niederer, P. and Anliker, M. (1980) “A Viscoelastic Model for Use in Predicting Arterial Pulse Waves”,
Journal of Biomechanical Engineering, 102, 318-325.
Olufsen, M. (1999) “Structured Tree Outflow Condition for Blood Flow in Larger Systemic Arteries”, American
Journal of Physiology, 276, 257-268.
74
october • fifteen • 2005
Differential gene expression poster
between healthy and melanoma
patients for T cells, B cells and
NK cells
Rebecca
Purpose
The purpose was to determine whether any genes were significantly under or
over expressed in the B, CD4, CD8, or NK cells of melanoma patients.
Meredith
Materials and Methods
Loew
Agilent’s feature extraction software scanned the microarrays and produced
45
txt files. R was then used to analyze the data. The limma package in R read in
the txt files and created RG objects, which stored information such as the red
and green frequencies for each probe and its background. From this, the control spot data and data for 172 genes frequently suffering from dye saturation
were removed using the flags provided by Agilent. A correlation matrix revealed
a strong correlation between the pooled samples and the averages of the patients represented. It also showed a good negative correlation between the dye
swap arrays. Next, the loess and variance stabilizing normalization methods were
compared by plotting the normalized data for genes known to be expressed. Since
vsn produced tighter clusters, it was selected. The multtest package from Bioconductor
was used for multiple testing, and Westfall and Young’s step-down permutation method was
used to correct for the multiple testing bias. In order to increase the power of the study, CD4, CD8, and B cells were
combined.
Rebecca CritchleyThorne
Peter Lee
Susan Holmes
Results
From the combined CD4, CD8, and B cells, 29 genes emerged significant. Hierarchical clustering and heatmaps
were created using this list of genes. Lists of upstream nucleotides were also compiled for significant genes for
the purpose of motif discovery.
Conclusion
The high number of significant genes is encouraging. Hopefully, network analysis will related these genes with
known pathways.
References
Holmes, S., He, M., Xu, T., Lee, P. (2005) Memory T cells have gene expression patterns intermediate between naive
and effector PNAS, Vol. 102, no. 15, 5519-5523. http://www.pnas.org/cgi/reprint/102/15/5519
Kent, W. J., Sugnet, C. W., Furey, T. S., Roskin, K. M., Pringle, T. H., Zahler, A. M., and Haussler, D. (2002) The Human
Genome Browser at UCSC . http://www.genome.ucsc.edu
R Development Core Team. (2005) R: A Language and environment for statistical computing. R Foundation for
Statistical Computing, Vienna, Austria. ISBN 3-900051- 07-0. http://www.R-project.org
Tong Xu, Chen-Tsen Shu, Elizabeth Purdom, Demi Dang, Diane Ilsley, Yaqian Guo, Jeffrey Weber, Susan P. Holmes,
and Peter P. Lee Microarray Analysis Reveals Differences in Gene Expression of Circulating CD8+ T Cells in Melanoma Patients and Healthy Donors. Cancer Res. 2004 64: 3661-3667.
75
poster
biomedical • computation • at • stanford
46
Mechanics of deformation of
bone cells
Ron Y. Kwon Purpose
Sheyi A. Ayeni
Christopher R.
Jacobs
Fluid flow can be a potent loading mechanism for eliciting biochemical responses in bone cells. However, understanding the mechanism
by which fluid flow-induced mechanical loads are transduced into
chemical signals is difficult since the mechanics of flow-induced bone
cell deformation are not well understood. The objectives of this study
were to (1) develop a methodology for determining fluid flow-induced cellular deformations in vitro, and (2) use the methodology to
investigate the mechanics of osteoblastic cells exposed to steady and
oscillatory fluid flow.
Materials and Methods
1.0 micron diameter fluorescent beads were added to MC3T3-E1 cells. For each cell,
its beads were imaged during no flow, oscillatory flow (1.0 Pa peak sinusoidal shear stress at 1 Hz) and
steady flow (10s at 1.0Pa shear stress). Cross-correlation fields were calculated by constructing kernels
from bead intensity fields, and convolving the kernels with the bead images. Bead coordinates were calculated from the cross-correlation fields, allowing displacement curves to be generated. For steady flow,
bead displacement curves of each cell were fit to the creep response of a Kelvin body (standard linear
solid) using a nonlinear least squares method. The 2D Green Lagrange strain tensor Eij was determined
by creating bead triads using Delaunay triangulation, interpolating the displacements within each triad,
and differentiating the resulting displacement fields. From Eij, the principal strain and orientation were
determined.
Results
Fitting of the creep response during steady flow to the Kelvin model resulted in relaxation times of 1.7 ±
0.8 (± S.E., n=3) (Fig 1), the first such values reported for osteoblastic cells. The similarity of the relaxation
times (~1.7s) to the period of oscillations (1s), as well as the larger displacements observed during steady
flow as opposed to oscillatory flow, suggests that differences in the response of MC3T3-E1 cells to oscillatory flow (1 Hz) and steady flow are due to the inability of the cells to undergo large deformations during
oscillatory flow. The average strains of 1.0~3.5% (Fig 2) are of the same magnitude as those reported to
elicit osteogenic responses in osteoblasts subjected to substrate stretch.
Conclusion
The methodology presented allows in vitro quantification of cellular deformation. Future studies will investigate the correlation between deformation and mechanosensitive responses such as kinase activity
or protein localization, with the overall goal of relating of fluid-flow induced deformation and subsequent
chemical signaling.
76
october • fifteen • 2005
poster
A novel software for plotting
multidimensional FACS data
Purpose
Flow cytometry (FACS) is a method for counting, examining, and sorting cells
that are suspended in a stream of fluid. Its applications are numerous in both
basic science and clinical research. The power of FACS stems from its ability to simultaneously measure the levels of many proteins (currently up to
twelve) on a single cell. However, current software for FACS analysis is limited to 2D visualizations, usually scatter or contour plots. Even in FlowJo, the
field’s most advanced software, a multidimensional interpretation of FACS
data requires considerable effort. We present a novel software “FACSwizard”
that is able to produce dynamic 3D and 4D representations of FACS data.
47
Ryan J.
Tibshirani
Materials and Methods
We have developed three techniques for data visualization.
(1) 3D setting: We have created scatter and contour plots that the user can freely rotate using the mouse.
Here, the main implementation challenge is to produce an accurate and fast contouring of the data; our
method is an application of the marching cubes algorithm, a technique used to render 3D visualizations
of human anatomy from CT and MR imaging.
(2) 4D setting, trellis plot: Data are incrementally stratified on two parameters, and correspondingly a
number of 2D plots displaying the other two parameters are laid out in a grid.
(3) 4D setting, dynamically-gated plot: The user performs custom stratification by drawing with a polygonal gating tool on a 2D plot. A second 2D plot displays the gated population along two other dimensions.
By using the mouse to drag the gate in the first plot, the user is able to simultaneously examine the newly
stratified population in the second plot.
Results
Our dynamic 3D plots provide a highly intuitive interpretation of 3D FACS data. As such, a single 3D plot
created by FACSwizard may provide more insight than three 2D plots produced by FlowJo. Our trellis plot
offers an immediate and global perspective of 4D data, which would be especially useful in a research
paper. Finally, our dynamically-gated plot is a novel idea that may be a very powerful tool for certain data.
In the case of data consisting of two cell surface markers and two phosphoproteins, for example, this
technique provides a rapid and flexible way of examining the phospho-responses of specific cell subsets.
We refer the reader to our website http://www.stanford.edu/~ryantibs/bcats05 for examples illustrating
these results.
Conclusion
Our software offers all of the standard 2D plotting utilities for FACS data, as well as novel 3D and 4D plotting techniques. These techniques allow fluid and intuitive manipulations of multidimensional plots, and
this can potentially yield new biological insights from the analysis of multiparametric FACS data.
77
poster
48
biomedical • computation • at • stanford
Shaohua
Sun
Registration of lung nodules
using a semi-rigid model
Purpose
The tracking of lung nodules in sequential CT scans for the same patient is helpful for determination of malignancy [1, 2]. We propose an automatic semi-rigid
registration scheme for tracking lung nodules over time that is robust to size/
shape changes.
Geoffrey D. Rubin
Sandy Napel
Material and Methods
Following the lung segmentation, registration is performed for each nodule,
considering only large structures surrounding the nodule. We first delete all
voxels whose gray levels are less than 200 in a 60mmx60mmx60mm cube centered on a nodule in the source scan, followed by 3D connected component
analysis. We retain the component containing the nodule and the 7 largest remaining components, and allow each of these to undergo an affine transformation
independently. In the target scan, corresponding structures are found by transforming
the source structures. The nodule location in the target scan is determined by finding the transformation parameter set that maximizes an energy function of the 8 structure pairs in the source and target volumes. This
function consists of normalized correlation coefficient measuring structural similarity and the variance of the
parameter sets measuring relative movement amongst the structures. The proposed maximization scheme is
a two-layered optimization method, employing a Simulated Annealing [3] framework.
Results
We tested our method with 10 nodule pairs by simulating nodules resolve, grow and shrink uniformly by a
factor of 2, and grow non-uniformly. A target scan was simulated by transforming a source lung CT scan; nodules were simulated using a CT simulator and embedded in the source and at the transformed location in the
target. The transformation from source to target consisted of translation, rotation, and non-rigid transformation based on inspiration and cardiac models. The mean absolute distance errors for the four cases were 1.13
mm+/-1.08 mm(s.d.), 0.97 mm+/-1.10 mm(s.d.), 0.72 mm+/-0.73 mm(s.d.), and 1.09 mm+/-1.36 mm(s.d.). We
also matched 92 nodules in 12 pairs of patient scans and compared our registration to a radiologist’s visual
determination. The mean absolute distance error was 1.16 mm+/-0.79 mm(s.d.).
Conclusion
The proposed semi-rigid model shows promise for matching nodules in a series of CT scans and is robust to
size/shape change over time.
References
[1] C.I. Henschke, D.F. Yankelevitz, D. Libby and M. Kimmel, “CT screening for lung cancer: the first ten years,”
Cancer J., vol. 8 Suppl 1, pp. S47-54, May-Jun. 2002.
[2] J.P.W. Pluim and J.M. Fitzpatrick, “Image registration,” IEEE Trans.Med. Imaging, vol. 22, pp. 1341-1343,
2003.
[3] L. Ingber, “Very fast simulated re-annealing,” Mathematical and Computer Modelling, vol. 12, pp. 967-973,
1989.
78
october • fifteen • 2005
A metropolis sampling framework poster
for inferring neural connectivity
in diffusion tensor images
Purpose
Anthony
Diffusion Tensor Imaging (DTI) is a magnetic resonance (MR) imaging method that
can be used to measure the local information about the structure of the white matter within the human brain. Combining DTI data with the computational methods Sherbondy
49
of MR Tractography, neuroscientists can estimate the locations and sizes of nerve
bundles (white matter pathways) that course through the human brain.
Current deterministic tractography algorithms have become quite good at reconstructing certain white matter structures such as the corona radiata and the superior and inferior longitudinal fasciculi, but these algorithms have also met with
consistent reproducible failure when attempting to reconstruct some sets of neural
pathways.
David Akers
Pat Hanrahan
One major set of challenging pathways include those that course temporally between the corpus
callosum and the cortical surface. There is strong evidence of the existence of these temporal connections between the left and right hemispheres. The evidence has been gathered from other measuring modalities, such as
functional MRI (fMRI), cadaver and primate studies, yet current tractography algorithms are invariably unsuccessful at recovering these pathways. Our work contributes a new Bayesian computational framework for performing
tractography through the sections of DTI data that have been a well recognized problem for previous tractography algorithms.
Materials and Methods:
Our tractography algorithm is designed to efficiently sample possible paths that connect user-defined regions of
the brain together. This approach allows a neuroscientist to use our algorithm in order to reason about connectivity in regions of DTI images that simply are not even possible with previous tractography algorithms.
We first show how to formulate our connectivity problem in terms suited for application of Bayesian inference
techniques. A defined posterior distribution on pathways encapsulates the neuroscientists’ prior knowledge of
pathway existence as well as beliefs of what would make a pathway more or less likely given some DTI information
along the pathway.
While it is true that Bayesian algorithms have been proposed by other researchers in this field, our algorithm offers
computational advantages over previous approaches when considering the problem of sampling pathways that
must connect two defined regions in the data. Specifically, we present a Markov Chain Monte Carlo path sampling algorithm to address the computational issues of sampling only pathways that connect the two regions.
Results
This work is illustrated by results from mathematically constructed DTI phantoms as well as real DTI data from a
normal human subject.
79
poster
biomedical • computation • at • stanford
Modeling of phosphorylation sites
in protein structures using the
FEATURE system
Shirley Wu Purpose
50
Megan Y. So
Russ B. Altman
Phosphorylation of proteins regulates a large number of biological processes in organisms. The ability to predict phosphorylation sites in new and
existing protein structures would contribute greatly to our understanding
and manipulation of these pathways. We use an existing method, FEATURE, to characterize the 3-D environment around these sites and to illuminate areas in which we can improve the modeling of variable, low
density data sets.
Materials and Methods
FEATURE is a supervised learning algorithm that uses a naive Bayes method
for distinguishing between provided positive and negative training sets of sites
in protein structures. The 3-D space around each data point is divided into radial shells,
and the physicochemical properties present in each shell are input into a vector. The significance of each
property in each shell given the two training sets is calculated and goes into a model that describes the
site. All protein structures used are derived from the Protein Data Bank (PDB). Phosphorylation site data
are derived from the PDB and from the Phospho.ELM database of experimentally verified phosphorylation sites. Negative training sites were randomly chosen from a non-redundant subset of the PDB. Sequence alignment to eliminate homology was done using ClustalW.
Results
Separation between positive and negative sites using positive sites derived from annotated phosphoresidues in the PDB was fair, with the Threonine model showing the best separation. Using positive sites
derived from Phospho.ELM, separation in the models was much worse. Model fingerprints showed very
few properties that differed significantly between positive and negative sites.
Conclusion
Reliable data on phosphorylation sites is still difficult to obtain. Although FEATURE was unable to build
high-quality models with this data using default settings, there may still be some retrievable signal within
the noise. More investigation into the selection of training sets and the target granularity of models may
greatly improve the performance of a structural model of phosphorylation sites.
References
Wei L, Altman RB. Recognizing protein binding sites using statistical descriptions of their 3D environments. Pac Symp Biocomput. 1998;:497-508.
H.M. Berman, J. Westbrook, Z. Feng, G. Gilliland, T.N. Bhat, H. Weissig, I.N. Shindyalov, P.E. Bourne: The Protein Data Bank. Nucleic Acids Research, 28 pp. 235-242 (2000).
Diella F et al. Phospho.ELM: A database of experimentally verified phosphorylation sites in eukaryotic
proteins. BMC Bioinformatics 2004, 5:79.
80
october • fifteen • 2005
Identifying regulatory mechanisms poster
associated with genetic variations
among individuals
Motivation
Su-In Lee
Genomic sequence polymorphisms create variations in the expression of
51
genes by perturbing the complex network of regulatory interactions. Much
work has focused on identifying quantitative trait loci (QTLs) that affect the
expression of genes1-3. However, there has been little work on identifying
the regulatory mechanisms associated with such sequence polymorphisms,
a task that can play an important role in understanding genetic diseases,
and suggesting targets for drug design.
Aimee Dudley
Dana Pe’er
Daphne Koller
Methods
We present a method based on probabilistic models4 that can provide a finegrained understanding of the regulatory mechanisms caused by genetic sequence
polymorphisms. Given the genomic expression and the marker-based genotype profile of many individuals, our method provides a mechanistic explanation for the association between gene expression and genetic variation. In our model, the expression of a gene is affected by three
kinds of factors: the sequence polymorphisms of the gene itself, the sequence polymorphisms of a regulator and the expression level of a regulator. The expression profile of a module of co-regulated genes is
induced by combinatorial relationships among these factors.
Our approach extends the previous genetic linkage analysis methods1-3 in three ways: (1) incorporating
the expression level of regulators as a factor in expression variation, (2) identifying combinatorial regulatory interactions, and (3) associating regulatory models with modules of genes, providing higher statistical significance for detecting the cause of the expression variation.
Results
We applied our method to a data set of recombinants generated by a cross between two strains of S.
cerevisiae1, resulting in fine-grained hypotheses about regulatory interactions associated with genetic
variations in yeast. We show that it outperforms the classical method1, both in statistical measures such
as the proportion of genetic variance explained, and in recovering biological relationships known in the
literature. Our method provided several well-supported novel hypotheses associating mechanistic explanations with polymorphisms. We experimentally verified one of these hypotheses in the wet-lab, validating the ability of our approach to detect indirect effects of a QTL on its target genes.
References
1.
Brem, R.B. & Kruglyak, L. PNAS 102, 1572-7 (2005).
2.
Morley, M. et al. Nature 430, 743-7 (2004).
3.
Schadt, E.E. et al. Nature 422, 297-302 (2003).
4.
Pearl, J. Probabilistic Reasoning in Intelligent Systems (1998).
81
poster
biomedical • computation • at • stanford
CONTRAST: de novo gene
prediction using a semi-Markov
conditional random field
Samuel S. Purpose
De novo gene prediction, defined as the identification of the location
and structure of protein coding genes using genomic sequences alone, reGross
mains an important problem. Although full-length experimentally verified
52
transcripts exist for approximately 60-85% of all human genes, traditional
experimental methods for identifying genes based on 5’ EST sampling and
cDNA clone sequencing are now reaching the point of diminishing returns.
As a result, efforts to identify new genes by RT-PCR from computationally predicted gene structures are taking on greater importance. De novo
gene predictors are particularly well-suited to this task as they do not rely on
cDNA or EST evidence, which is unlikely to exist for currently unknown human
genes. We present CONTRAST (CONditionally TRAined Search for Transcripts),
the first de novo gene predictor based on a probabilistic model known as a semiMarkov conditional random field (sCRF).
Chuong B. Do
Serafim Batzoglou
Materials and Methods
Almost every successful modern gene predictor is based on a generative probabilistic model known as a hidden Markov model (HMM). Although HMMs are easy to train, they have the disadvantage that they do not
directly model model the conditional probability P(Y|X) of a gene prediction Y given a sequence X. Instead,
normal HMM training optimizes the joint probability P(X,Y) = P(Y|X) P(X), which means a set of parameters
that leads to worse predictions may be prefered if it leads to a better value of P(X), a quantity irrelevant to the
problem of gene prediction. sCRFs avoid this problem by directly
modeling the conditional probability distribution while at the same time relaxing many of the strict independence assumptions required by HMMs. This allows for much more expressive models using a far richer set of
possible sequence features. We implemented a sCRF-based gene predictor, CONTRAST, and tested it on the
Drosophila melanogaster genome.
Results
In our tests on the D. melanogaster genome, CONTRAST performs significantly better than GENSCAN, one of
the most accurate single genome gene predictors. In particular, CONTRAST is able to avoid GENSCAN’s wellknown tendancy toward the prediction of many false positive genes.
Conclusion
We have shown that gene predictors based on sCRFs can be significantly more accurate than those based on
HMMs. There are many promising avenues of future research to be pursued in sCRF-based gene prediction. In
particular, we plan to extend CONTRAST’s relatively simple gene model, add 5’ and 3’ UTR prediction capabilities, and add the ability to use multiple alignments to increase predictive accuracy.
82
october • fifteen • 2005
Removal of respiratory motion
artifacts and the construction of
4D PET
poster
53
Brian
Thorndyke
Purpose
Respiratory motion remains a challenging problem in positron emission tomography (PET) imaging of thoracic and abdominal cancer.
While PET has proven to be a valuable adjunct in the detection,
delineation and characterization of metabolically hyperactive tissue, the radionuclide signal may be significantly degraded by the
motion of the abdominothoracic organs. The most common solution is to “gate” the scan, effectively limiting image acquisition to
short-time interval snapshots corresponding to a particular point in
the respiratory cycle. One major drawback to gated PET is reduced
signal-to-noise ratio owing to the decimation of coincidence events. A
second problem is the limitation to a single gated interval, precluding the
possibility of following the PET image over time. We describe a method, coined
“restrospective stacking”, which both restores image quality and enables the generation of a timeresolved (4D) PET image from end-inhale to end-exhale.
Eduard Schreibmann
Lei Xing
Materials and Methods
Retrospective stacking uses b-spline deformable image registration to combine amplitude-binned
PET data along the entire respiratory cycle into a single respiratory end-point. The motion maps derived from the b-spline registration process can subsequently be used to generate a PET image at any
arbitrary point in the cycle. The technique was applied to F-18 FDG PET images of a pancreatic cancer
patient with a liver metastasis, whose lesion was poorly visible through gated PET. A restrospectively stacked image was generated at end-inhale, and this image was then interpolated at several
points from end-inhale to end-exhale. Cross section visualizations and contrast-to-noise ratios (CNRs)
at each respiratory point provided the basis of comparison between ungated, gated and retrospectively stacked PET.
Results
Both retrospective stacking and gated acquisition removed lesion motion artifacts. Retrospective
stacking, however, greatly enhanced the visibility of the tumor by increasing the CNR of the malignant to surrounding tissue by over 3-fold. The lesion and PET-visible organ motion was followed from
end-inhale to end-exhale using b-spline interpolation, and the image enhancement was maintained
throughout.
Conclusion
Patient data have demonstrated that retrospective stacking can correct for respiratory motion artifacts, significantly improving tumor visibility while providing a view of the tumor and surrounding
PET-visible tissue at any point in the respiratory cycle.
83
poster
54
biomedical • computation • at • stanford
Ion size effects and the PoissonBoltzmann equation
Vincent
Bangping
Chu
The ion atmospheres around nucleic acids play a significant role in their
structure and function. Due to their charged backbones, formation of tertiary structure in nucleic acids occurs upon the addition of counterions in
solution which screen the phosphate charges. Greater understanding of
the role of electrostatics in nucleic acid structure formation may yield insight into many outstanding problems in biology, such as the ribozyme
folding and gene regulation, and may lead to new drug designs and advanced molecular sensor development.
Jan Lipfert
Yu Bai
Poisson-Boltzmann (PB) theory has historically been the most widely
used framework in modeling electrostatics around biological moleRhiju Das
cules. Unfortunately, because it does not consider the finite sizes of ions
Sebastian Doniach in solution it is inaccurate, especially near the highly charged nucleic acid
surfaces. We present an extension based on earlier work that allows us to
Daniel Herschlag capture
these finite size effects in a lattice gas framework.
We have successfully prototyped the new model in Matlab for simple systems.
We are currently implementing and testing the new model in an existing PB Solver,
APBS, making extensive use of computing facilities in our group and the Bio-X cluster in order to
tackle even larger, biologically relevant systems.
References
Bai, Y., Das, R., Millett, I., Herschlag, D., and Doniach, S. Probing Counterion Modulated
Repulsion and Attraction Between Nucleic Acid Duplexes in Solution. Proc. Natl. Acad.
Sci. USA, 102(4):1035{1040, 2005.
Baker NA, Sept D, Joseph S, Holst MJ, McCammon JA. Electrostatics of nanosystems: application to microtubules and the ribosome. Proc. Natl. Acad. Sci. USA 98, 10037-10041 2001.
I. Borukhov, D. Andelman, and H. Orland, Physical Review Letters 79 (3), 435-438 (1997).
84
october • fifteen • 2005
Three-dimensional correlation poster
analysis of the macromolecules
on synaptic vesicles
Purpose
Jing Xu
We studied the protein associations of the 50-nm-large synaptic vesicles (SVs)
55
in a model system, frog neuromuscular junctions (NMJ) from the structural
perspective, aiming to provide insights into the clustering, mobilization,
docking and fusion of these smallest functional organelles that carry neurotransmitters.
Mark L. Harlow
David Ress
David Yip
Materials and Methods
Sample preparation: Frog neuromuscular junctions were prepared with the
Mira Raman
conventional procedures for TEM.
Electron Microscope Tomography (EMT): This is a technology that gener- Robert M. Marshall
ates 3D volume data from stained biological tissue sections, enables studies
Uel J. McMahan
of the structures and localizations of cell organelles or protein aggregates. A
software package called EM3D was used to reconstruct and analyze the volume data.
Correlation analysis and probability map generation: Contact points from any two
vesicles were aligned based on a cross-correlation function given three dimensions of
rotational freedom as rigid body. And a large number of vesicles were aligned until the best mutual correlation was reached, the algorithm of which was adapted from [2].
Results
Each SV in the frog NMJ terminal is linked to other SVs by filamentous structures that are 3-30 nm long and
4-10 nm in diameter, and each SV contains scaffolding luminal filaments. The pattern of the position of contacts of these filaments to the vesicles was found by generating a probability map of the contacts from a large
number of vesicles. The high degree of clustering of the probability map indicates a constancy of the distribution. Most of the contacts are confined to five domains. The two maps found for the contacts of filaments from
outside and that from inside the vesicle are strikingly similar, which provides strong evidence that both types
of filaments contain the tails of proteins in the membrane and such membrane proteins are stereotypically organized. Furthermore, this distribution of vesicle proteins was used to determine the “orientations” of the SVs
within the nerve terminal, and it was found that the portion of membrane in the vesicles that are going to be
docked at the cell membrane are becoming oriented facing the membrane as the vesicles are within 125nm
distance from the membrane.
Conclusions
The SVs have similar macromolecular assembly and the SVs must be oriented in a specific way in order to be
docked at the active zone where there are fused with the cell membrane and release the neurotransmitters.
Reference
[1] Ress et al, Structure, 12 (10):1763-1774, 2004
[2] Penczek P et al, Ultramicroscopy, 40: 33-53, 1992
[3] Harlow et al, Nature 409:479-484, 2001
85
biomedical • computation • at • stanford
Notes
86
october • fifteen • 2005
Notes
87
biomedical • computation • at • stanford
Notes
88
october • fifteen • 2005
BioNe us
A grassroots approach to Bio- X
BioNexus is a newly-formed Stanford
student/post-doc organization seeking to promot e
and facilitate a community of interdisciplinary
research and recreation within Bio-X.
We are planning social and academic activities to
bring Bio-X together , from the bottom up .
Join the Nexus!
http://bione us.stanford.edu
051165
9/29/04
10:03 AM
Page 1
Weekend
Road Trip Specials.
Giving you a terrific package is what Enterprise does best - especially on the weekends.
We even have great rates designed to fit your college budget. So wherever your weekend
plans take you, call Enterprise for great cars, low rates and free pick-up!
*As with all our renters, those in the 18/20 age group must meet certain qualifications to rent. Ask for details.
Pick-up is subject to geographic and other restrictions.
Just dial (650) 833-8060.
Size: 6” X 5”
BW
non-bleed
Output at 100%
89
biomedical • computation • at • stanford
Combining Experience, Connections, and Capital
to Create Winning Companies in both
Information and Life Sciences
acquired by ArQule
Capital commitments since 1982 in excess of $1 Billion
90
General Partners
Venture Partners
Tony Di Bona
Ammar H. Hanafi
Michael W. Hunkapiller
Douglas E. Kelly, M.D.
J. Leighton Read, M.D.
Daniel I. Rubin
John F. Shoch
Craig C. Taylor
David J. Ezequelle
Peter N. Loukianoff
David W. Pidwell
W. Ferrell Sanders
www.alloyventures.com
october • fifteen • 2005
Why Choose a Mac?
Count the ways you’ll love your new Apple Computer.
Simply take a Mac out of its box, plug it in and start surfing, working or
rocking. And when you want to add something new – such as a digital
camera or iPod – you just plug it in and start using it. Too good to be
true? No, the Mac is exactly that good.
It just works.
As easy as iPod.
Want a crash course on the
Mac? Lesson 1: It doesn’t
crash. Built by the same folks
who designed its rock-solid operating
system and its award-winning
applications, the Mac always starts up
and never lets you down.
Your iPod has a history. And it
starts with the Mac. Intuitive
interface, elegant design,
seamless integration with Apple
software and a hearty dose of just
plain cool. If you love your iPod, you’ll
love a Mac.
Picture-perfect photos.
It’s a musical instrument.
If every picture tells a story,
you’ll live happily ever after
with your new Mac. Plug in your
our
camera and the Mac does the rest,
helping you import, edit, share and
organize your photos faster than you
can say “cheese.”
More than a place to turn up
iTunes, the Mac lets you make
music, too. Whether you’re
recording your next big hit, burning
a mix CD or podcasting your
passion, you’ll be doing it all on the
same computer countless pros trust.
Home movies in HD.
Online streamlined.
With a Mac, you get an editing
suite, a special effects house
and a state-of-the-art cinema
all in one package. From recording
your HD masterpiece to burning it to
DVD, a Mac grants you auteur status.
With a Mac in your corner,
you can get online in
minutes. And once you’re
there you might as well try all the built in
goodies — Mail, Safari, .Mac, iChat AV, AirPort
and more — the Mac has to offer.
Join the party.
It loves roadtrips.
It does Windows.
It’s beautiful.
Buy a Mac and you become
part of the solution. The Apple
experience includes top-rated
service online, in our retail stores and
via telephone. Plus, you join a
worldwide network of Mac users,
happy to share their stories.
Your Mac may soon seem likee
your digital oasis, but that
doesn’t make you an island.
You’ll be able to connect to
o the
office network, share Office files with
colleagues, and even use many of your
PC hardware devices.
Mobility means more on
a Mac. Besides powering
the two most appealing
notebooks on the planet,
iBook and PowerBook, Mac OS X
Tiger gives you easy connectivity
from anywhere you dare to roam.
To learn what separates
an Apple computer from
every other machine on
the market, just open
your eyes. The Mac’s
design fold function into form, so
the perfect computer looks the part.
Believe your ears.
Welcome the
impossibly small
iPod nano
to the family.
apple.com/ipodnano
Got Questions?
Got questions about your purchase or
anything Apple related? Just drop an
e-mail to the friendly Stanford University
Apple Campus Reps at
[email protected],
[email protected], or
[email protected] and they will be
sure to help you out!
Where can I purchase?
Don’t forget to use your education discount!
Stanford Bookstore
White Plaza
519 Lasuen Mall
(650) 725-6136 ext 456
www.stanfordbookstore.com
Want to read more?
http://www.apple.com/switch
91
bcats2005
thankyou
Administrative Help
and Guidance
Previous Organizers
Zhen LIn
Zachary Pincus
Brian Naughton
Jessica Tenenbaum
David Buckeridge
Symposium Sponsors
Founding Sponsors
Stanford University Center for
Carol Maxwell (Stanford)
Bio-Medical Computation
Ted Hopper (Stanford)
NIH – NIBIB
Tiffany Murray (Stanford)
Bio-X
Rosalind Ravasio (Stanford)
Fiona Sincock (Stanford)
Gold Sponsors
Heideh Fattaey (Stanford)
Apple
Symposium Volunteers
Jill Sakata (Stanford)
Genentech (William Wood)
Jonathan Dugan (Stanford)
Alloy Ventures (Leighton Reed)
Timothy McBride
Yue Lu (Stanford)
Yael Garten
Xinlei Qiu (Stanford)
David Gobaud
Silver Sponsors
Vincent Foecke (Stanford)
Greg Goldgorf
Roche (John Shon)
Marcy Rivera (Roche)
Yang Huang
Agilent
Collyne Rusin (Roche)
Nikesh Kotecha
Life Sciences Society
Annette Adler
Christina Lu
Wilson Sonsini Goodrich
(Agilent Technologies)
Cesar Rodriguez
& Rosati
Jane Weber (Alloy Ventures)
Kahye Song
(Benjamin Glenn)
Silvia Herrero (Apple)
Valentin Spitkovsky
Justin Chang (Apple)
Jessica Tenenbaum
Vicky Markstein (Life Sciences
Li Zhang
Other Sponsors
Society)
Enterprise
Stefan Zappe (Life Sciences
Stanford Bookstore
Society)
Stanford OTL
Julie Rubinstein (Enterprise)
Jennifer Kesler (Genentech)
John Brown (WSGR)
biomedicalcomputationatstanford