Transcript Standard

Extending the Internet
Throughout the Physical World
Keynote to
The EC-US Taskforce on
Biotechnology Research
Arlington, VA
Sept 9, 2001
Larry Smarr
Department of Computer Science and Engineering
Jacobs School of Engineering, UCSD
Director, California Institute for Telecommunications and Information Technology
Towards a
Global Biological Knowledge Grid
• Capture and Integrate Multiple Scales of Science
–
–
–
–
–
–
Genomes, Proteins, Metabolic Pathways
Cellular Systems
Organism Models
Ecological Systems
Geographic Biodiversity
Environmental Interactions
• Adapting to the Emerging Information Infrastructure
–
–
–
–
–
Wireless Access--Anywhere, Anytime
Distributed Sensors, Data, People, Computers
From the Web to the Grid
Highly Parallel Light Waves Through Fiber
Emergence of a Distributed Planetary Computer
Dynamic Growth in Mobile Internet
Forecast of Internet users worldwide
3G Adds Mobility, QoS, and High Speeds
Subscribers (millions)
2,000
1,800
1,600
1,400
1,200
1,000
Mobile Internet
800
600
400
Fixed Internet
200
0
1999
2000
2001
2002
Source: Ericsson
2003
2004
2005
California Has Undertaken a
Grand Experiment in Partnering
The California Institute for Bioengineering,
Biotechnology,
and Quantitative Biomedical Research
The Center for
Information Technology Research
in the Interest of Society
UCD
UCB
UCSF UCM
UCSC
The California
NanoSystems Institute
UCSB
UCLA
UCI
UCSD
The California Institute
for Telecommunications
and Information Technology
The UCSD “Living Grid Laboratory”—
Fiber, Wireless, Compute, Data, Software
•Commodity Internet, Internet2
•CENIC’s ONI, Cal-REN2, Dig. Cal.
•PACI Distributed Terascale Facility
Wireless WAN
SDSC
• High-speed optical core
CS
Eng. / Cal-(IT)2
Hosp
Med
Chem
• Wireless LANs
SIO
½ Mile
Source: Phil Papadopoulos, SDSC
The High Performance
Wireless Research and Education Network
• Cal-(IT)2 Will Build
on This Pioneering
Experiment
• Add New Ecological
Sensor Arrays
• Try Out New
Wireless
Technologies
• Data Analysis
• Outreach and
Education
NSF Funded
PI, Hans-Werner Braun, SDSC,
UCSD
Co-PI, Frank Vernon, SIO, UCSD
45mbps Duplex Backbone
As Our Bodies Move On-Line
Bioengineering and Bioinformatics Merge
FDA
Approved
Aug. 2001
• New Sensors—Israeli Video Pill
– Battery, Light, & Video Camera
– Images Stored on Hip Device
• Next Step—Putting You On-Line!
– Wireless Internet Transmission
– Key Metabolic and Physical Variables
– Model -- Dozens of 25 Processors and 60
Sensors / Actuators Inside of our Cars
• Post-Genomic Individualized Medicine
– Combine
– Genetic Code
– Body Data Flow
– Use Powerful AI Data Mining Techniques
www.givenimaging.com
Adding Brilliance to Wireless Sensors
With Systems-on-Chip
Applications
Critical New Role of
Power Aware Systems
Sensors Embedded
Software
Processors
Memory
Protocol
Processors
Radio
DSP
Ad Hoc Hierarchical Network
of Brilliant Sensors
Source: Sujit Dey, UCSD ECE
Internet
Moore’s Law—
Simple 2D Shrinking Reaches End by 2015
Feature size
(nanometers)
Intel8080
1000nm
Intel386
Intel486
Pentium
PentiumPro
PentiumIII
IA-64
100nm
DARPA
10nm
1nm
1970
1980
Bipolar, NMOS
1990
2000
CMOS
Nanosciences
Molecular Electronics/
Quantum / Bio
3-D CMOS + - HYBRIDS
2010
2020
2030
?
Source: Shankar Sastry, DARPA ITO
2040
2050
The Perfect Storm:
Convergence of Engineering with Bio, Physics, & IT
500x
Magnification
Nanogen
MicroArray
2 mm
VCSELaser
400x
Magnification
IBM Quantum Corral
Iron Atoms on Copper
Human Rhinovirus
5 nanometers
Nanobioinfotechnology
Why the Grid is the Future
Scientific American, January 2001
Layered Software Approach to
Building the Planetary Grid
Science Portals & Workbenches
Twenty-First Century Applications
Access
Grid
Computational
Grid
Access Services &
Technology
Computational Services
Grid Services
(resource independent)
Grid Fabric
(resource dependent)
“A source book for the history
of the future” -- Vint Cerf
Networking, Devices and Systems
Edited by Ian Foster and Carl Kesselman
www.mkp.com/grids
P
e
r
f
o
r
m
a
n
c
e
The Grid Physics Network Is Driving the
Creation of an International Grid
•
Paul Avery (Univ. of Florida) and Ian Foster (U. Chicago and ANL), Lead PIs
– Largest NSF Information Technology Research Grant
– 20 Institutions Involved
– Built on Globus Middleware
LHC
Sloan Digital
Sky Survey
CMS
ATLAS
The EUROGRID
Creates an EU Virtual Machine Room
• UNICORE
– Java
Middleware
• Driven by
Applications
• Links to Key
Databases
• One Interface
to Multiple
Machines
STAR TAP:
Science Technology And Research Transit Access Point
Canada
Japan
Korea (2)
Taiwan
Singapore (2)
Australia (2)
China
Chile, Brazil ANSP, Brazil RNP, Mexico
US: Abilene, DREN, ESnet, NISN, NREN, vBNS/vBNS+
Norway
Iceland
Sweden
Finland
Denmark
Russia
France
Netherlands
CERN
Israel
Ireland
Belgium
Europe/DANTE
United Kingdom
www.startap.net
http://www.ucaid.edu/abilene/html/itnpeerparticipants.html (Abilene ITN)
http://www.canet3.net/optical/peering_info/intl_peering.html (CA*net3 ITN)
Star Light
International Wavelength Switching Hub
AsiaPacific
SURFnet,
CERN
CANARIE
Seattle
Portland
NYC
AsiaPacific
TeraGrid
Caltech
SDSC
*ANL, UIC, NU,
UC, IIT, MREN
Source: Tom DeFanti, Maxine Brown
AMPATH
AMPATH
The NSF TeraGrid
Partnerships for Advanced Computational Infrastructure
This will Become the National Backbone to Support Multiple
Large Scale Science and Engineering Projects
Applications
Visualization
Caltech
Argonne
0.5 TF
0.4 TB Memory
86 TB disk
1 TF
0.25 TB Memory
25 TB disk
TeraGrid Backbone (40 Gbps)
Data
SDSC
4.1 TF
2 TB Memory
250 TB disk
Compute
NCSA
8 TF
4 TB Memory
240 TB disk
Advancing Realism in
Modeling Cell Structures
• Pre-Blue Horizon (mid-1990s):
– Model Electrostatic Forces of a
Structure up to 50,000 Atoms
– a Single Protein or Small Assembly
• Pre-TeraGrid (2001):
– Model One Million Atoms
– Simulate Drawing a Drug Molecule
Through a Microtubule or Tugging
RNA Into a Ribosome
• TeraGrid (2003):
– Models of 10 Million Atoms
– Model Function, Structure
Movement, and Interaction at the
Cellular Level for Drug Design and to
Understand Disease
Baker, N., Sept, D., Joseph, S., Holst,
M., and McCammon, J. A. PNAS 98:
10037-10040 (2001)
Source: Fran Berman
Prototyping the Grid Cyber-Infrastructure
for a Biomedical Imaging Research Network
Forming a National-Scale Grid
Federating Multi-Scale Neuro-Imaging
Data from Centers with High Field
MRI and Advanced 3D Microscopes
Deep Web
Harvard
NCRR Imaging
Source: Mark Ellisman, UCSD
BIRN
and Computing
Resources UCSD
Cal Tech
SDSC
UCLA
UCSD
Surface Web
Cal-(IT)2
Duke
Wireless “Pad”
Web Interface
Part of the UCSD CRBS
Center for Research on Biological Structure
From Telephone Conference Calls to
Access Grid International Video Meetings
Creating a Virtual Global Research Lab
Access Grid Lead-Argonne
NSF STARTAP Lead-UIC’s Elec. Vis. Lab
Vast Data Sets Will Require
High Resolution Data Analysis Facilities
Celera
Control Room
SDSC
SIO
Cal-(IT)2
Control Room
Cox Communications
Teraburst Networks
Panoram Technologies
Newsday Photo
Ira Schwarz
Grid-Enabled Collaborative Analysis
of Ecosystem Dynamics Datasets
Chesapeake Bay Data in Collaborative Virtual Environment
Common Portal Architecture
Customized for Biological Sciences
Web Browser - Portal Interface
State Values
Portal Engine
Data Gather
XML
HTML
Legacy and Problem Specific
Databases, Collections, & Literature
User Preferences
Analysis Tools
- Genome, Protein, &
Metabolic Pathways
- Cellular Models
- Integrative Systems
- Species Identification
- GIS Biodiversity
- Data Mining
- ...
A Global IT Strategy Is Needed
to Integrate the Emerging Plant Genomes
Immense Computing Power Will Be Required
to Lead in Post-Genomic Research
• ”We Don’t Need an Evolution in Computing,
We Need a Revolution”—Craig Venter
• Sandia and Celera Will Collaborate On:
– Advanced Algorithms
– Visualization Technologies for Analyzing
Massive Quantities of Experimental Data From
High-Throughput Instruments
• Equivalent to 100,000 Pentium 4’s!
• Prototype by 2004
Biology is at the Leading Edge of Using the
Emerging Planetary Computer
Application Software Has Been Downloaded to Over 30,000 PCs
Over 500 CPU-Years Computed
Total Storage 50 Terabytes, Peak Speed 13 Teraflops
In Silico
Drug Design
Art Olson,
The Scripps
Research
Institute
A Planetary MegaComputer—
Distributed Computing & Mass Storage
• Napster Meets Seti@Home !
• Assume Ten Million PCs in Five Years
– Average Speed Ten GigaFLOP
– Average Free Storage 100 GB
• Planetary Computer Capacity
– 100,000 TetaFLOP Speed
– 1 Million Terabyte Storage
• Global Distributed Server for Mobile Clients
Will a New Form of Intelligence
Join Human Kind?
1 Million x
•Will the Grid Become Self–Organizing
–Powered
–Aware?
Source: Hans Moravec
www.transhumanist.com/volume1/power_075.jpg