Title Slide First Line Title Slide Second Line
Download
Report
Transcript Title Slide First Line Title Slide Second Line
Cyberinfrastructure and Internet2
Eric Boyd
Deputy Technology Officer
Internet2
What is Cyberinfrastructure (CI)?
• A strategic orientation supported by NSF
• Calls for large-scale public investment to
encourage the evolution of widely distributed
computing via the telecommunications
network
• Goal is to deploy the combined capacity of
multiple sites to support the advance of
current research, initially in science and
engineering
The Distributed CI Computer
Instrumentation
Control
Help
Desk
Researcher
Security
Viewing
Data
Generation
Program
Security
Collab
Tools
Control
Training
Education
And
Outreach
Publishing
Policy and
Funding
Management
Security and
Access
Access
Control
Human
Support
Funding
Agencies
Authentication
Resource
Providers
Campuses
3D
Imaging
Display
Tools
Authorization
Security
Security
Analysis
Simulation
Computation
Program
Input
Archive
Retrieval
Data Sets
Storage
Search
Data
Directories
Schema
Metadata
Ontologies
Data
.
Input
Network
Security
Display and
Visualization
The Network is the Backplane for the
Distributed CI Computer
Instrumentation
Control
Help
Desk
Researcher
Security
Viewing
Data
Generation
Program
Security
Collab
Tools
Control
Training
Education
And
Outreach
Publishing
Policy and
Funding
Management
Security and
Access
Access
Control
Human
Support
Funding
Agencies
Authentication
Resource
Providers
Campuses
3D
Imaging
Display
Tools
Authorization
Security
Security
Analysis
Simulation
Computation
Program
Input
Archive
Retrieval
Data Sets
Storage
Search
Data
Directories
Schema
Metadata
Ontologies
Data
.
Input
Network
Security
Display and
Visualization
Challenge and Opportunity
• Challenge:
• The R&E community thinks of CI primarily
in terms of building distributed computing
clusters
• Opportunity:
• The network is a key component of CI
• Internet2 is leading the development of
solutions for the network component of CI
LHC epitomizes the CI Challenge
Current Situation
• Large Hadron Collider (LHC) at CERN
will go operational in 2008
• Over 68 U.S. Universities and National
Laboratories are poised to receive data
• More than 1500 scientists are waiting
for this data
• Are campus, regional, and national
networks ready for the task?
7
Local Infrastructure
US Tier 4
(1500 US
scientists)
Scientists
Analyze Data
US Tier 3
(68 orgs)
Internet2/Connectors
Internet2/Connectors
Scientists
Request Data
Atlas (6-7)
CMS (7)
US Tier 2
(15 orgs)
Provides
Data to
Tier 3
GEANT-ESNet-Internet2
Tier 1
(12 orgs)
FNAL
BNL
Shared Data Storage
and Reduction
LHCOPN
Tier 0
8
CERN
Raw Data
Peak Flow Network Requirements
Local Infrastructure
Tier 1 or 2 to Tier 3: Estimate: Requires 1.6 Gbps per transfer (2
TB's in 3 hours)
Internet2/Connectors
Internet2/Connectors
Tier 1 to Tier 2: Requires 10-20 Gbps
GEANT-ESNet-Internet2
Tier 0 to Tier1: Requires 10-40 Gbps
LHCOPN
CERN
9
Science Network Requirements Aggregation Summary
(slide courtesy of ESNet)
Science
Drivers
End2End
Reliability
Connectivity
2006
End2End
Band
width
2010
End2End
Band
width
1 TB/day
5 TB/day
300 Mbps
1.5 Gbps
• DOE sites
• US Universities
625 Mbps
250 Gbps
Science Areas
/ Facilities
Advanced
Light Source
-
Bioinformatics
-
• DOE sites
• US Universities
• Industry
12.5
Gbps in
two years
Traffic
Characteristics
Network Services
• Bulk data
• Remote control
• Guaranteed bandwidth
• PKI / Grid
• Bulk data
• Remote control
• Point-to-multipoint
• Guaranteed bandwidth
• High-speed multicast
Chemistry /
Combustion
-
• DOE sites
• US Universities
• Industry
-
10s of
Gigabits
per
second
• Bulk data
• Guaranteed bandwidth
• PKI / Grid
Climate
Science
-
• DOE sites
• US Universities
• International
-
5 PB per
year
• Bulk data
• Remote control
• Guaranteed bandwidth
• PKI / Grid
• Bulk data
• Remote control
• Guaranteed bandwidth
• Traffic isolation
• PKI / Grid
High Energy
Physics (LHC)
99.95+%
(Less than
4 hrs/year)
• US Tier1 (DOE)
• US Tier2 (Universities)
• International (Europe,
Canada)
5 Gbps
10 Gbps
60 to 80
Gbps
(30-40
Gbps per
US Tier1)
Science Network Requirements Aggregation Summary
(slide courtesy of ESNet)
Science
Drivers
End2End
Reliability
Connectivity
2006
End2End
Band
width
2010
End2End
Band
width
• DOE sites
• US Universities
• Industry
200+
Mbps
1 Gbps
• Bulk data
• Remote control
• Guaranteed bandwidth
• Guaranteed QoS
• Deadline scheduling
• DOE sites
• US Universities
• Industry
• International
• DOE sites
• US Universities
• Industry
• International
• DOE sites
• US Universities
• International
10 Gbps
20 to 40
Gbps
• Bulk data
• Remote control
• Guaranteed bandwidth
• Guaranteed QoS
• Deadline Scheduling
• PKI / Grid
Backbone
Band
width
parity
Backbone
band
width
parity
• Bulk data
12 Gbps
70 Gbps
• Bulk data
• DOE sites
640 Mbps
2 Gbps
• Bulk data
Science Areas
/ Facilities
Magnetic
Fusion Energy
99.999%
(Impossible
without full
redundancy)
NERSC
-
NLCF
-
Nuclear
Physics
(RHIC)
-
Spallation
Neutron
Source
High
(24x7
operation)
Traffic
Characteristics
Network Services
• Guaranteed bandwidth
• PKI / Grid
CI Components
• Supercomputing / Cycles /
Computational
• Supercomputing / Storage (Non-volatile)
• Interconnecting Networks (Campuses,
Regionals, Backbones)
• Cyberinfrastructure Software
• Analysis / Visualization
CI Components
Applications
Bulk
Transport
2-Way
Interactive
Video
Real-Time
Communications
….
….
….
Performance
Infrastructure / Tools
Middleware
Phoebus
Measurement
Nodes
Network
….
Control
Plane
Control Plane
Nodes
Network
Cyberinfrastructure
Applications call on Network Cyberinfrastructure
Internet2 Network CI Software
• Dynamic Circuit Control Infrastructure
• DRAGON (with ISI, MAX)
• Oscars (with ESnet)
• Middleware (Federated trust Infrastructure)
•
•
•
•
Shibboleth
Signet
Grouper
Comanage
• Performance Monitoring Infrastructure
• perfSONAR (with ESnet, GEANT2 JRA1, RNP, many others)
• BWCTL, NDT, OWAMP, Thrulay
• Distributed System Infrastructure
• Topology Service (with University of Delaware)
• Distributed Lookup Service (with University of Delaware, PSNC)
Internet2 Network CI Standardization
• Dynamic Circuit Control Protocol (IDC)
• DICE-Control, GLIF
• Measurement Schema / Protocol
• OGF NMWG
• IETF IPPM
• perfSONAR Consortium
• Middleware Arena
• Liberty Alliance
• OASIS
• Possible emerging corporate consortium
• Topology Schema / Protocol
• OGF NML-WG
• perfSONAR Consortium
• DICE-Control
Internet2’s CI Vision
• Internet2’s CI vision:
• Be a networking cyber-service provider
• Be a trust cyber-service provider
• Be a CI technology developer.
Internet2’s CI Position
• Internet2’s position:
• Backbone network provider
• Federated trust infrastructure provider
• Forum for collaboration by members of the
R&E community
• Gives Internet2 a unique vision and
strategy for Cyberinfrastructure.
Internet2’s CI Constituencies
•
•
•
•
•
•
•
•
Collaborators
University Members
Regional Networks
Regional CI Organizations
High Performance Computing Centers
Federal Partners
International Partners
CI Integrators
Early Thoughts: Internet2’s CI Strategy (1)
• Requirements
• Informed by our membership
• Agenda set by our governance mechanisms
• Offer, and in some cases develop, services and
technology that are key components of a coherent CI
software suite.
• CI-enhanced Networks: IP Network, Dynamic Circuit
Network
• Services: InCommon, USHER
• New Technologies: DCN software, perfSONAR, Shibboleth
• Systems Integration: Assemble open source communication
tools into a common veneer.
• Emphasize a systems approach towards CI.
Early Thoughts: Internet2’s CI Strategy (2)
• Take a “toolkit” approach
• Make sure it still looks like a wall jack to end user
• Push for best practices for campuses
• What to do
• How to do it
• Community learns as a whole / avoid reinventing the wheel
• Contribute to the support structure for use of CI
• Open source CI software
• Centers of Excellence
• Training
Early Thoughts: Internet2’s CI Strategy (3)
• Play the role of community CI coordinator,
convening community conversations.
• Partner with other community coordinators (e.g.
Teragrid, EDUCAUSE).
• Play a convening function in order to facilitate the
development, use, and dissemination of CI
• Take a lead in international outreach efforts
• Facilitate conversations among various federal
agencies (e.g. DOE, NSF, NIH), each of which is
developing its own CI