Distributed Data Access and Analysis
Download
Report
Transcript Distributed Data Access and Analysis
Distributed Data Access and Analysis
for Next Generation HENP Experiments
Harvey Newman, Caltech
CHEP 2000, Padova
February 10, 2000
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
LHC Computing: Different from
Previous Experiment Generations
Geographical dispersion: of people and resources
Complexity: the detector and the LHC environment
Scale: Petabytes per year of data
~5000 Physicists
250 Institutes
~50 Countries
Major challenges associated with:
Coordinated Use of Distributed Computing Resources
Remote software development and physics analysis
Communication and collaboration at a distance
R&D: A New Form of Distributed System: Data-Grid
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Four Experiments
The Petabyte to Exabyte Challenge
ATLAS, CMS, ALICE, LHCB
Higgs and New particles; Quark-Gluon Plasma; CP Violation
Data written to “tape”
0.1 to
(~2010)
1
(~2020 ?)
~5 Petabytes/Year and UP
(1 PB = 1015 Bytes)
Exabyte (1 EB = 1018 Bytes)
Total for the LHC Experiments
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
To Solve: the LHC “Data Problem”
While the proposed LHC computing and data handling
facilities are large by present-day standards,
They will not support FREE access, transport or processing
for more than a minute part of the data
Balance between proximity to large computational and data
handling facilities, and proximity to end users and more
local resources for frequently-accessed datasets
Strategies must be studied and prototyped, to ensure both:
acceptable turnaround times, and efficient resource utilisation
Problems to be Explored
How to meet demands of hundreds of users who need transparent
access to local and remote data, in disk caches and tape
stores
Prioritise hundreds of requests of local and remote communities,
consistent with local and regional policies
Ensure that the system is dimensioned/used/managed
optimally,
for theData
mixed
workload
February
10, 2000: Distributed
Access and
Analysis for HENP Experiments
Harvey B Newman (CIT)
MONARC General Conclusions
on LHC Computing
Following discussions of computing and network requirements,
technology evolution and projected costs, support requirements etc.
The scale of LHC “Computing” requires a worldwide effort to
accumulate the necessary technical and financial resources
A distributed hierarchy of computing centres will lead to better use
of the financial and manpower resources of CERN, the Collaborations,
and the nations involved, than a highly centralized model focused at
CERN
The distributed model also provides better use of
physics opportunities at the LHC by physicists and students
At the top of the hierarchy is the CERN Center, with the ability to perform
all analysis-related functions, but not the ability to do them completely
At the next step in the hierarchy is a collection of large, multi-service
“Tier1 Regional Centres”, each with
10-20% of the CERN capacity devoted to one experiment
There will be Tier2 or smaller special purpose centers in many regions
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Bandwidth Requirements Estimate (Mbps) [*]
ICFA Network Tas
Year
1998
2000
2005
0.050.25
(0.5 - 2)
0.2 - 2
(2 - 10)
0.8 - 10
(10 - 100)
BW Utilized by a University
Group
0.25 - 10
1.5 - 45
34 - 622
BW to a Home-laboratory
or Regional Center
1.5 - 45
34 155
622 5000
BW to a Central Laboratory
Housing One or More Major
Experiments
34 - 155
155 622
2500 10000
BW on a Transoceanic Link
1.5 - 20
34-155
622 5000
BW
Utilized Per Physicist
(and Peak BW Used)
See http://l3www.cern.ch/~newman/icfareq98.html
Circa 2000: Predictions roughly on track:
“Universal” BW Growth” by ~2X Per Year;
622 Mbps on Links European and Transatlantic by ~2002-3
Terabit/sec US Backbones (e.g. ESNet) by ~2003-5
Caveats: Distinguish raw bandwidth and effective line capacity;
Maximum end-to-end rate for individual data flows
“QoS”/ IP has a way to go
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
D388,
D402,
D274
CMS Analysis and
Persistent Object Store
Online
CMS
Offline
L1
Slow Control
Detector
Monitoring
L2/L3
“L4”
Filtering
Persistent Object Store
Object Database Management System
Simulation
C121
Calibrations, Group
Analyses
Common Filters and
Pre-Emptive Object
Creation
User Analysis
On Demand
Object Creation
Data Organized In a(n
Object) “Hierarchy”
Raw, Reconstructed (ESD),
Analysis Objects (AOD),
Tags
Data Distribution
All raw, reconstructed
and master parameter DB’s
at CERN
All event TAG and AODs,
and selected reconstructed
data sets
at each regional center
HOT data (frequently
accessed) moved to RCs
Goal of location and medium
transparency
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
GIOD Summary
GIOD has
C51
Constructed a Terabyte-scale
Hit
Track
Detector
C226
set of fully simulated CMS
events and used these to
create a large OO database
Learned how to create large
database federations
Completed the “100” (to 170)
Mbyte/sec CMS Milestone
Developed prototype
reconstruction and analysis
codes, and Java 3D OO
visualization demonstrators,
that work seamlessly with
persistent objects over
networks
Deployed facilities and
database federations as
useful testbeds for
Computing Model studies
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Data Grid Hierarchy (CMS Example)
1 TIPS = 25,000 SpecInt95
~PBytes/sec
Online System
Bunch crossing per 25 nsecs.
100 triggers per second
Event is ~1 MByte in size
Tier 1
France Regional
Center
~100 MBytes/sec
PC (today) = 10-15 SpecInt95
Offline Farm
~20 TIPS
~100 MBytes/sec
~622 Mbits/sec
or Air Freight
Tier 0
Germany Regional
Center
CERN Computer
Center
Italy Regional
Center
Fermilab
~4 TIPS
~2.4 Gbits/sec
Tier 2
~622 Mbits/sec
Tier 3
Institute
Institute Institute
~0.25TIPS
Physics data cache
E277
Tier2 Center
Tier2 Center
Tier2 Center
Tier2 Center Tier2 Center
~1 TIPS ~1 TIPS ~1 TIPS ~1 TIPS
~1 TIPS
Workstations
Institute
100 - 1000
Mbits/sec
Tier 4
Physicists work on analysis “channels”.
Each institute has ~10 physicists working
on one or more channels
Data for these channels should be
cached by the institute server
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
LHC (and HEP) Challenges
of Petabyte-Scale Data
Technical Requirements
Optimize use of resources with next generation middleware
E163 Co-Locate and Co-Schedule Resources and Requests
C74,
C292
Enhance database systems to work seamlessly
across networks: caching/replication/mirroring
Balance proximity to centralized facilities, and
to end users for frequently accessed data
Requirements of the Worldwide
Collaborative Nature of Experiments
Make appropriate use of data analysis resources in
each world region, conforming to local and regional policies
Involve scientists and students in each world region
in front-line physics research
Through an integrated collaborative environment
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Time-Scale: CMS
Recent “Events”
A108
A PHASE TRANSITION in our understanding of the role of CMS
Software and Computing occurred in October - November 1999
“Strong Coupling” of S&C Task,Trigger/DAQ, Physics TDR,
detector performance studies and other main milestones
Integrated CMS Software and Trigger/DAQ planning for the next
round:
May 2000 Milestone
Large simulated samples are required: ~ 1 Million events fully
simulated a few times during 2000, in ~1 month
A smoothly rising curve of computing and data handling needs
from now on
Mock Data Challenges from 2000 (1% scale) to 2005
Users want substantial parts of the functionality formerly
planned for 2005, Starting Now
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Roles of Projects
for HENP Distributed Analysis
RD45, GIOD: Networked Object Databases
Clipper,GC;
High speed access to Objects or File
data
FNAL/SAM
for processing and analysis
SLAC/OOFS Distributed File System + Objectivity
Interface
NILE, Condor:
Fault Tolerant Distributed Computing
with
Heterogeneous CPU Resources
MONARC: LHC Computing Models:
Architecture, Simulation, Strategy, Politics A391
PPDG:
First Distributed Data Services and
Data Grid System Prototype
ALDAP:
Database Structures and AccessE277
Methods for Astrophysics and HENP Data
GriPhyN:
Production-Scale Data Grid
Simulation/Modeling, Application + Network
Instrumentation, System Optimization/Evaluation
February 10, 2000:
APOGEE
Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
MONARC: Common Project
Models Of Networked Analysis
At Regional Centers
Caltech, CERN, Columbia, FNAL, Heidelberg,
Helsinki, INFN, IN2P3, KEK, Marseilles, MPI
Munich, Orsay, Oxford, Tufts
PROJECT GOALS
Develop “Baseline Models”
Specify the main parameters
characterizing the Model’s
performance: throughputs, latencies
Verify resource requirement baselines:
(computing, data handling, networks)
Univ
1
Univ
2
Tier2 Ctr
20k SI95
20 TB
Disk
Robot
Univ
FNAL/BNL
70k SI95
70 Tbyte
Disk; Robot
TECHNICAL GOALS
M
Define the Analysis Process
Define RC Architectures and Services
Provide Guidelines for the final Models
Provide a Simulation Toolset for Further Model Circa
Model studies
2005
F148
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
CERN
350k SI95
350 Tbytes
Disk;
Robot
Harvey B Newman (CIT)
MONARC Working Groups/Chairs
“Analysis Process Design”
P. Capiluppi (Bologna, CMS)
“Architectures”
Joel Butler (FNAL, CMS)
“Simulation”
Krzysztof Sliwa (Tufts, ATLAS)
“Testbeds”
Lamberto Luminari (Rome, ATLAS)
“Steering”
Laura Perini (Milan, ATLAS)
Harvey Newman (Caltech, CMS)
& “Regional Centres Committee”
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
MONARC Architectures WG:
Regional Centre Facilities & Services
Regional Centres Should Provide
All technical and data services required to do physics analysis
All Physics Objects, Tags and Calibration data
Significant fraction of raw data
Caching or mirroring calibration constants
Excellent network connectivity to CERN and the region’s users
Manpower to share in the development of common validation
and production software
A fair share of post- and re-reconstruction processing
Manpower to share in ongoing work on Common R&D Projects
Excellent support services for training, documentation,
troubleshooting at the Centre or remote sites served by it
Service to members of other regions
Long Term Commitment for staffing, hardware evolution and support
for R&D, as part of the distributed data analysis architecture
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
MONARC and Regional Centres
MONARC RC FORUM: Representative Meetings Quarterly
Regional Centre Planning well-advanced, with optimistic outlook,
in US (FNAL for CMS; BNL for ATLAS), France (CCIN2P3), Italy, UK
Proposals submitted late 1999 or early 2000
Active R&D and prototyping underway, especially in US, Italy,
Japan; and UK (LHCb), Russia (MSU, ITEP), Finland (HIP)
Discussions in the national communities also underway in
Japan, Finland, Russia, Germany
There is a near-term need to understand the level and sharing of
support for LHC computing between CERN and the outside
institutes, to enable the planning in several countries to advance.
MONARC Uses traditional 1/3:2/3 sharing assumption
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Regional Center Architecture
Example by I. Gaines (MONARC)
Tape Mass Storage
& Disk Servers
Database Servers
Network
from
CERN
Network
from Tier 2
& simulation
centers
Tapes
Local
institutes
Production
Reconstruction
Production
Analysis
Individual
Analysis
CERN
Raw/Sim ESD
ESD AOD
AOD DPD
AOD DPD
and plots
Tapes
Scheduled
Chaotic
Physics groups
Physicists
Scheduled,
predictable
experiment/
physics groups
C169
Physics
Software
Development
Tier 2
R&D Systems
and Testbeds
Info servers
Code servers
Web Servers
Telepresence
Servers
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Desktops
Training
Consulting
Help Desk
Harvey B Newman (CIT)
Data Grid: Tier2 Layer
Create an Ensemble of (University-Based) Tier2
Data Analysis Centres
E277
Site Architectures Complementary to the the Major
Tier1 Lab-Based Centers
Medium-scale Linux CPU farm, Sun data server, RAID disk array
Less need for 24 X 7 Operation Some lower component costs
Less production-oriented, to respond to local and regional analysis
priorities and needs
Supportable by a small local team and physicists’ help
One Tier2 Center in Each Region (e.g. of the US)
Catalyze local and regional focus on particular sets of physics goals
Encourage coordinated analysis developments emphasizing particular
physics aspects or subdetectors. Example: CMS EMU in Southwest US
Emphasis on Training, Involvement of Students at Universities
in Front-line Data Analysis and Physics Results
Include a high quality environment for desktop remote collaboration
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
MONARC Analysis Process Example
DAQ/RAW
Slow
Control/Cal
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Monarc Analysis Model Baseline:
ATLAS or CMS “Typical” Tier1 RC
CPU Power
Disk space
Tape capacity
Link speed to Tier2
Raw data
ESD data
Selected ESD
Revised ESD
AOD data
Revised AOD
TAG/DPD
~100 KSI95
~100 TB
300 TB, 100 MB/sec
10 MB/sec (1/2 of 155 Mbps)
1%
10-15 TB/year
100%
100-150 TB/year
25%
5 TB/year
[*]
25%
10 TB/year [*]
100%
2 TB/year
[**]
100%
4 TB/year
[**]
100%
200 GB/year
Simulated data
25%
25 TB/year
(repository)
[*] Covering Five Analysis Groups; each selecting ~1%
of Annual ESD or AOD data for a Typical Analysis
[**] Covering All Analysis Groups
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
MONARC Testbeds WG:
Isolation of Key Parameters
Some Parameters Measured,
Installed in the MONARC Simulation Models,
and Used in First Round Validation of Models.
Objectivity AMS Response Time-Function, and its
dependence on
Object clustering, page-size, data class-hierarchy
and access pattern
Mirroring and caching (e.g. with the Objectivity DRO option)
Scalability of the System Under “Stress”:
Performance as a function of the number of jobs,
relative to the single-job performance
Performance and Bottlenecks for a variety of data
access patterns
Tests over LANs and WANs D235, D127
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
MONARC Testbeds WG
Test-bed configuration defined and widely deployed
“Use Case” Applications Using Objectivity:
GIOD/JavaCMS, CMS Test Beams,
ATLASFAST++, ATLAS 1 TB Milestone
Both LAN and WAN tests
ORCA4 (CMS)
A108
First “Production” application
Realistic data access patterns
Disk/HPSS
“Validation” Milestone Carried Out, with Simulation WG
C113
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
MONARC Testbed Systems
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Multitasking Processing Model
A Java 2-Based, CPU- and code-efficient simulation
for distributed systems has been developed F148
Process-oriented discrete event simulation
Concurrent running tasks share resources (CPU, memory, I/O)
“Interrupt” driven scheme:
For each new task or when one task is finished, an interrupt is
generated and all “processing times” are recomputed.
It provides:
An efficient mechanism
to simulate multitask
processing
An easy way to apply
different load balancing
schemes
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Role of Simulation
for Distributed Systems
Simulations are widely recognized and used as essential tools
for the design, performance evaluation and optimisation
of complex distributed systems
From battlefields to agriculture; from the factory floor to
telecommunications systems
Discrete event simulations with an appropriate and
high level of abstraction
Just beginning to be part of the HEP culture
Some experience in trigger, DAQ and tightly coupled
computing systems: CERN CS2 models (Event-oriented)
MONARC (Process-Oriented; Java 2 Threads + Class Lib)
These simulations are very different from HEP “Monte Carlos”
“Time” intervals and interrupts are the essentials
Simulation is a vital part of the study of site architectures,
network behavior, data access/processing/delivery strategies,
for HENP Grid Design and Optimization
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Example : Physics Analysis at
Regional Centres
Similar data processing
jobs are performed in
each of several RCs
Each Centre has “TAG”
and “AOD” databases
replicated.
Main Centre provides
“ESD” and “RAW” data
Each job processes
AOD data, and also a
a fraction of ESD and
RAW data.
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Example: Physics Analysis
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Simple Validation Measurements
The AMS Data Access Case
Simulation
Measurements
180
Distribution of 32 Jobs’
Processing Time
C113
Mean Time per job [ms]
4 CPUs Client
monarc01
160
35
140
LAN
30
RawDB
Data
120
25
100
20
80
15
60
10
40
5
20
0
100
0
0
5
10
15
20
25
Nr. of concurrent jobs
30
105
110
115
35
Simulation
mean 109.5
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Measurement
mean 114.3
Harvey B Newman (CIT)
120
MONARC Phase 3
INVOLVING CMS, ATLAS, LHCb, ALICE
TIMELY and USEFUL IMPACT:
Facilitate the efficient planning and design of mutually
compatible site and network architectures, and services
Among the experiments, the CERN Centre and Regional
Centres
Provide modelling consultancy and service to the
experiments and Centres
Provide a core of advanced R&D activities, aimed at LHC
computing system optimisation and production prototyping
Take advantage of work on distributed data-intensive computing
for HENP this year in other “next generation” projects [*]
For example PPDG
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
MONARC Phase 3
Technical Goal: System Optimisation
Maximise Throughput and/or Reduce Long Turnaround
Phase 3 System Design Elements
RESILIENCE, resulting from flexible management of each data
transaction, especially over WANs
SYSTEM STATE & PERFORMANCE TRACKING, to match and
co-schedule requests and resources, detect or predict faults
FAULT TOLERANCE, resulting from robust fall-back strategies
to recover from bottlenecks, or abnormal conditions
Base developments on large scale testbed prototypes
at every stage: for example ORCA4
[*] See H. Newman, http://www.cern.ch/MONARC/progress_report/longc7.html
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
MONARC Status
MONARC is well on its way to specifying baseline Models
representing cost-effective solutions to LHC Computing.
Discussions have shown that LHC computing has
a new scale and level of complexity.
A Regional Centre hierarchy of networked centres
appears to be the most promising solution.
A powerful simulation system has been developed, and is a
very useful toolset for further model studies.
Synergy with other advanced R&D projects has been identified.
Important information, and example Models have been
provided:
Timely for the Hoffmann Review and discussions of LHC
Computing over the next months
MONARC
Phase 3 has been Proposed
Based on prototypes, with increasing detail and realism
Coupled to Mock Data Challenges in 2000
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
The Particle Physics Data Grid (PPDG)
DoE/NGI Next Generation Internet Project
ANL, BNL, Caltech, FNAL, JLAB, LBNL,
SDSC, SLAC, U.Wisc/CS
Site to Site Data
Replication Service
PRIMARY SITE
SECONDARY SITE
CPU, Disk,
Tape Robot
Data Acquisition,
CPU, Disk,
Tape Robot
100 Mbytes/sec
Coordinated reservation/allocation techniques;
Integrated Instrumentation, DiffServ
First Year Goal: Optimized cached read access to 1-10
Gbytes,
drawn from a total data set of up to One Petabyte
Multi-Site Cached File Access Service
PRIMARY SITE
DAQ, Tape,
CPU,
Disk, Robot
Satellite Site
Tape, CPU,
Satellite
Site
Disk,
Robot
Tape, CPU,
Disk, Robot
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
University
University
CPU,
Disk,
University
CPU,
Disk,
University
Users
CPU,
Disk,
University
Users
CPU, Disk,
Users
CPU, Disk,
Users
Users
Harvey B Newman (CIT)
PPDG: Architecture for Reliable High
Speed Data Delivery
Object-based and
File-based Application
Services
File Access
Service
Matchmaking
Service
File Replication
Index
Cost Estimation
Cache Manager
File Fetching
Service
Mass Storage
Manager
File Mover
File Mover
End-to-End
Network Services
Site Boundary
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Security Domain
Harvey B Newman (CIT)
Distributed Data Delivery and
LHC Software Architecture
Software Architectural Choices
Traditional, single-threaded applications
Allow for data arrival and reassembly
OR
Performance-Oriented (Complex)
I/O requests up-front; multi-threaded; data driven;
respond to ensemble of (changing) cost estimates
Possible code movement as well as data
movement
Loosely coupled, dynamic
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
ALDAP (NSF/KDI) Project
ALDAP:
Accessing Large Data Archives
in Astronomy and Particle Physics
NSF Knowledge Discovery Initiative (KDI)
CALTECH, Johns Hopkins, FNAL(SDSS)
C226
Explore advanced adaptive database structures, physical
data storage hierarchies for archival storage of next
generation astronomy and particle physics data
Develop spatial indexes, novel data organizations,
distribution and delivery strategies, for
Efficient and transparent access to data across networks
Example (Kohonen) Maps for data “self-organization”
Create prototype network-distributed data query execution
systems using Autonomous Agent workers
Explore commonalities and find effective common solutions
for particle physics and astrophysics data
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Beyond Traditional Architectures:
Mobile Agents (Java Aglets)
“Agents are objects with rules and legs” -- D. Taylor
Application
Mobile Agents: Reactive, Autonomous, Goal Driven, Adaptive
Execute Asynchronously
Reduce
Network Load: Local Conversations
Overcome Network Latency; Some Outages
Adaptive Robust, Fault Tolerant
Naturally Heterogeneous
Extensible Concept: Agent Hierarchies
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
D9
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Grid Services Architecture [*]:
Putting it all Together
Applns
HEP Data-Analysis
Related Applications
Appln
Toolkits
Remote
data
toolkit
Grid
Services
Protocols, authentication, policy, resource
management, instrumentation, data discovery, etc.
Grid
Fabric
Remote
comp.
toolkit
Remote
viz
toolkit
Remote
collab.
toolkit
Remote
... sensors
toolkit
Archives, networks, computers, display devices, etc.;
associated local services
[*] Adapted from Ian Foster
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
E403
Harvey B Newman (CIT)
Grid Hierarchy Goals: Better Resource
Use and Faster Turnaround
Efficient resource use and improved responsiveness
through:
Treatment of the ensemble of site and network resources
as an integrated (loosely coupled) system
E163
Resource discovery, query estimation (redirection),
co-scheduling, prioritization, local and global allocations
Network and site “instrumentation”: performance
tracking, monitoring, forward-prediction, problem
E345
trapping and handling
Exploit superior network infrastructures (national,
land-based) per unit cost for frequently accessed data
Transoceanic links relatively expensive
Shorter links normally higher throughput
Ease development, operation, management and security,
through the use of layered, (de facto) standard services
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Grid Hierarchy Concept:
Broader Advantages
Greater flexibility to pursue different physics interests,
priorities, and resource allocation strategies by region
Lower tiers of the hierarchy More local control
Partitioning of users into “proximate” communities
into for support, troubleshooting, mentoring
Partitioning of facility tasks, to manage and focus
resources
“Grid” integration and common services are a principal
means for effective worldwide resource coordination
An Opportunity to maximize global funding resources and
their effectiveness, while meeting the needs for analysis
and physics
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Grid Development Issues
Integration of applications with Grid Middleware
Performance-oriented user application software architecture
needed, to deal with the realities of data access and delivery
Application frameworks must work with system state and
policy information (“instructions”) from the Grid
ODBMS’s must be extended to work across networks
“Invisible” (to the DBMS) data transport, and catalog update
Interfacility cooperation at a new level, across
world regions
Agreement on the use of standard Grid components,
services, security and authentication
Match with heterogeneous resources, performance levels,
and local operational requirements
Consistent policies on use of local resources by remote
communities
Accounting and “exchange of value” software
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Grid Hierarchy Concept:
Broader Advantages
Greater flexibility to pursue different physics interests,
priorities, and resource allocation strategies by region
Lower tiers of the hierarchy More local control
Partitioning of users into “proximate” communities
into for support, troubleshooting, mentoring
Partitioning of facility tasks, to manage and focus
resources
“Grid” integration and common services are a principal
means for effective worldwide resource coordination
An Opportunity to maximize global funding resources and
their effectiveness, while meeting the needs for analysis
and physics
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Content Delivery Networks:
a Web-enabled Pre- “Data Grid”
Worldwide Integrated Distributed Systems
for Dynamic Content Delivery Circa 2000
Akamai, Adero, Sandpiper Server Networks
1200 Thousands of Network-Resident Servers
25 60 ISP Networks
25 30 Countries
40+ Corporate Customers
$ 25 B Capitalization
Resource Discovery
Build “Weathermap” of Server Network (State Tracking)
Query Estimation; Matchmaking/Optimization;
Request rerouting
Virtual IP Addressing
Mirroring, Caching
(1200) Autonomous-Agent Implementation
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
The Need for a “Grid”: the Basics
Computing for LHC will never be “enough” to fully exploit the physics
potential, or exhaust the scientific potential of the collaborations
The basic Grid elements are required to make the ensemble of
computers, networks, storage management systems, and function as a
self-consistent system, implementing consistent (and complex)
resource usage policies.
A basic “Grid” will an information gathering/ workflow guiding/
monitoring/ and repair-initiating entity, designed to ward off resource
wastage (or meltdown) in a complex, distributed and somewhat “open”
system.
Without such information, experience shows that effective global use of
such a large, complex and diverse ensemble of resources is likely to
fail; or at the very least be sub-optimal
The time to accept the charge to build a Grid, for sober and
compelling reasons, is now
Grid-like systems are starting to appear in industry and commerce
But Data Grids on the LHC scale will not be in production until
significantly after 2005
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)
Summary
The HENP/LHC Data Analysis Problem
Petabyte scale compact binary data, and computing
resources distributed worldwide
Development of an integrated robust networked data
access processing and analysis system is mission-critical
An aggressive R&D program is required
to develop reliable, seamless systems that work across
an ensemble of networks
An effective inter-field partnership is now developing
through many R&D projects (PPDG, GriPhyN, ALDAP…)
HENP analysis is now one of the driving forces
for the development of “Data Grids”
Solutions to this problem could be widely applicable in
other scientific fields and industry, by LHC startup
National and Multi-National “Enterprise Resource Planning”
February 10, 2000: Distributed Data Access and Analysis for HENP Experiments
Harvey B Newman (CIT)