Grid Computing – Introduction
Download
Report
Transcript Grid Computing – Introduction
Grid Computing –
Introduction
Sathish Vadhiyar
Generic Grid
Architecture/Components
Grid Access & Info
Problem Solving
Environments
User Portals
Service Layers
Authentication
Scheduling &
Co- Scheduling
Computers
Naming &
Files
Application Science
Portals
Resource Discovery
& Allocation
Events
Data bases Online instruments
Resource Layer
High speed networks and routers
Fault Tolerance
Software
OK, I have built some software.
Is mine a Grid software?
Ian Foster’s three-point checklist:
1. coordinates resources not subject to
centralized control
2. using standard, open, general-purpose
protocols and interfaces
3. to deliver non-trivial qualities of
service
Some Myriad Definitions
“Coordinated resource sharing and problem
solving in dynamic, multi-institutional virtual
organizations”
“Anatomy of the grid – highly flexible sharing relationships,
sophisticated and precise levels of control over use of shared
resources, sharing of varied resources, diverse usage modes.”
“Controlled sharing – not free
access”
“Infrastructure enabling integrated, collaborative use
of resources”
“Sharing resources can vary dynamically vary over time”
More colorful definitions keep coming
Common keywords: Coordinated, shared, multi-institutions,
controlled, usage, collaboration
Differences with Other
Technologies
Enterprise-level distributed computing – limited
cross-organizational support
Current distributed computing approaches do not
provide a general resource-sharing framework that
addresses Virtual Organization (VO) requirements.
WWW – just client-server. Lacks richer interaction
models
Technologies like CORBA, Java, DCOM – single
organization, limited scope
Some of the Grid techniques complement existing
techniques.
Grids vs Conventional Distributed
Computing (Nemeth and Sunderam)
Distributed Computing
Virtual Pool of nodes
Set of nodes static. Users have login access. They explicitly know
about nodes
VM constructed out of a priori knowledge
Resource assignment implicit
Resource owning
Grid Computing
Virtual Pool of wide range of resources
Set of nodes static/dynamic. Resources dynamic and diverse – can
vary in number, can vary in performance
Difficult for user to get a priori knowledge
User abstraction at resource layers
Resource sharing
Apps. – resource requirements more than can be solved on machines
“owned”
Continued
Nemeth and Sunderam
Motivating examples
SETI@home
To search new life and civilizations
Use individual computers’ idle time through
running SETI@home screen saver
Screen savers retrieves data, analyzes and
reports results back to SETI project
Looking for extra-terrestrial signal over a 12second period
Each work unit takes 10 to 50 hours on an
average computer – 2.4 to 3.8 trillion floating
point operations
Steps and Statistics
Data collected from Arecibo telescope in Puerto Rico onto
tapes and shipped to SETI@home lab in UC, Berkeley.
Break tapes -> work units -> given to users
Find candidate signals reported from users
Other steps:
•Checking data integrity
Statistics
from 1999-2004
•Removing radio frequency interference
(RFI)
Users
•Identify final candidates
Results received
Statistics:
Total CPU time
208,174,383 work units
Floating Point
1,261 tapes
Operations
Images and statistics from SETI
web site
Total
5054812
1459999962
1988719.151 years
5.278185e+21
Average CPU time
11 hr 55 min 56.3 sec
per work unit
Climateprediction.net
Forecast climate in 21st century
3 steps – explore current model,
validate against past climate,
forecast 21st century climate
Different models (in terms of
initial conditions, forcing
[volcanoes, solar activity etc.],
parameters [approximations or
ranges of fixed values in the
model. E.g. ice size in ocean,
friction between different ocean
layers]) distributed to different
users
Massive ensemble experiment
From
climateprediction.net
Steps
Experiment
1
Explore model
sensitivity to
parameters
Goal
Identify suitable ranges
of parameters.
2
Simulation of 19502000
Assess model skill by
making a probability
based forecast of the
past climate.
3
Simulation of 20002100
Make a probability based
forecast of future
climate.
From climateprediction.net
Methodology
Each simulation includes 3
phases:
• Calibration (15yrs)
• Pre-industrial CO2 run
(15yrs)
• Double CO2 run (15yrs)
Run the model with a range of
initial conditions and
parameters for the period
1950-2000. Compare model
outputs with observations to
assess how well the model
performs.
Run the model with a range of
initial conditions, forcings
and parameters for the
period 2000-2100.
Prime number generation - GIMPS
Finding Mersenne prime numbers – 2P-1
GIMPS is to find largest known Mersenne
prime numbers
41st Mersenne prime found recently 224,036,583-1 with 7,235,733 decimal digits !!!
GIMPS found seven
For mostly fun
1000s of Pentium PCs involved. Setup similar
to SETI@home
PCs do primality tests
Other @home Projects
genome@home – designing new genes that form
working proteins in cells. To study protein evolution.
Individual PCs design protein sequences
folding@home – to study why proteins fold/misfold.
Each PC simulates a particular kind of protein folding
evolution@home – to understand and simulate
evolution
Compute-against-cancer – to study cancer cells and to
design new cancer drugs
FightAids@home – screen millions of candidate drug
compounds
Distributed.net – cryptography, secret key challenges
More can be found in
http://boinc.berkeley.edu/projects.php
The Telescience project
Grid for remote
accessing microscopes,
data analysis and
visualization
To study complex
interactions of
molecular and cellular
biological structures and
hence understand brain
diseases
Interactively steer a
400,000-volt electron
microscope at UC San
Diego
From TeleScience web site
References
http://www.globus.org/research/papers/chapter2.pdf
What is the Grid? A three point checklist. Ian Foster. GRIDToday,
July 20, 2002.
The Anatomy of the Grid: Enabling scalable virtual organizations. I.
Foster, C. Kesselman, S. Tuecke. IJSA. 15(3), 2001.
A Complete History of the Grid. Dr. Rob Baxter. Pdf
Zsolt Nemeth, Mauro Migliardi, Dawid Kurzyniec and Vaidy
Sunderam. A comparative analysis of PVM/MPI and computational
grids. In EuroPVM/MPI 2002.
Zsolt Nemeth and Vaidy Sunderam. A comparison of conventional
distributed computing environments and computational grids. ICCS
2002.
Zsolt Nemeth and Vaidy Sunderam. A formal framework for defining
grid systems. CCGrid 2002.