Seaborg - MSU Computer Science

Download Report

Transcript Seaborg - MSU Computer Science

Seaborg
Cerise Wuthrich
CMPS 5433
Seaborg

Manufactured by IBM
 Distributed Memory Parallel Supercomputer
 Based on IBM’s SP RS/6000 Architecture
Seaborg
 Used
by National Energy Research
Scientific Computing Center (funded by
Department of Energy) at Berkeley Lab
 Named for Glenn Seaborg – Nobel
laureate chemist who discovered 10
atomic elements, including plutonium
IBM SP RS/6000 Architecture
SP –
Scalable
Power
Parallel
 RS – RISC
System
 Composed
of nodes

Nodes
 416
nodes with 16 processors/node
 380
compute nodes
 20 nodes used for disk storage
 6 login nodes
 2 network nodes
 8 spares
Front and back view of nodes
Node Architecture
 16
IBM Power3
processors per
node
 Between 16 – 64
GB Memory per
node
 2 network
adapters per
node
Processors

IBM Power3 processors each running at 375 MHz
 Power – Performance Optimized With Enhanced
RISC
 PowerPC processors are RISC-based symmetric
multiprocessors (every processor is functionally
identical) with 64-bit addressability
 Connected to L2 cache by bus running at 250 MHz
 Dynamic Branch Prediction
 Instruction prefetching
 FP units are fully pipelined
 4 FLOP/cycle x 375 MHz = 1500 Million or 1.5
GFLOPS/sec
Power PC 3 processor
32 KB
8 MB
64KB
Power3 Processor
15 million transistors
Interconnection Network
 Nodes
connected with high bandwidth,
low latency IBM SP2 switch
 Can be connected in various topologies
depending on number of nodes
 Each switchboard has up to 32 links
 16
links to nodes
 16 links to other
switchboards
Interconnection Network
Star Topology
used for up to
80 nodes and
still guarantee
4 independent
shortest paths
Interconnection Network
Intermediate
switchboards
must be added
for 81-256
nodes
Interconnection Network
 The
combination of HW and SW of the
switch system is known as the CSS –
Communication SubSystem
 Network is highly available
Latency in the network
 Within
nodes,
latency is 9
microseconds
 Between nodes,
using Message
Passing Interface,
the latency is 17
microseconds
Scalability
 Architecture
can handle from 1 – 512
nodes
 The current version of Seaborg (2003)
is twice the size of the original (2001)
Memory
 Within
each node, between 16 & 64 GB
of shared memory
 Between nodes, there is distributed
memory
 Parallel programs can be run using
distributed memory message passing,
shared memory threading or a
combination
I/O
 20
nodes run the distributed parallel I/O
system called GPFS – General Parallel
File System
 44 Terabytes of disk space
 Each node runs its own copy of AIX –
IBM’s Unix-based OS
Production Status/Cost
 $33
Million for the first version put into
operation in June 2001
 At the time, it was the 2nd most powerful
computer in the world and the most
powerful one for unclassified research
 In 2003, number of nodes was doubled
Customers
 2100
researchers at national labs and
universities across the country
 Restricted to Department of Energy
funded massively parallel processing
projects
 Located at the National Energy
Research Computing Center
Applications

Massively Parallel Scientific Research
 Gasoline Combustion Simulation
 Fusion Energy Research
 Climate Modeling
 Materials Science
 Computational Biology
 Particle Simulations
 Plasma Acceleration
 Large Scale Simulation of Atomic Structures
Interesting Features

In 2004, 2.4 times
as many requests
as resources
available
 Uses POE (Parallel
Operating
Environment) and
LoadLeveler to
schedule jobs
Survey Results – Why do you
use Seaborg?





Need massively parallel
computer
High speed
Achieves level of
numerical accuracy
Can run several
simulations in parallel
Easy to connect using
ssh

Fastest and most
efficient computer
available for my
research
 Long queue times
are great
 Large enough
memory for my
needs
Survey Results – How could
Seaborg be improved?

I think too many nodes are scheduled for
many jobs. Scaling is not good in many
cases.
 “..Virtually impossible to do interactive work”
 “Debuggers are terrible.”
 “Compilers and debuggers are a step down
from the Cray.”
 Giving preference to high concurrency jobs
makes smaller jobs wait
Questions