Parallel computing

Download Report

Transcript Parallel computing

Lecture 2 :
Introduction to Multicore Computing
Bong-Soo Sohn
Associate Professor
School of Computer Science and Engineering
Chung-Ang University
Multicore Processor

A single computing component with two or
more independent cores




Core (CPU): computing unit that reads/executes program instructions
Ex) dual-core, quad-core, hexa-core, octa-core, …
share cache or not
symmetric or asymmetric
Intel Quad-core processors (Sandy Bridge)
Multicore Processor

Multiple cores run multiple instructions at the
same time

Increase overall program speed

performance gained by multi-core processor

strongly dependent on the software algorithms and
implementation.
Manycore processor (GPU)

multi-core architectures with an especially
high number of cores (hundreds or more)


CUDA




Ex) nVidia GeForce GTX 780 Ti
Compute Unified Device Architecture
parallel computing platform and programming model created
by NVIDIA and implemented by the graphics processing units
(GPUs) that they produce
GPGPU (General Purpose Graphics Processing Unit)
OpenCL

Open Standard parallel programming platform
Thread/Process

Both


Process


run in separate memory space
Thread



Independent sequence of execution
A process with two
threads of execution on
a single processor
Run in shared memory space in a process
One process may have multiple threads
Multithreaded Program

a program running with multiple threads that is executed
simultaneously
What is Parallel Computing?

Parallel computing


Examples of parallel machines:




using multiple processors in parallel to solve problems
more quickly than with a single processor
A cluster computer that contains multiple PCs combined
together with a high speed network
A shared memory multiprocessor by connecting multiple
processors to a single memory system
A Chip Multi-Processor (CMP) contains multiple processors
(called cores) on a single chip
Concurrent execution comes from desire for
performance; unlike the inherent concurrency in a
multi-user distributed system
Parallelism vs Concurrency

Parallel Programming



Using additional computational resources to
produce an answer faster
Problem of using extra resources effectively?
Example

summing up all the numbers in an array with multiple n
processors
Parallelism vs Concurrency

Concurrent Programming



Correctly and efficiently controlling access by
multiple threads to shared resources
Problem of preventing a bad interleaving of
operations from different threads
Example

Implementation of dictionary with hashtable



operations insert, update, lookup, delete occur
simultaneously (concurrently)
Multiple threads access the same hashtable
Web Visit Counter
Parallelism vs Concurrency

Often Used interchangeably

In practice, the distinction between
parallelism and concurrency is not
absolute

Many multithreaded programs have
aspects of both
Parallel Programming Techniques

Shared Memory


OpenMP, pthreads
Distributed Memory

MPI

Distributed/Shared Memory
 Hybrid (MPI+OpenMP)

GPU Parallel Programming


CUDA programming (NVIDIA)
OpenCL
Parallel Processing Systems

Small-Scale Multicore Environment






Notebook, Workstation, Server
OS supports multicore
POSIX threads (pthread) , win32 thread
GPGPU-based supercomputer
Development of CUDA/OpenCL/GPGPU
Large-Scale Multicore Environment




Supercomputer : more than 10,000 cores
Clusters
Servers
Grid Computing
Parallel Computing vs. Distributed Computing

Parallel Computing



all processors may have access to a shared memory to
exchange information between processors.
more tightly coupled to multi-threading
Distributed Computing



multiple computers communicate through network
each processor has its own private memory
(distributed memory).
executing sub-tasks on different machines and then merging
the results.
Parallel Computing vs. Distributed Computing
Distributed Computing
Parallel Computing

No Clear Distinction
Cluster Computing vs. Grid Computing

Cluster Computing




a set of loosely connected computers that work together so
that in many respects they can be viewed as a single system
good price / performance
memory not shared
Grid Computing


federation of computer resources from multiple locations to
reach a common goal (a large scale distributed system)
grids tend to be more loosely coupled, heterogeneous, and
geographically dispersed
Cluster Computing vs. Grid Computing
Cloud Computing

shares networked computing resources rather than
having local servers or personal devices to handle
applications.

“Cloud” is used as a metaphor for “Internet"
meaning "a type of Internet-based computing,“

different services - such as servers, storage and
applications - are delivered to an user’s computers
and smart phones through the Internet.
Good Parallel Program

Writing good parallel programs






Correct (Result)
Good Performance
Scalability
Load Balance
Portability
Hardware Specific Utilization
Moore’s Law : Review



Doubling of the number of transistors on integrated circuits
roughly every two years.
Microprocessors have become smaller, denser, and more
powerful.
processing speed, memory capacity, sensors and even the
number and size of pixels in digital cameras.All of these are
improving at (roughly) exponential rates
Computer Hardware Trend

Chip density is continuing
increase ~2x every 2years
Clock speed is not
(in high clock speed, power
consumption and heat generation
is too high to be tolerated.)
 # of cores may double instead

No more hidden parallelism
(ILP;instruction level parallelism)
to be found


Transistor# still rising
Clock speed flattening sharply

Need Multicore programming!

Source: Intel, Microsoft (Sutter) and
Stanford (Olukotun, Hammond)
Examples of Parallel Computer

Chip MultiProcessor (CMP)



Symmetric Multiprocessor (SMP)


Sun Fire E25K
Heterogeneous Chips


Intel Core Duo
AMD Dual Core
Cell Processor
Clusters
 Supercomputers
Intel Core Duo

Two 32-bit Pentium processors
 Each has its own 32K L1 cache
 Shared 2MB or 4MB L2 cache
 Fast communication through shared L2
 Coherent shared memory
AMD Dual Core Opteron



Each with 64K L1 cache
Each with 1MB L2 cache
Coherent shared memory
Intel vs. AMD

Main difference : L2 cache position

AMD




More core private memory
Easier to share cache coherency info with other CPUs
Preferred in multi chip systems
Intel



Core can use more of the shared L2 at times
Lower latency communication between cores
Preferred in single chip systems
Generic SMP

Symmetric MultiProcessor (SMP) System







multiprocessor hardware architecture
two or more identical processors are connected to a single
shared memory
controlled by a single OS instance
Most common multiprocessor systems today use an SMP
architecture
Both Multicore and multi-CPU
Single logical memory image
Shared bus often bottleneck
GPGPU : NVIDIA GPU

Tesla K20





GPU : 1 Kepler GK110
2496 cores; 706MHz
Tpeak 3.52Tflop/s – 32bit floating point
Tpeak 1.17Tflop/s – 64bit floating point
GTX 680

1536 CUDA cores; 1.0GHz
Hybrid Programming Model

Main CPU performs hard to parallelize portion
 Attached processor (GPU) performs compute
intensive parts
Summary

All computers are now parallel computers!

Multi-core processors represent an important new
trend in computer architecture.



Decreased power consumption and heat generation.
Minimized wire lengths and interconnect latencies.
They enable true thread-level parallelism with great
energy efficiency and scalability.
Summary

To utilize their full potential, applications will need to
move from a single to a multi-threaded model.



Parallel programming techniques likely to gain importance.
Hardware/Software
the software industry needs to get back into the state
where existing applications run faster on new
hardware.
Why writing (fast) parallel
programs is hard
Principles of Parallel Computing





Finding enough parallelism (Amdahl’s Law)
granularity
Locality
Load balance
Coordination and synchronization
All of these things makes parallel programming
even harder than sequential programming.
Finding Enough Parallelism


Suppose only part of an application seems parallel
Amdahl’s law


let s be the fraction of work done sequentially, so
(1-s) is fraction parallelizable
P = number of processors
Speedup(P) = Time(1)/Time(P)
<= 1/(s + (1-s)/P)
<= 1/s
• Even if the parallel part speeds up perfectly
performance is limited by the sequential part
Overhead of Parallelism

Given enough parallel work, this is the biggest barrier to
getting desired speedup

Parallelism overheads include:




cost of starting a thread or process
cost of communicating shared data
cost of synchronizing
extra (redundant) computation

Each of these can be in the range of milliseconds (=millions
of flops) on some systems

Tradeoff: Algorithm needs sufficiently large units of work to
run fast in parallel (I.e. large granularity), but not so large that
there is not enough parallel work
Locality and Parallelism
Conventional
Storage
Proc
Hierarchy
Cache
L2 Cache


L2 Cache
L2 Cache
L3 Cache
L3 Cache
L3 Cache
Memory
Memory
Memory
Large memories are slow, fast memories are small
Storage hierarchies are large and fast on average
Parallel processors, collectively, have large, fast cache


Proc
Cache
the slow accesses to “remote” data we call “communication”
Algorithm should do most work on local data
potential
interconnects

Proc
Cache
Load Imbalance

Load imbalance is the time that some
processors in the system are idle due to



insufficient parallelism (during that phase)
unequal size tasks
Algorithm needs to balance load