PowerPoint - Computer Sciences Dept.
Download
Report
Transcript PowerPoint - Computer Sciences Dept.
Self-configuring Condor Virtual
Machine Appliances for Ad-Hoc
Grids
Renato Figueiredo
Arijit Ganguly, David Wolinsky, J. Rhett Aultman, P.
Oscar Boykin,
ACIS Lab, University of Florida
http://wow.acis.ufl.edu
Advanced Computing and Information Systems laboratory
Outline
Motivations
Background
Condor Virtual Appliance: features
On-going and future work
Advanced Computing and Information Systems laboratory
2
Motivations
Goal: plug-and-play deployment of Condor grids
•
•
High-throughput computing; LAN and WAN
Collaboration: file systems, messaging, ..
Synergistic approach: VM + virtual network + Condor
• “WOWs” are wide-area NOWs, where:
• Nodes are virtual machines
• Network is virtual: IP-over-P2P (IPOP) overlay
• VMs provide:
• Sandboxing; software packaging; decoupling
• Virtual network provides:
• Virtual private LAN over WAN; self-configuring and
capable of firewall/NAT traversal
• Condor provides:
• Match-making, reliable scheduling, … unmodified
Advanced Computing and Information Systems laboratory
3
Condor WOWs - outlook
5. VMs obtain IP addresses from MyGrid
Virtual DHCP server, join virtual IP network,
discover available manager(s), and join pool
4. Download base and custom
VM images; boot up
5b. VMs obtain IP addresses from OtherGrid
Virtual DHCP server, join virtual IP network,
discover available manager(s), and join pool
3. Create virtual IP namespace
for pool: MyGrid:10.0.0.0/255.0.0.0
Prime custom image with virtual
namespace, desired tools
Bootstrap manager(s)
10.0.0.2
10.0.0.1
2. Download image; boot using
free VM monitor (e.g. VMware
Player or Server)
10.0.0.2
10.0.0.3
10.0.0.4
10.0.0.1
10.0.0.3
10.0.0.4
1. Prime base VM image with O/S, Condor,
Virtual network; publish (Web/Torrent)
Advanced Computing and Information Systems laboratory
4
Condor WOW snapshot
Gainesville
Zurich
Long Beach
Advanced Computing and Information Systems laboratory
5
Roadmap
The basics:
1.1 VMs and appliances
1.2 IPOP: IP-over-P2P virtual network
1.3 Grid Appliance and Condor
The details:
2.1 Customization, updates
2.2 User interface
2.3 Security
2.4 Performance
Usage experience
Advanced Computing and Information Systems laboratory
6
1.1: VMs and appliances
System VMs:
Homogenous system
Sandboxing
Co-exist with
unmodified hosts
Virtual appliances:
• VMware, KVM, Xen
• Hardware/software configuration packaged in
•
easy to deploy VM images
Only dependences: ISA (x86), VMM
Advanced Computing and Information Systems laboratory
7
1.2: IPOP virtual networking
Key technique: IP-over-P2P tunneling
•
•
Interconnect VM appliances
WAN VMs perceive a virtual LAN environment
IPOP is self-configuring
•
•
Avoid administrative overhead of VPNs
NAT and firewall traversal
IPOP is scalable and robust
IPOP networks are isolated
• P2P routing deals with node joins and leaves
•
•
One or more private IP address spaces
Decentralized DHCP serves addresses for each space
Advanced Computing and Information Systems laboratory
8
1.2: IPOP virtual networking
Structured overlay network topology
•
•
•
Bootstrap 1-hop IP tunnels on demand
Discover NAT mappings; decentralized hole punching
VM keeps IPOP address even if it migrates on WAN
• [Ganguly et al, IPDPS 2006, HPDC 2006]
App
IPOP
Node A
B
IPOP
Node B
App
A
tap0
tap0
(10.0.0.3)
(10.0.0.2)
eth0
(128.227.136.244)
eth0
P2P Overlay
Advanced Computing and Information Systems laboratory
(139.70.24.100)
9
1.3 Grid appliance and Condor
Base: Debian Linux; Condor; IPOP
• Works on x86 Linux/Windows/MacOS;
•
VMware, KVM/QEMU
157MB zipped
Uses NAT and host-only NICs
Managed negotiator/collector VMs
Easy to deploy schedd/startd VMs
• No need to get IP address on host network
• Flocking is easy – virtual network is a LAN
Advanced Computing and Information Systems laboratory
10
2.1: Customization and updates
VM image: Virtual Disks
Disks are logically stacked
• Portable medium for data
• Growable after distribution
• Leverage UnionFS file system
• Three stacks:
• Base – O/S, Condor, IPOP
• Module – site specific configuration (e.g. nanoHUB)
• Home – user persistent data
Major updates: replace base/module
• Minor updates: automatic, apt-based
Advanced Computing and Information Systems laboratory
11
2.2: User interface (Windows host)
Host-mounted loop-back Samba folder
VM console: X11 GUI
Loopback
SSH
Advanced Computing and Information Systems laboratory
12
2.2: User interface (Mac host)
Host-mounted loop-back Samba folder
VM console: X11 GUI
Loopback
SSH
Advanced Computing and Information Systems laboratory
13
2.2: User interface (Linux host)
Host-mounted loop-back Samba folder
VM console: X11 GUI
Loopback
SSH
Advanced Computing and Information Systems laboratory
14
2.3 Security
Appliance firewall
• eth0: block all outgoing Internet packets
• Except DHCP, DNS, IPOP’s UDP port
• Only traffic within WOW allowed
• eth1 (host-only): allow ssh, Samba
IPsec
• X.509 host certificates
• Authentication and end-to-end encryption
• VM joins WOW only with signed certificate bound
to its virtual IP
• Private net/netmask: ~10 lines of IPsec
configuration for an entire class A network!
Advanced Computing and Information Systems laboratory
15
2.4: Performance
User-level C# IPOP implementation (UDP):
•
•
Link bandwidth: 25-30Mbit/s
Latency overhead: ~4ms
Connection times:
• ~5-10s to join P2P ring and obtain DHCP address
• ~10s to create shortcuts, UDP hole-punching
89.35
100.00
79.92
80.18
90.00
80.00
70.00
Time
60.00
SimpleScalar 3.0
(cycle-accurate CPU simulator)
50.00
40.00
30.00
20.00
10.00
0.00
Physical
VMWare
Xen
Advanced Computing and Information Systems laboratory
16
Experiences
Bootstrap WOW with VMs at UF and partners
•
•
Currently ~300 VMs, IPOP overlay routers (Planetlab)
Exercised with 10,000s of Condor jobs from real users
• nanoHUB: 3-week long, 9,000-job batch (BioMoca)
•
submitted via a Condor-G gateway
P2Psim, CH3D, SimpleScalar
Pursuing interactions with users and the
Condor community for broader dissemination
Advanced Computing and Information Systems laboratory
17
Time scales and expertise
Development of baseline VM image:
Development of custom module:
Deployment of VM appliance:
• VM/Condor/IPOP expertise; weeks/months
• Domain-specific expertise; hours/days/weeks
• No previous experience with VMs or Condor
• 15-30 minutes to download and install VMM
• 15-30 minutes to download and unzip
•
appliance
15-30 minutes to boot appliance,
automatically connect to a Condor pool, run
condor_status and a demo condor_submit job
Advanced Computing and Information Systems laboratory
18
On-going and future work
Enhancing self-organization at the Condor level:
• Structured P2P for manager publish/discovery
• Distributed hash table (DHT); primary and flocking
• Condor integration via configuration files, DHT scripts
• Unstructured P2P for matchmaking
• Publish/replicate/cache classads on P2P overlay
• Support for arbitrary queries
• Condor integration: proxies for collector/negotiator
Decentralized storage, cooperative caching
•
•
•
Virtual file systems (NFS proxies)
Distribution of updates, read-only code repositories
Caching and COW for diskless, net-boot appliances
Advanced Computing and Information Systems laboratory
19
Acknowledgments
National Science Foundation NMI, CI-TEAM
SURA SCOOP (Coastal Ocean Observing and Prediction)
http://wow.acis.ufl.edu
Publications, Brunet/IPOP code (GPL’ed C#), Condor Grid
appliance
Advanced Computing and Information Systems laboratory
20
Questions?
Advanced Computing and Information Systems laboratory
21
Self-organizing NAT traversal, shortcuts
Sends
CTM
request
Node A
Node B
CTM request: connect to me at my NAT IP:port
- A starts exchanging IP packets with B
- Traffic inspection triggers request to create shortcut
- Connect-to-me (CTM)
- “A” tells “B” its known address(es):
- “A” had learned NATed public IP/port when it joined overlay
Advanced Computing and Information Systems laboratory
22
Self-organizing NAT traversal, shortcuts
Link request: NAT endpoint (IP:port)A
Node A
Node B
CTM reply through overlay: send NAT (IP:port)B
- “B” sends CTM reply – routed through overlay
- “B” tells “A” its address(es)
- “B” initiates linking protocol by attempting to connect to
“A” directly
Advanced Computing and Information Systems laboratory
23
Self-organizing NAT traversal, shortcuts
A Gets
CTM reply;
initiates
linking
Node A
Node B
- B’s linking protocol message to A pokes hole on B’s NAT
- A’s linking protocol message to B pokes hole on A’s NAT
CTM protocol establishes direct shortcut
Advanced Computing and Information Systems laboratory
24
Performance considerations
CPU-intensive application, Condor
• SimpleScalar 3.0d execution-driven computer
architecture simulator
89.35
100.00
79.92
80.18
90.00
80.00
70.00
Time
60.00
50.00
40.00
30.00
20.00
10.00
0.00
Physical
VMWare
Advanced Computing and Information Systems laboratory
Xen
25
Performance considerations
I/O: PostMark
12
• Version 1.51
• Parameters:
• Minimum file
size: 500 bytes
• Maximum file
size: 4.77 MB
• Transactions:
5,000
9.93
10
8
MBs
11.94
5.38
6
Read
4.47
4.28
Write
3.56
4
2
0
Advanced Computing and Information Systems laboratory
Host
Vmware
Xen
26
Performance considerations
User-level C# IPOP implementation (UDP):
•
•
Link bandwidth: 25-30Mbit/s (LAN)
Latency overhead: ~4ms
Connection times:
• (Fine-tuning has reduced mean acquire time to ~ 6-10s, with
degree of redundancy n=8)
Advanced Computing and Information Systems laboratory
27
Condor Appliance on a desktop
VM Hardware configuration
Swap
User files
Domainspecific
tools
Linux,
Condor,
IPOP
Advanced Computing and Information Systems laboratory
28
Related Work
Virtual Networking
Internet Indirection Infrastructure (i3)
IPv6 tunneling
• VIOLIN
• VNET; topology adaptation
• ViNe
• Support for mobility, multicast, anycast
• Decouples packet sending from receiving
• Based on Chord p2p protocol
• IPv6 over UDP (Teredo protocol)
• IPv6 over P2P (P6P)
Advanced Computing and Information Systems laboratory
29