Module 1 - MSL Community Connection

Download Report

Transcript Module 1 - MSL Community Connection

Virtualizing your Datacenter
with Windows Server 2012 R2 & System Center 2012 R2
Module
1
2
3
4
5
Title
Introduction to Microsoft Virtualization & Host Configuration
VM Clustering & Configuration
VM Mobility & Replication, Network Virt and Service Templates
Private Clouds & System Center 2012 R2 Datacenter
Virtualization with the Hybrid Cloud, VMware Management,
Integration & Migration
Microsoft
Virtual
Academy
Microsoft
Virtual
Academy
SCVMM01
DC01
HYPER-V01
HYPER-V02
FS01
Automation
Orchestrator
vCenter Orchestrator
Service Mgmt.
Service Manager
vCloud Automation Center
Protection
Monitoring
Data Protection Manager
System Center 2012 R2
Operations Manager
vSphere Data Protection
vCloud Suite
vCenter&Ops
Mgmt. Suite
vCenter
Self-Service
App Controller
vCloud Director
VM Management
Virtual Machine Manager
vCenter Server
Hypervisor
Hyper-V
vSphere Hypervisor
Automation
Service Mgmt.
Orchestrator
Standard
Datacenter
Service
# of Physical
CPUs per Manager
2
License
Protection
Monitoring
Self-Service
VM Management
Hypervisor
vCenter
Orchestrator
vCloud Suite Licensing
System Center 2012 R2 Licensing
2
2 + Host
Unlimited
Data Protection
Manager
# of Managed OSE’s
per License
Std.
Adv.
Ent.
vCloud
Automation
Center
# of Physical CPUs
1
1
1
per License
Unlimited
VMs on Hosts
vSphere Data
Protection
# of Managed OSE’s
per License
Includes all SC Mgmt.
Components
Yes
Yes
Includes vSphere
5.1 Enterprise Plus
Yes
Yes
Yes
Includes SQL Server
for Mgmt. Server Use
Yes
Yes
Includes vCenter 5.5
No
No
No
Operations Manager
App Controller
$1,323
Open No Level (NL) &
Software Assurance
(L&SA) 2 year Pricing
$3,607
Virtual Machine Manager
Windows Server 2012 R2 Inc. Hyper-V
Hyper-V Server 2012 R2 = Free Download
vCenter Ops Mgmt. Suite
No
vCloud No
Director
Includes all required
database licenses
Retail Pricing per
CPU (No S&S)
$4,995
$7,495
vCenter Server
No
$11,495
vSphere 5.5 Standalone Per CPU Pricing (Excl. S&S):
Standard = $995
Enterprise = $2,875
Enterprise Plus = $3,495
vSphere Hypervisor
Massive scalability for the
most demanding workloads
Hosts
•
Support for up to 320 logical processors
& 4TB physical memory per host
•
Support for up to 1,024 virtual machines
per host
Clusters
•
Support for up to 64 physical nodes &
8,000 virtual machines per cluster
Virtual Machines
•
Support for up to 64 virtual processors
and 1TB memory per VM
•
Supports in-guest NUMA
System
Host
VM
Cluster
1.
2.
Windows Server
2012 R2 Hyper-V
vSphere
Hypervisor
vSphere 5.5
Enterprise Plus
Logical Processors
320
320
320
Physical Memory
4TB
4TB
4TB
Virtual CPUs per Host
2,048
4,096
4,096
Virtual CPUs per VM
64
8
641
1TB
1TB
1TB
1,024
512
512
Guest NUMA
Yes
Yes
Yes
Maximum Nodes
64
N/A2
32
8,000
N/A2
4,000
Resource
Memory per VM
Active VMs per Host
Maximum VMs
vSphere 5.5 Enterprise Plus is the only vSphere edition that supports 64 vCPUs. Enterprise edition supports 32 vCPU per VM with all other editions supporting 8 vCPUs per VM
For clustering/high availability, customers must purchase vSphere
vSphere Hypervisor / vSphere 5.x Ent+ Information: http://www.vmware.com/pdf/vsphere5/r55/vsphere-55-configuration-maximums.pdf and http://www.vmware.com/products/vsphere-hypervisor/faq.html,
http://www.vmware.com/files/pdf/vsphere/VMware-vSphere-Platform-Whats-New.pdf
Centralized, Scalable
Management of Hyper-V
•
Supports up to 1,000 Hyper-V hosts &
25,000 virtual machines per VMM Server
•
Supports Hyper-V hosts in trusted &
untrusted domains, disjointed
namespace & perimeter networks
•
Supports Hyper-V from 2008 R2 SP1
through to 2012 R2
•
Comprehensive fabric management
capabilities across Compute, Network &
Storage
•
End to end VM management across
heterogeneous hosts & clouds
{
Deep Discovery Prior to
Hyper-V Deployment
4
Through integration with the BMC, VMM can
wake a physical server & collect information to
determine appropriate deployment
2
1. OOB Reboot
2. Boot from PXE
3. Authorize PXE boot
5
3
4. Download VMM customized WinPE
5. Execute a set of calls in WinPE to collect
hardware inventory data (network
adapters and disks)
6. Send hardware data back to VMM
6
1
Virtualization Deployment with VMM
Centralized, Automated Bare
Metal Hyper-V Deployment
Post-deep discovery, VMM will deploy a
Hyper-V image to the physical server
1. OOB Reboot
3
2. Boot from PXE
3. Authorize PXE boot
4. Download VMM customized WinPE
5. Run generic command execution scripts
and configure partitions
6. Download VHD & Inject Drivers
The host is then domain joined, added to
VMM Management & post-install scripts
executed
5
Capability
Microsoft
VMware
Deployment from DVD
Yes
Yes
Deployment from USB
Yes
Yes
PXE Deployment - Stateful
Yes – WDS, MDT, SCCM, SCVMM
Yes – PXE/Auto Deploy1
PXE Deployment - Stateless
No
Yes – Auto Deploy
Virtualization Host Configuration
Granular, Centralized
Configuration of Hosts
Virtual Machine Manager 2012 R2 provides
complete, centralized hardware configuration
for Hyper-V hosts
Hardware – Allows the admin to configure
local storage, networking, BMC settings etc.
Storage – Allows the admin control granular
storage settings, such as adding an iSCSI or
FC array LUN to the host, or an SMB share.
Virtual Switches – A detailed view of the
virtual switches associated with physical
network adaptors.
Migration Settings – Configuration of Live
Migration settings, such as LM network,
simultaneous migrations
iSCSI & Fibre
Channel
Multi-Path
I/O Support
Offloaded
Data Transfer
Native 4K
Disk Support
Integrate with existing
storage investments
quickly and easily
Inbox for resiliency,
increased performance
& partner extensibility
Offloads storageintensive tasks to the
SAN
Take advantage of
enhanced density
and reliability
VMM Storage Management
Centralized Management &
Provisioning of Storage
System Center
Virtual Machine Manager 2012 R2
Storage Management
VMM can discover & manage local and
remote storage, including SANs, Pools, LUNs,
disks, volumes, and virtual disks.
VMM supports iSCSI & Fibre Channel Block
Storage & File-based Storage
VMM integrates with WS SMAPI for discovery
of:
•
SMI-S, SMP, and Spaces Devices
•
Disk & Volume management
•
iSCSI/FC/SAS HBA initiator management
R2: 10x faster enumeration of storage
Block Storage
File Storage
Integrated iSCSI Target
Transform Windows Server
2012 R2 into an iSCSI SAN
Integrated Role within Windows Server &
manageable via GUI, PowerShell
Ideal for Network & Diskless Boot, Server
Application Storage, Heterogeneous Storage
& Development, Test & Lab Deployments
Supports up to 64TB VHDX, Thin
Provisioning, Dynamic & Differencing. Also
supports secure zeroing of disk for Fixed size
disk deployments.
Scalable up to 544 sessions & 256 LUNs per
iSCSI Target Server & can be clustered for
resilience
Complete VMM Management via SMI-S
VMM iSCSI & Fibre Channel Integration
Improved Support for Fibre
Channel Fabrics
Once discovered, VMM can centrally manage
key iSCSI & Fibre Channel capabilities.
iSCSI - Connects Hyper-V hosts to iSCSI
portal and logs on to iSCSI target ports
including multiple sessions for MPIO.
Fibre Channel - Add target ports to Zone
•
Zone Management, Member
Management, Zoneset Management
Once connected, VMM can create and assign
LUNs, initialize disks, create partitions,
volumes etc.
VMM can also remove capacity, unmounts
volumes, mask LUNs etc.
Storage
Spaces
Storage
Tiering*
Data
Deduplication
Hyper-V over
SMB 3.0
Transform high-volume,
low cost disks into
flexible, resilient
virtualized storage
Pool HDD & SSD and
automatically move hot
data to SSD for
increased performance
Reduce file storage
consumption, now
supported for live VDI
virtual hard disks*
Ease of provisioning,
increased flexibility &
seamless integration
with high performance
*New in Windows Server 2012 R2
Inbox solution for Windows to
manage storage
Virtualize storage by grouping industrystandard disks into storage pools
Pools are sliced into virtual disks, or Spaces.
Spaces can be Thin Provisioned, and can be
striped across all physical disks in a pool.
Mirroring or Parity are also
supported.
Windows then creates a volume on the
Space, and allows data to be placed on the
volume.
Spaces can use DAS only (local to the chassis,
or via SAS)
Optimizing storage
performance on Spaces
Disk pool consists of both high performance
SSDs and higher capacity HDDs
Hot data is moved automatically to SSD and
cold data to HDD using
Sub-File-Level data movement
With write-back-caching, SSD absorb random
writes that are typical in virtualized
deployments
Admins can pin hot files to SSDs manually to
drive high performance
New PowerShell cmdlets are available for the
management of storage tiers
SSD Tier - 400GB EMLC SAS SSD
Hot Data
Store Hyper-V VMs on SMB
3.0 File Shares
Simplified Provisioning & Management
Low OPEX and CAPEX
Adding multiple NICs in File Servers unlocks
SMB Multichannel – enables higher
throughput and reliability. Requires NICs of
same type and speed.
Using RDMA capable NICs unlocks SMB
Direct offloading network I/O processing to
the NIC.
SMB Direct provides high throughput and
low latency and can reach 40Gbps (RoCE) and
56Gbps (Infiniband) speeds
\\SOFSFileServerName\VMs
File Storage Integration
Comprehensive, Integrated
File Storage Management
VMM supports network shares via SMB 3.0
on NAS device from storage vendors such as
EMC and NetApp
VMM supports integration and management
with standalone and clustered file servers
VMM will quickly discover and inventory
selected File Storage
VMM allows the selection, and now, the
classification of existing File Shares to
streamline VM placement
VMM allows IT Admin to assign Shares to
Hyper-V hosts for VM placement, handling
ACL’ing automatically.
Scale-Out File Server
Low Cost, High Performance,
Resilient Shared Storage
Clustered file server for storing Hyper-V
virtual machine files, on file shares
Scale Out File Server (4 Nodes)
FS1
FS2
FS3
FS4
High reliability, availability, manageability, and
performance that you would expect from a
SAN
Clustered
Spaces
Active-Active file shares - file shares online
simultaneously
Clustered
Pools
Increased bandwidth – as more SOFS nodes
are added
CHKDSK with zero downtime & CSV Cache
Created & Managed by VMM, both from
existing Windows Servers & Bare Metal
JBOD
Storage
via
Shared
SAS
Scale-Out File Server Deployment
Centralized, Managed
Deployment of File Storage
VMM can not only manage standalone File
Servers, but can deploy Scale-Out File
Servers, even to Bare Metal
For Bare Metal deployment, a physical profile
determines the characteristics of the File
Server
Existing Windows Servers can be transformed
into a SOFS, right within VMM
Once imported, VMM can transform
individual disks into highly available,
dynamic pools, complete with classification.
VMM can then create the resilient Spaces &
File Shares within the Storage Pool
Storage & Fabric Classification
Granular Classification of
Storage & FC Fabrics
VMM can classify storage at a granular level
to abstract storage detail:
•
Volumes (including local host disks &
Direct Attached Storage)
•
File Shares (Standalone & SOFS-based)
•
Storage Pools & SAN LUNs
•
Fibre Channel Fabrics - Helps to identify
fabric using friendly names.
Support for efficient & simplified
deployment of VMs to classifications
Now integrated with Clouds
In-box Disk Encryption to
Protect Sensitive Data
Data Protection, built in
•
Supports Used Disk Space Only
Encryption
•
Integrates with TPM chip
•
Network Unlock & AD Integration
Multiple Disk Type Support
•
Direct Attached Storage (DAS)
•
Traditional SAN LUN
•
Cluster Shared Volumes
•
Windows Server 2012 File Server Share
Hyper-V
(2012 R2)
vSphere Hypervisor
vSphere 5.5
Enterprise Plus
iSCSI/FC Support
Yes
Yes
Yes
3rd Party Multipathing (MPIO)
Yes
No
Yes (VAMP)1
Yes (ODX)
No
Yes (VAAI)2
Yes (Spaces)
No
Yes (vSAN)3
Yes
No
Yes4
Yes (SMB 3.0)
Yes (NFS)
Yes (NFS)
Data Deduplication
Yes
No
No
Storage Encryption
Yes
No
No
Capability
SAN Offload Capability
Storage Virtualization
Storage Tiering
Network File System Support
1.
2.
3.
4.
vSphere API for Multipathing (VAMP) is only available in Enterprise & Enterprise Plus editions of vSphere 5.5
vSphere API for Array Integration (VAAI) is only available in Enterprise & Enterprise Plus editions of vSphere 5.5
vSphere vSAN is still in beta
vSphere Flash Read Cache has a write-through caching mechanism only, so reads only are accelerated. vSAN also has SSD
caching capabilities built in, acting as a read cache & write buffer.
vSphere Hypervisor / vSphere 5.x Ent+ Information: http://www.vmware.com/pdf/vsphere5/r55/vsphere-55-configuration-maximums.pdf, http://www.vmware.com/files/pdf/vsphere/VMwarevSphere-Platform-Whats-New.pdf, http://www.vmware.com/products/vsphere/compare.html,
Integrated Solution for
Network Card Resiliency
•
Vendor agnostic and shipped inbox
•
Provides local or remote management
through Windows PowerShell or UI
•
Enables teams of up to 32 network
adapters
•
Aggregates bandwidth from multiple
network adapters whilst providing traffic
failover in the event of NIC outage
•
Includes multiple nodes: switch dependent
and independent
•
Multiple traffic distribution algorithms:
Hyper-V Switch Port, Hashing and
Dynamic Load Balancing
Virtual
adapters
Team network
adapter
Team network
adapter
Connecting VMs to each
other, and the outside world
3 Types of Hyper-V Network
• Private = VM to VM Communication
• Internal = VM to VM to Host (loopback)
• External = VM to Outside & Host
Each vNIC can have multiple VLANs attached to it,
however if using the GUI, only a single VLAN ID can
be specified.
Set-VMNetworkAdapterVlan -VMName VM01
-Trunk -AllowedVlanIdList 14,22,40
Creating an external network transforms the chosen
physical NIC into a switch and removes TCP/IP stack
and other protocols
Optional host vNIC is created to allow
communication of host out of the physical NIC
Hyper-V Host
VM1
VM2
Layer-2 Network Switch for
Virtual Machine Connectivity
Hyper–V host
Virtual machine
Extensible Switch
•
Virtual Ethernet switch that runs in the
management OS of the host
•
Exists on Windows Server Hyper-V, and
Windows Client Hyper-V
•
Managed programmatically
•
Extensible by partners and customers
•
Virtual machines connect to the
extensible switch with their
virtual network adaptor
•
Can bind to a physical NIC or team
•
Bypassed by SR-IOV
Virtual machine
Network
application
Virtual network
adapter
Virtual machine
Network
application
Network
application
Virtual network
adapter
Virtual network
adapter
Hyper-V
Extensible Switch
Physical network
adapter
Physical switch
Layer-2 Network Switch for
Virtual Machine Connectivity
Granular In-box Capabilities
•
Isolated (Private) VLAN (PVLANs)
•
ARP/ND Poisoning (spoofing)
protection
•
DHCP Guard protection
•
Virtual Port ACLs
•
Trunk Mode to VMs
•
Network Traffic Monitoring
•
PowerShell & WMI Interfaces for
extensibility
Hyper–V host
Virtual machine
Virtual machine
Network
application
Virtual network
adapter
Virtual machine
Network
application
Network
application
Virtual network
adapter
Virtual network
adapter
Hyper-V
Extensible Switch
Physical network
adapter
Physical switch
Build Extensions for Capturing,
Filtering & Forwarding
2 Platforms for Extensions
•
Network Device Interface Specification
(NDIS) filter drivers
•
Windows Filtering Platform (WFP)
callout drivers
Extensions
•
NDIS filter drivers
•
WFP callout drivers
•
Ingress filtering
•
Destination lookup and forwarding
•
Egress filtering
Virtual Machine
Virtual Machine
Parent Partition
VM NIC
Host NIC
Virtual Switch
Extension Protocol
Capture
Extensions
Extension
A
Filtering
Extensions
Extension
C
Forwarding
Extension
Extension
D
Extension Miniport
Physical NIC
Hyper-V Extensible Switch architecture
VM NIC
Build Extensions for Capturing,
Filtering & Forwarding
Many Key Features
Virtual Machine
Virtual Machine
Parent Partition
VM NIC
Host NIC
•
Extension monitoring & uniqueness
•
Extensions that learn VM life cycle
•
Extensions that can veto state changes
Extension Protocol
•
Multiple extensions on same switch
Capture
Extensions
Extension
A
Several Partner Solutions Available
•
Cisco – Nexus 1000V & UCS-VMFEX
•
NEC – ProgrammableFlow PF1000
•
5nine – Security Manager
•
InMon - SFlow
Virtual Switch
Filtering
Extensions
Extension
C
Forwarding
Extension
Extension
D
Extension Miniport
Physical NIC
Hyper-V Extensible Switch architecture
VM NIC
Hyper-V
(2012 R2)
vSphere
Hypervisor
vSphere 5.5
Enterprise Plus
Integrated NIC Teaming
Yes
Yes
Yes
Extensible Network Switch
Yes
No
Replaceable
Confirmed Partner Solutions
5
N/A
2
Private Virtual LAN (PVLAN)
Yes
No
Yes1
ARP Spoofing Protection
Yes
No
vCloud/Partner2
DHCP Snooping Protection
Yes
No
vCloud/Partner2
Virtual Port ACLs
Yes
No
vCloud/Partner2
Trunk Mode to Virtual Machines
Yes
No
Yes3
Port Monitoring
Yes
Per Port Group
Yes3
Port Mirroring
Yes
Per Port Group
Yes3
Advanced Networking Capability
1.
2.
3.
The vSphere Distributed Switch (required for PVLAN capability) is available only in the Enterprise Plus edition of vSphere 5.5 and
is replaceable (By Partners such as Cisco/IBM) rather than extensible.
ARP Spoofing, DHCP Snooping Protection & Virtual Port ACLs require the vCloud Networking & Security package, which is part
of the vCloud Suite or a Partner solution, all of which are additional purchases
Trunking VLANs to individual vNICs, Port Monitoring and Mirroring at a granular level requires vSphere Distributed Switch, which
is available in the Enterprise Plus edition of vSphere 5.5
vSphere Hypervisor / vSphere 5.x Ent+ Information: http://www.vmware.com/products/cisco-nexus-1000V/overview.html, http://www-03.ibm.com/systems/networking/switches/virtual/dvs5000v/,
http://www.vmware.com/technical-resources/virtualization-topics/virtual-networking/distributed-virtual-switches.html, http://www.vmware.com/files/pdf/techpaper/Whats-New-VMware-vSphere51-Network-Technical-Whitepaper.pdf, and http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9902/data_sheet_c78-492971.html, http://www.vmware.com/products/vcloudnetwork-security,
Comprehensive Network Management
Integrated management of the
software defined network
Top of rack switch management and
integration for configuration and compliance
Logical network management: named
networks that serve particular functions
in your environment i.e. backend
IP address pool management and integration
with IP address management
Host and VM network switch management
Load balancer integration and
automated deployment
Network virtualization deployment
and management
Top of Rack Switch Integration
Synchronize & Integrate ToR
Settings with VMM
Physical switch management and integration
built into VMM using in-box or partnersupplied provider
Switches running Open Management
Infrastructure (OMI)
Communicating using
WS-MAN
Switch Management PowerShell Cmdlets
Common management interface across
multiple network vendors
Automate common network management
tasks
Manage compliancy between VMM, Hyper-V
Hosts & physical switches.
OMI
OMI
OMI
Logical Networks
Abstraction of Infrastructure
Networks with VMM
Logical networks are named networks that
serve particular functions i.e. “Backend,”
“Frontend,” or “Backup”.
Used to organize and simplify network
assignments
Logical network is a container for network
sites, IP subnet & VLAN information
Supports VLANs & PVLAN Isolation
Hosts & Host Groups can be associated with
Logical Networks
IP Addresses can be assigned to Host & VM
NICs from Static IP Pools
Static IP Pool Management in VMM
IP Address Management for
Hosts & Virtual Machines
VMM can maintain centralized control of host
& VM IP address assignment
IP Pools defined and associated with a
Logical Network & Site
VMM supports specifying IP range, along
with VIPs & IP address reservations
Each IP Pool can have Gateway, DNS & WINS
Configured.
IP address pools support both IPv4 and IPv6
addresses, but not in the same pool.
IP addresses assigned on VM creation, and
retrieved on VM deletion
The Logical Switch
Centralized Configuration of
Network Adaptors across Hosts
Combines key VMM networking constructs to
standardize deployment across multiple hosts
within the infrastructure:
•
Uplink Port Profiles
•
Virtual Port Profiles for vNICs
•
Port Classifications for vNICs
•
Switch Extensions
Logical Switches support compliance &
remediation
Logical Switches support Host NIC Teaming
& Converged Networking
Native Port Profile for Uplinks
Port Classification
Native Port Profile for vNIC
Port Classification
Native Port Profile for vNIC
Uplink Port Profiles
Host Physical Network Adaptor
Configuration with VMM
Uplink Port Profile – centralized
configuration of physical NIC settings that
VMM will apply upon assigning a Logical
Switch to a Hyper-V host.
Teaming – Automatically created when
assigned to multiple physical NICs, but
admin can select LB algorithm &
teaming mode
Sites – Assign the relevant network sites &
logical networks that will be supported by
this uplink port profile
Virtual Port Profiles
Host Physical Network Adaptor
Configuration with VMM
Virtual Port Profile – Used to pre-configure
VM or Host vNICs with specific settings.
Offloading – Admins can enable offload
capabilities for a specific vNIC Port Profile.
Dynamic VMq, IPsec Task Offload & SR-IOV
are available choices.
Security – Admins can enable key Hyper-V
security settings for the vNIC Profile, such as
DHCP Guard, or enable Guest Teaming.
QoS – Admins can configure QoS bandwidth
settings for the vNIC Profile so when
assigned to VMs, their traffic may be
limited/guaranteed.
Increased efficiency of network
processing on Hyper-V hosts
Without VMQ
•
Hyper-V Virtual Switch is responsible for
routing & sorting packets for VMs
•
This leads to increased CPU processing, all
focused on CPU0
With VMQ
•
Physical NIC creates virtual network
queues for each VM to reduce host CPU
With Dynamic VMQ
•
Processor cores dynamically allocated for
a better spread of network traffic
processing
Hyper-V Host
Hyper-V Host
Hyper-V Host
Integrated with NIC hardware
for increased performance
•
Standard that allows PCI Express devices
to be shared by multiple VMs
•
More direct hardware path for I/O
•
Reduces network latency, CPU utilization
for processing traffic and increases
throughput
•
SR-IOV capable physical NICs contain
virtual functions that are securely
mapped to VM
•
This bypasses the Hyper-V Extensible
Switch
•
Full support for Live Migration
Virtual Machine
VM Network Stack
Synthetic NIC
Virtual Function
Hyper-V
Extensible Switch
SR-IOV NIC
VF
VF
VF
Achieve desired levels of
networking performance
Bandwidth Management
•
Establishes a bandwidth floor
•
Assigns specified bandwidth for each type
of traffic
•
•
Helps to ensure fair sharing during
congestion
Can exceed quota with no congestion
2 Mechanisms
•
Enhanced packet scheduler (software)
•
Network adapter with DCB support
(hardware)
Relative minimum
bandwidth
Normal
priority
W=1
High
priority
Strict minimum
bandwidth
Bronze
tenant
Critical
W=2
W=5
Silver
tenant
100 MB
Hyper-V Extensible Switch
Gold
tenant
200 MB
Hyper-V Extensible Switch
1 Gbps
Bandwidth
oversubscription
Gold
tenant
Gold
tenant
500 MB
Gold
tenant
500 MB
Hyper-V Extensible Switch
NIC Teaming
1 Gbps
1 Gbps
500 MB
500 MB
Port Classifications
Abstract Technical Depth from
Virtual Network Adaptors
Port Classifications – provides a global
name for identifying different types of virtual
network adapter port profiles
Cross-Switch - Classification can be used
across multiple logical switches while the
settings for the classification remain specific
to each logical switch
Simplification – Similar to Storage
Classification, Port Classification used to
abstract technical detail when deploying VMs
with certain vNICs. Useful in Self-Service
scenarios.
Constructing the Logical Switch
Combining Building Blocks to
Standardize NIC Configuration
Simple Setup – Define the name and
whether SR-IOV will be used by VMs.
SR-IOV can only be enabled at switch
creation time.
Switch Extensions – Pre-installed/
Configured extensions available for use with
this Logical Switch are chosen at this stage
Teaming – Decide whether this logical switch
will bind to individual NICs, or to NICs that
VMM should team automatically.
Virtual Ports – Define which port
classifications and virtual port profiles can be
used with this Logical Switch
Deploying the Logical Switch
Applying Standardized
Configuration Across Hosts
Assignment – VMM can assign logical
switches directly to the Hyper-V hosts.
Teaming or No Teaming – Your logical
switch properties will determine if multiple
NICs are required or not
Converged Networking – VMM can create
Host Virtual Network Adaptors for isolating
host traffic types i.e. Live Migration, CSV, SMB
3.0 Storage, Management etc. It will also
issue IP addresses from it’s IP Pool. This is
useful with hosts that have just 2 x 10GbE
adaptors but require multiple separate,
resilient networks.