Welcome Note! How To Survive SCU 2015 Europe!

Download Report

Transcript Welcome Note! How To Survive SCU 2015 Europe!

Windows Server 2016
Markus Erlacher
CEO
itnetX AG
www.itnetx.ch
Bringing the cloud to you
What if you could have the control of the datacenter
and the power of the cloud?
Reduced developer chaos and shadow IT
Simplified, abstracted experience
Added consistency to modern development
One portal, one user experience
Write once, deploy anywhere
Nano Server
Voice of the Customer
 Reboots impact my business


Why do I have to reboot because of a patch to a component I never use?
When a reboot is required, the systems need to be back in service ASAP
 Server images are too big



Large images take a long time to install and configure
Transferring images consumes too much network bandwidth
Storing images requires too much disk space
 Infrastructure requires too many resources


If the OS consumes fewer resources, I can increase my VM density
Higher VM density lowers my costs and increases my efficiency & margins
I want just the components
I need
and nothing more
Microsoft Server Journey
Server Roles/Features
GUI Shell
Windows/
WindowsNT
Full Server
Server Core
Windows NT to
Windows Server
Windows
NT /
2003
Windows
Server 2003
Windows Server 2008
and
Windows Server
2008
Windows Server 2008 R2
Windows Server 2008 R2
Minimal Server
Interface
Server Core
Windows Server 2012
and 2012
Windows Server
Windows Server 2012 R2
Windows Server 2012 R2
Why we need Nano Server
 Azure
 Patches and reboots interrupt service delivery
 (*VERY large # of servers) * (large OS resource consumption)
 Provisioning large host images competes for network
resources
Why we need Nano Server
 Cloud Platform System (CPS)
 Cloud-in-box running on 1-4 racks using System Center &
Windows Server
 Setup time needs to be shortened
 Patches and reboots result in service disruption
 Fully loaded CPS would live migrate > 16TB for every host OS
patch
 Network capacity could have otherwise gone to business uses
 Reboots: Compute host ~2 minutes / Storage host ~5 minutes
We need server configuration optimized for the
cloud
Nano Server - Next Step in the Cloud Journey
 A new headless, 64-bit only, deployment option for
Windows Server
 Deep refactoring focused on
Basic Client
 CloudOS infrastructure
 Born-in-the-cloud applications
 Follow the Server Core pattern
Nano Server
Experience
Server with Local
Admin Tools
Server Core
Nano Server - Roles & Features
 Zero-footprint model
 Server Roles and Optional Features live outside of Nano Server
 Standalone packages that install like applications
 Key Roles & Features
 Hyper-V, Storage (SoFS), and Clustering
 Core CLR, ASP.NET 5 & PaaS
 Full Windows Server driver support
 Antimalware Built-in
 System Center and Apps Insight agents to follow
Nano Server in Windows Server 2016
 An installation option, like Server Core
 Not listed in Setup because image must be customized with
drivers
 Separate folder on the Windows Server media
 Available since the Windows
Server Technical Preview 2
released at Ignite
Installing Drivers
 For the leanest image, install just the drivers your
hardware requires
 Dism /Add-Driver /driver:<path>
 Nano Server includes a package of all drivers in Server
Core
 Dism /Add-Package /PackagePath:.\packages\Microsoft-NanoServerOEM-Drivers-Package.cab
 To run Nano Server as a VM install
 Dism /Add-Package /PackagePath:.\packagesMicrosoft-NanoServer-
Deploying Nano Server
 Generate a VHD from NanoServer.wim
 Download Convert-WindowsImage.ps1 from the Script
Center:
https://gallery.technet.microsoft.com/scriptcenter/ConvertWindowsImageps1-0fe23a8f
 Run Convert-WindowsImage.ps1 -Sourcepath <path to wim>
-VHD <path to new VHD file> –VHDformat VHD -Edition 1.
 Dism /Apply-Image
Installing Roles and Features
 Nano Server folder has a Packages sub-folder
 Dism /Add-Package /PackagePath:.\packages\<package>
 Dism /Add-Package /PackagePath:.\packages\enus\<package>
Installing Agents and Tools on Nano Server
 No MSI support in Nano Server
 Current builds of Nano Server require xcopy or custom
PowerShell script
 Nano Server Installer in the works, which will provide




Install
Uninstall
Inventory
Online and offline installation support
Installing Software on Nano Server
 No MSI support in Nano Server
 Current builds of Nano Server require xcopy or custom
PowerShell script
 Nano Server Installer in the works, which will provide




Install
Uninstall
Inventory
Online and offline installation support
Demo: Nano Server Deployment
Hyper-V
We are winning
virtualization share
x86 Server Virtualization Share
For The Past 5+ Years
Microsoft
Hyper-V
Server
Q1 CY2008
Q3 CY2009
Q3 CY2012
Q3 CY2014
Change
Windows
Server 2008
Released
Windows
Server 2008 R2
Released
Windows
Server 2012
Released
CURRENT
Since
Hyper-V
Released
0.0%
11.8%
25.9%
30.6%
+30.6 Pts
40.0%
46.6%
51.4%
46.2%
+6.2 Pts
ESX
Source: IDC WW Quarterly Server Virtualization Tracker, December 2014. Hyper-V and ESX + vSphere shares based on percent market share among all x86
new hypervisor deployments (nonpaid and paid). x86 hypervisor shipments include those sold on new servers, new nonpaid hypervisor deployments
aboard new servers, and new hypervisor sales and nonpaid hypervisor deployments on installed base servers. Share gains for Hyper-V and ESX + vSphere
come in part from market share transfers from older products from same vendors.
Security
Challenges in protecting high-value assets
Confidently protect sensitive customer data:
Designed for ‘zero-trust’ environments
Shielded VMs
Spotlight capabilities
Shielded Virtual Machines can only run in
fabrics that are designated as owners of
that virtual machine
Shielded Virtual Machines will need
to be encrypted (by BitLocker or other
means) in order to ensure that only the
designated owners can run this virtual
machine
You can convert a running
Generation 2 virtual machine into a
Shielded Virtual Machine
Host Guardian
Service
Secure Boot Support for Linux
 Providing kernel code integrity protections for Linux guest
operating systems.
 Works with:
• Ubuntu 14.04 and later
• SUSE Linux Enterprise Server 12
Secure Boot Support for Linux
 PowerShell to enable this:
Set-VMFirmware "Ubuntu" -SecureBootTemplate
MicrosoftUEFICertificateAuthority
Resiliency & Availability
Storage and Cluster Resiliency
Virtual Machine Storage Resiliency
 Storage fabric outage no longer means that virtual machine
crash
• Virtual machines pause and resume automatically in
response to storage fabric problems
Virtual Machine Cluster Resiliency
 VMs continue to run even when a node falls out of cluster
membership
 Resiliency to transient failures
 Repeat offenders are “quarantined”
Shared VHDX
Improved Shared VDHX
 Host Based Backup of Shared VHDX
 Online Resize of Shared VHDX
Replica Support for Hot Add of VHDX
 When you add a new virtual hard disk to a virtual machine
that is being replicated – it is automatically added to the
not-replicated set. This set can be updated online.
Set-VMReplication "VMName" -ReplicatedDisks
(Get-VMHardDiskDrive "VMName")
Runtime Memory Resize
 Dynamic memory is great, but more can be done.
 For Windows Server 2016 guests, you can now increase and
decrease the memory assigned to virtual machines while
they are running.
Hot add / remove of network adapters
 Network adapters can be added and removed from
Generation 2 virtual machines while they are running.
Servicing & Upgrades
Rolling Cluster Upgrade
 You can now upgrade a 2012 R2 Hyper-V cluster to
Windows Server 2016 with:
• No new hardware
• No downtime
• The ability to roll-back safely if needed
New VM Upgrade Process
 Windows Server 2016:
• Hyper-V will not automatically upgrade virtual machines
• Upgrading a virtual machine is a manual operation that is
separate from upgrading the host
• Individual virtual machines can be moved back to earlier
versions, until they have been manually upgraded
New VM Upgrade Process
 Windows Server 2016:
 PowerShell only:
Update-VMConfigurationVersion
Changing how we handle VM servicing
 Windows Server 2016:
• VM drivers (integration services) updated when needed
• Require latest available VM drivers for that guest operating
system
• Drivers delivered directly to the guest operating system via
Windows Update
Scale Improvements
Evolving Hyper-V Backup
 New architecture to improve reliability, scale and
performance.
• Decoupling backing up virtual machines from backing up
the underlying storage.
• No longer dependent on hardware snapshots for core
backup functionality, but still able to take advantage of
hardware capabilities when they are present.
Built in change tracking for Backup
 Most Hyper-V backup solutions today implement kernel
level file system filters in order to gain efficiency.
• Makes it hard for backup partners to update to newer
versions of Windows
• Increases the complexity of Hyper-V deployments
 Efficient change tracking for backup is now part of the
platform
VM Configuration Changes

•
•

•
New virtual machine configuration file
Binary format for efficient performance at scale
Resilient logging for changes
New file extensions
.VMCX and .VMRS
Operations
Production Checkpoints
 Delivers the same Checkpoint experience that you had in
Windows Server 2012 R2 – but now fully supported for
Production Environments
• Uses VSS instead of Saved State to create checkpoint
• Restoring a checkpoint is just like restoring a system backup
PowerShell Direct to Guest OS
 You can now script PowerShell in the Guest OS directly from
the Host OS
 No need to configure PowerShell Remoting
 Or even have network connectivity
 Still need to have guest credentials
Network Adapter Identification
 You can name individual network adapters in the virtual
machine settings – and see the same name inside the guest
operating system.
Add-VMNetworkAdapter
PowerShell -VMName
in host:
“TestVM" -SwitchName "Virtual Switch" -Name "Fred" -Passthru |
Set-VMNetworkAdapter -DeviceNaming on
 PowerShell in guest:
Get-NetAdapterAdvancedProperty | ?{$_.DisplayName -eq "Hyper-V Network Adapter Name"} |
select Name, DisplayValue
ReFS Accelerated VHDX Operations
 Taking advantage of an intelligent file system for:
• Instant fixed disk creation
• Instant disk merge operations
Isolation
Distributed Storage QoS


•
•
Windows Server 2016:
Leveraging Scale Out File Server to allow you to:
Define IOPs reserves for important virtual hard disks
Define a IOPs reserve and limit that is shared by a group of
virtual machines / virtual hard disks
Host Resource Protection
 Dynamically identify virtual machines that are not “playing
well” and reduce their resource allocation.
Networking
Software Defined Networking
 Bringing Software Defined Networking to the next level






VXLAN and NVGRE support
Virtual Firewall
Software Load Balancer
Improved Gateways
RDMA using vNICs
And much much more…
One more thing….
Containers
What are Containers
 LXC (Linux Containers) is an operating-system-level
virtualization environment for running multiple
isolated Linux systems (containers) on a single Linux
control host. Containers provide operating systemlevel virtualization through a virtual environment that
has its own process and network space, instead of
creating a full-fledged virtual machine.
Bare-Metal
Virtual Machines
Container Run-time
Container Ecosystem
Container Run-Time
Container Images
Image Repository
Applications
Linux
Application
Frameworks
Containers
Deploying Containers
Requires Hyper-V Hypervisor
Deploying Containers
Deploying Containers
Requires Hyper-V
Hypervisor
Demo: Containers
Software Defined Storage
The (v)Next Level
Current SDS with Server 2012 R2 (1 / 2)
Access point for Hyper-V
Scale-out data access
Data access resiliency
Cluster Shared Volumes
Single consistent namespace
Fast failover
Storage Spaces
Storage pooling
Virtual disks
Data Resiliency
Hardware
- Standard volume hardware
- Fast and efficient networking
- Shared storage enclosures
- SAS SSD
- SAS HDD
Software Defined Storage System
Scale-Out File Server
Shared JBOD Storage
Current SDS with Server 2012 R2 (2 / 2)
Tiered Spaces leverage file system intelligence
File system measures data activity at sub-file granularity
Heat follows files
Admin-controlled file pinning is possible
Hyper-V Nodes
I/O Activity
Accumulates
Heat at Sub-File
Granularity
Data movement
Automated promotion of hot data to SSD tier
Configurable scheduled task
Write-Back Cache (WBC)
Helps smooth effects of write bursts
Uses a small amount of SSD capacity
IO to SSD bypass WBC
Large IO bypass WBC
Complementary
Together, WBC and the SSD tier address data’s short-term and long-term
performance needs
Hot Data
Cold Data
What we hear from
customers…..
I can’t implement a Microsoft Storage Solution because…
“I wanna do HyperConverged”
“Replication on Storage-Level
is missing”
“I don’t trust Microsoft doing
Storage”
Storage Spaces Direct
Storage Spaces Direct at a glance
•
•
•
•
Non-Shared Storage (inbox or local attached)
Distributed via new Software Storage Bus
Enables easy scale out
Leverages all the benefits of SMB3.x / RDMA
Storage Spaces Direct logical View
•
•
•
•
•
•
•
•
•
•
Scenarios
Hyper-converged
Converged (Disaggregated)
Software Storage Bus Cache
•
•
•
Cache scoped to local machine
Read and Write cache (see table below)
Automatic configuration when enabling S2D
•
•
Special partition on each caching device
•
Leaves 32GB for pool and virtual disks metadata
Round robin binding of SSD to HDD
Storage Configuration
Caching devices
Capacity devices
Caching behavior
SATA SSD + SATA HDD
All SATA SSD
All SATA HDD
Read + Write
NVMe SSD + SATA HDD
All NVMe SSD
All SATA HDD
Read + Write
NVMe SSD + SATA SSD
All NVMe SSD
All SATA SSD
Write only
Caching Devices
Capacity Devices
Software Storage Bus
•
Virtual storage bus spanning all nodes
•
•
Clusport: Initiator (virtual HBA)
ClusBlft: Target (virtual disk / enclosures)
•
SMB3/SMB Direct transport
•
Intelligent I/O Management
•
•
•
Prioritizing(App vs System)
De-randomization of random IO
Drives sequential IO pattern on rotational
media
Application
Cluster Shared Volumes File System (CSVFS)
File System
File System
Virtual Disks
Virtual Disks
SpacePort
SpacePort
ClusPort
ClusBflt
Physical Devices
Block over SMB
ClusPort
ClusBflt
Physical Devices
Storage Spaces Direct: Requirements
•
•
•
•
•
4 up to 16 Storage Nodes
10 Gbps RDMA
Min. 2 SSD / 1 NMVe per Node (Cache / Journal)
Min. 4 SSD / HDD per Node (capacity)
Supported HW Model (VMs for LAB)
Storage Spaces Direct Development Partners
Cisco UCS C3160 Rack Server
Dell PowerEdge R730xd
Fujitsu Primergy RX2540 M1
HP Apollo 2000 System
Intel® Server Board S2600WT-Based Systems
Lenovo System x3650 M5
Quanta D51PH
Demo
Storage Spaces Direct
Storage Replica
SR is used for DR preparation
Storage Replica at a glance







Volume based block-level storage replication
synchronous or asynchronous
HW agnostic (any type of source / destination volume)
SMB3 as transport protocol
Leverages RDMA / SMB3 Encryption Multichanneling
I/Os pre-aggregated prior to transfer
Managed via Powershell, cluadmin, ASR
Storage Replica Layer
Source
Destination
SR: Streched Cluster




Single Cluster, spanning DCs
Asymmetric Storage
Automatic Failover
Synchronous only
Datacenter A
Datacenter B
Streched Cluster
SR: Cluster to Cluster
 Multiple Cluster
 Manual Failover
 Synchronous or Asynchronous
Datacenter A
Cluster A
Datacenter B
Cluster B
SR: Server to Server
 To separate Servers
 Manual Failover
 Synchronous or Asynchronous
Datacenter A
Server A
Datacenter B
Server B
Synchronous workflow
Applications
(local or remote)
1
5
2
Source Server
Node (SR)
t
Data
4
2
Log
Destination Server
Node (SR)
t1
Data
3
Log
Asynchronous workflow
Applications
(local or remote)
1
3
4
Source Server
Node (SR)
t
Data
6
2
Log
Destination Server
Node (SR)
t1
Data
5
Log
Storage Replica: Requirements
•
•
•
•
•
•
•
Any volume (SAS, SAN, iSCSI, Local)
<5 ms round trip between sites for synchron mirror
RDMA
Identical size for source / target volume
SSDs for log disks recommended (min. 8GB size)
Identical physical disk geometry (phys. sector size)
Turn on SR write ordering for distributed app data
Demo
Storage Replica
ReFS (Resilient File System)
ReFS - Data Integrity
Metadata Checksums
Storage Spaces 3-way mirror with ReFS
Checksums protect all filesystem metadata
User Data Checksums
Optional checksums protect file data
Checksum Verification
•
•
Occurs on every read of checksum-protected data
During periodic background scrubbing
Healing of Detected Corruption
•
•
•
Healing occurs as soon as corruption is detected
Healthy version retrieved from Spaces’ alternate versions
(i.e. mirrors or parity data), if available
ReFS uses the healthy version to automatically have
Storage Spaces repair the corruption
•
•
•
Checksums verified on reads
On checksum mismatch, mirrors are consulted
Good copies used to heal bad mirror
ReFS - Resiliency and Availability
Availability
•
ReFS designed to stay online and keep your data accessible when all else fails
• Even when corruptions cannot be healed via Storage Spaces’ resiliency (mirror- and paritygenerated data versions)
Online Repair
•
•
CHKDSK-Like repairs performed without taking the volume offline
No downtime due to repair operations!
On-Volume Backups of Critical Metadata
•
•
•
Backups of critical metadata are automatically maintained on the volume
Online repair process consults backups if checksum-based repair fails
Provides additional protection for volume-critical metadata, in addition that provided by mirror
and parity spaces
ReFS - Speed and Efficiency
Efficient VM Checkpoints and Backup
VHD(X) checkpoints cleaned up without physical data copies
• Data migrated between parent and child VHD(X) files as a
ReFS metadata operation
• Reduction of I/O to disk
• Increased speed
• Reduces impact of checkpoint clean-up to foreground
workloads
Hyper-V Checkpoint Merging
Accelerated Fixed VHD(X) Creation
Fixed VHD(X) files zeroed with just a metadata operation
• Impact of zeroing I/O to foreground workloads eliminated
• Decreases VM deployment time
Quick Dynamic VHD(X) Expansion
Dynamic VHD(X) files zeroed with a metadata operation
• Impact of zeroing I/O for expanded regions eliminated
• Reduces latency spike for foreground workloads
Hyper-V VHDX Creation / Extension
Storage QoS
Distributed Storage QoS
Mitigate noisy neighbor issues
Monitor end to end storage performance
Deploy at high density with confidence
Storage Quality of Service (QoS)
Control and monitor storage performance
•
•
•
•
•
•
Rate Limiter
SCALE OUT FILE SERVER
•
•
Policy
Manager
IO Scheduler
Storage QoS Policy Types
Single Instance
Thank you!