z/Linux - Computer Measurement Group

Download Report

Transcript z/Linux - Computer Measurement Group

zLinux Measurements
and Architecture
Clea Zolotow
Senior Technical Staff Member
IBM IT Delivery
(borrowed extensively from the ECM z/Linux Team.)
April 2009
© 2009 IBM Corporation
IBM Virtualization – Enterprise Data Center Journey
Agenda
•
IBM Commitment/Announcement Highlights
―IBM Transformation and “Big Green”
―Internal Infrastructure Challenge, Approach and Benefits
•
IBM Virtualization Update
―Virtualization Progress
―Business Case Approach and Client View of Savings
―Application and Workload Selection
―Successful Techniques and Lessons Learned
© 2009 IBM Corporation
Project ‘Big Green’
Major proof point for Project Big Green
Double compute
capacity with no
increase in consumption
or impact by 2010
To accelerate “green” technologies and services
 IBM will consolidate and virtualize thousands of
servers onto approximately 30 IBM System z™
mainframes
To offer a roadmap for clients to address the IT energy crisis while
leveraging IBM hardware, software, services, research, and financing
teams
 Substantial savings expected in multiple dimensions:
energy, software and system support costs
To create a global “green” team of almost 1,000 energy efficiency
specialists from across IBM
 The consolidated environment will use 80% less
energy and 85% less floor space
IBM to reallocate $1 billion each year
Re-affirming a long standing IBM commitment
Energy conservation efforts from 1990 – 2005 have resulted in a 40%
reduction in CO2 emissions and a quarter billion dollars of energy savings
 This transformation is enabled by the System z
sophisticated virtualization capability
Annually invest $100M in infrastructure to support remanufacturing and
recycling best practices
© 2009 IBM Corporation
IT Organizations are Challenged by Operational Issues
Challenges
Rising costs of systems and networking operations
Costs &
Service
Delivery
Explosion in volume of data and information
Difficulty in deploying new applications and services
Security of your assets & your clients’ information
Business
Resiliency
& Security
Landslide of compliance requirements
Systems and applications need to be available
Rising energy costs & rising energy demand
Energy
Requirements
Power & thermal issues inhibit operations
Environmental compliance & governance mandates
© 2009 IBM Corporation
IBM’s Globally Integrated Enterprise Data Centers
IBM Strategic Delivery Model
Data Center Efficiencies Achieved
 Consolidation of infrastructure, applications
 Enterprise architecture optimization
 Global resource deployment
TECHNOLOGY
IBM Metrics
1997
Today
CIOs
128
1
Host data centers
155
7
Web hosting centers
80
5
Network
31
1
15,000
4,700
Applications
Global
Resources
Strategic
IGA
Location
Strategic
Web
Location
for IGA
Ethernet &
Power9
Networks
Next Level of Infrastructure Challenge




Floor space challenges in key facilities
Underutilized assets in outdated Web infrastructure
Continued infrastructure cost pressure
Increase % IT spending to transformation initiatives
© 2009 IBM Corporation
Stages of Adoption: IBM Journey
Highly responsive and
business goal driven
Rapid deployment of new
infrastructure and services
Drives IT efficiency
Physical consolidation of data
centers, networks and applications
Simple like-for-like server and storage
virtualization
Service tools, energy facilities mgmt
IBM Research “Cloud”
Business-driven service
management pilots
Globally Integrated Enterprise
Significant progress toward highly virtualized
environment to enable pooled System z,
Power Systems, System x and storage
Green production and advanced data center
facilities
Shared service delivery model
© 2009 IBM Corporation
Enterprise Business Value – Expectations
Business
case
Early modeling identified significant potential for savings through zLinux virtualization
Performed TCO virtualization assessment on IBM portfolio as cross-IBM effort
• System z, SW Migration Services, STG Lab Services, IBM Academy, ITO Migration Factory
Identified substantial savings opportunity
• Energy
• Labor
• Floor space
• Software
Annual energy usage to be reduced by 80%
Energy
savings
Quality
service
Total floor space to be reduced by 85%
•
•
11,045 square feet for distributed solution
1,643 square feet for System z solution
Leverages maturity of System z stack products - high availability, resiliency
Reduces complexity and increases stability, centralizes service mgmt
Potential for faster provisioning speed (months  days)
Dynamic allocation of compute power
Provides world-class security
Distributed Solution
Comparison of Annual
Energy Usage for
Workloads
System z Solution
Kilowatt hours (K)
Cost* ($K)
Kilowatt hours (K)
Cost* ($K)
Power
24,000
$2,400
4,796
$479
Cooling**
14,400
$1,440
2,877
$287
Total Energy
38,400
$3,840
7,673
$767
* Electrical cost calculated at rate of .10 per kW
** Cooling is 60% of power cost
© 2009 IBM Corporation
IBM Virtualization – Enterprise Data Center Journey
Agenda
•
IBM Commitment/Announcement Highlights
―IBM Transformation and “Big Green”
―Internal Infrastructure Challenge, Approach and Benefits
•
IBM Virtualization Update
―Virtualization Progress
―Business Case Approach and Client View of Savings
―Application and Workload Selection
―Successful Techniques and Lessons Learned
© 2009 IBM Corporation
IBM System z Linux Virtualization Progress
Established phased approach for quick wins
Migrated initial servers from early ‘wave’ teams
•
•
•
Phase 1
“Quick wins”
Thousands of servers inventoried
Multiple successful migrations delivering benefits as expected
Decommission pipeline of hundreds of servers for reuse or removal
Phase 2
Highest Savings First
Comprehensive project plan and management system in place
•
•
•
Phase 3-n
Integrated business priorities with transformational objectives
‘Work in progress’ approach to maximize server migrations
Pipeline, process, technical, finance and communications support
Finish Virtualization
Developed internal business case (RACE*)
•
Created detailed cash flow and labor analysis, migration expense, iterated
Technical solution, education plan and operational plan developed
•
Built upon IBM prior consolidation/simplification efforts, utilizing IBM offerings and capabilities
Highest level of support from IBM senior executive team
*Formerly zRace
© 2009 IBM Corporation
IBM System z Linux Virtualization Progress
IBM implementing New Enterprise Data Center through achievements in
•
•
Server and storage virtualization
Energy efficiency and resiliency improvements
Benefits are on track with expectations
•
•
•
Cumulative 5-Year Cost
Comparison
Migration management key
Business case is compelling
Using System z10 technology, the number of machines could be cut by
about half, with greater savings in energy, floor space, software and
support costs
Distributed
Cum
z9 Cumulative
Lessons Learned, including:
•
•
•
Enterprise strategy and sponsorship needed to drive business case
and execution
Compelling business imperative accelerates execution and drives support
Enterprise view of migration managed by waves drives experience;
savings for investment
1st
Year
2nd
Year
3rd
Year
4th
Year
5th
Year
IBM experience is driving Time to Value initiatives, integrated into IBM capabilities
•
•
•
Dramatic reduction in labor through new processes supporting workload migrations
Fall in/out analysis, working with business units, to close gaps in workload pipeline
Piloting new testing strategy, processes & tools to automate
© 2009 IBM Corporation
Virtualization Benefits are Significant; Migration Management is Key
Expected Benefits of Virtualization
Substantial savings in multiple dimensions:
energy, software and
system support costs
80% less energy, 85% less floor space
for consolidated environment
Improved inventory hygiene, including application
to server mapping
Large Scale Migration Challenges Exist
Decision-making: Integrating Enterprise and
Business Unit view
Mindset/Culture related to distributed
and mainframe worlds
Workload selection - multidimensional nature of
selection process
Dramatically faster provisioning
Dated inventory records that are not centrally
maintained
Improved security and resiliency
Detailed data required for internal business case
Higher quality through reduced complexity,
increased stability and availability
Project and program complexity – integrating
multiple priorities
Clients are able to leverage IBM experience and
capabilities to accelerate value
© 2009 IBM Corporation
Business Case Leveraged RACE Tool, Iterative Approach
Utilized RACE commercial modeling tool
•
Foundation for internal business case, constructed
specific environmental variables
. Created financial plan for “known universe”
•
Identified relevant sample (5-10%) of most likely servers to be
migrated and gathered financial profile information for each
Engaged SME’s within IBM
•
Provided business case assumptions (i.e.
depreciation/maintenance), modified as appropriate
Iterative Process
•
Continuously engaged with core SME’s to ensure most
current information
Project Metrics
•
Weekly report of migrated servers and their disposition status (reuse
or disposal using GARS*) and Energy Certificate status
-
Working to incorporate actuals into the Business Case such that we
can refresh our assumptions
*IBM Global Asset Recovery Services
© 2009 IBM Corporation
TCO: A Range of IT Cost Factors – Often Not Considered
Availability
•
•
High availability
Hours of operation
Backup / Restore / Site Recovery
•
•
•
•
•
Backup
Disaster Scenario
Restore
Effort for Complete Site Recovery
SAN effort
•
•
•
•
•
Space
Power
Network Infrastructure
Storage Infrastructure
Initial Hardware Costs
Software Costs
Maintenance Costs
•
Investment for one platform – reproduction
for others
Controlling and Accounting
•
•
Analyzing the systems
Cost
Operations Effort
•
•
•
•
Monitoring, Operating
Problem Determination
Server Management Tools
Integrated Server Management – Enterprise
Wide
System Programming
―Keeping consistent OS and SW Level
―Database Effort
•
Middleware
•
Integrated Functionality vs. Functionality
to be implemented (possibly with 3rd
party tools)
•
•
Balanced System
―SW Distribution (across firewall)
•
•
•
•
•
Planned outages
•
Workload Management across physical
borders
•
•
Business continuity
•
•
•
End User Service
Application
―Technology Upgrade
―System Release change without interrupts
Operating Concept
•
•
•
Development of an operating procedure
Feasibility of the developed procedure
Automation
Resource Utilization and Performance
•
•
Mixed Workload / Batch
Resource Sharing
―shared nothing vs. shared everything
•
•
•
•
Integration of / into Standards
Further Availability Aspects
―SW Maintenance
Additional development/implementation
•
Authentication / Authorization
User Administration
Data Security
Server and OS Security
RACF vs. other solutions
Deployment and Support
Infrastructure Cost
•
•
•
•
•
•
•
Integration
Security
Unplanned outages
Automated Take Over
Uninterrupted Take Over (especially for
DB)
Availability effects for other applications
/ projects
End User Productivity
Virtualization
Skills and Resources
•
•
Personnel Education
Availability of Resources
Parallel Sysplex vs. Other Concepts
Response Time
Performance Management
Peak handling / scalability
Routinely Assessed
Cost Factors
© 2009 IBM Corporation
Client View of TCO Comparison for Similar Distributed Workload vs. System z
Linux results in Potential 60-75% Gross Costs Savings / 5 yrs
Operating Cost: Distributed vs. Mainframe
Dramatic Simplification
Relative Cost
Labor
Connectivity
Software
Distributed
System z
Linux
%
Reduction
Software
Licenses
26,700
1,800
93%
Ports
31,300
960
97%
Cables
19,500
700
96%
15,700
7,000
55%
Unit
HW Maintenance
HW Acquisition
Facilities
Distributed Cost
System z Linux Cost
Potential Savings: Categories as a
% of Gross Savings
HW Acquisition*
17%
HW Maintenance
11%
Facilities
6%
Labor
19%
Software
45%
Connectivity
2%
* HW Acquisition compares server/disk refresh of distributed environment to
the cost of acquiring new mainframes/storage
Physical
Network
Connections
Results will vary based on several factors including # of servers and work load types
© 2009 IBM Corporation
Diagnose
Energy Efficiency Certificates Deliver Savings
Build
By formally decommissioning servers, IBM is able to demonstrate
energy savings and receive energy efficiency credits (EECs)
Manage &
Green Measure
Data
Center
Virtualize
Cool
Client requirements
Lower energy costs and achieve business benefit of Energy
Efficiency
Demonstrate Energy Efficiency Commitment
Solution
Virtualized workloads onto System z platform and reduced energy
consumption
Hundreds of servers in pipeline to be redeployed, sent to GARS*
and/or energy efficiency certificates issued
IBM applied for EECs for eligible decommissioned servers to receive
Energy Efficiency Credits
What is an Energy Efficiency Credit?
A Neuwing EEC (Energy Efficiency Credit) is a
measured & verified Megawatt Hour (MWh) of
Energy Savings i.e., Energy Efficiency
EECs quantify, measure,
verify, certify and monetize
data center energy efficiency
projects
GARS for asset reuse, recycling and/or reclamation
November 2, 2007 Press Release
Benefits
Quantifable energy reductions, tradable certificates
Demonstrated commitment to energy efficiency
IBM Launches World's First Corporate-Led Energy
Efficiency Certificate Program
In Conjunction With Neuwing Energy, Program Will Provide
Clients Documentation and Third-Party Verification of the
Energy Saving Results of Their Projects. Read more
*IBM Global Asset Recovery Services
© 2009 IBM Corporation
IBM is Using a ‘Work in Process' Approach to Manage
the Migration
Management Approach and Reporting
 Process approach borrowed from
factory line management
 Metrics for each process and
sub-process
 Quality measured with process
fallout – tracked by cause
 Daily status calls for issue
resolution
 Weekly status reporting for CIO
and management team
© 2009 IBM Corporation
Decommission Process Overview
Server available
as a result of
virtualization
efforts
Server
Ready
Check for technical viability and asset
value to determine if h/w is a
redeployment candidate
If redeployed
Request completed to
coordinate shipping
and update property
control
If not redeployed
Complete Machine
List Database and
ship to GARS*
Apply to Neuwing for
energy efficiency
certificates
Tracking tool is updated to reflect
disposition of the assets in the project
Capture savings in business plan and
business case
*IBM Global Asset Recovery Services
© 2009 IBM Corporation
Enterprise Approach to Workload Migration
Environment View
Location View
Managed ‘Offerings’
Development
Intranet
BU Environments
Boulder
Poughkeepsie
Portsmouth
Raleigh
Rochester
Southbury
Migration
Candidates
Sourcing
Application View
Business Unit Partnership
Technology View
Domino
Email
Static Web
DB2
Linux on x86
© 2009 IBM Corporation
Each Workload is Evaluated for Suitability Based on Technical
Attributes
Priority Workloads for
Consolidation:
WebSphere® applications
Domino® Applications
Tivoli®,
A
IX
e le
m
ks
b y
a
a
a
l
l
r vai on pe eeds
f
in re a me CPU y n
a
M
r
Proximity
a ra
d
tw inf ine mo
to other
f
a
a me
t
So m
data on
us ge
on
s
mainframe
a
er ver
Other
w
a
o
components
L nd
of app on
a
Selected tools:
WebSphere® and internally
developed
mainframe
High and/or
transactional
I/O
WebSphere MQ
S
Hi
®
on of
gh
t
w
an s
AI a
d us
hi ta X® re a
/U v
gh in
®
m ed NIX aila
em C
® bl
o e
or PU
Already
y p nly
ne e a
easily
ed ks
virtualized on
s
/U
N
IX
AIX®/UNIX®
Test in
support of
prod on
Default
is chosen
when no other
criteria
apply
apply
AIX®/UNIX®
Grey area
(not clearly one platform or another)
DB2® Universal Database™
Al
l
Already
virtualized
on Intel®
Li
n
Small
L
in ux ®
So
counts of
u
f
x ne
isolated
o n ed
on twa
images
m sn
In re a
ai o
te v
nf t m
l® ail
on ab ram et
ly le
e by
®
In
te
l
Technical Attributes
 Software compatibility or
platform dependency
ts
en
ith le
 Workload characteristics
m
w ilab m
e
ir
ds va te
qu
oa e a / sys
l
Decision
Not fully
virtualized
e
r
k r
criteria/thresholds
or wa que
e
u
t i
W
optimized
iq
of un
can be tailored
bys the
client
n
U
n
o nly
o
© 2009 IBM Corporation
Blue Harmony is an enterprise transformation program that
will enable IBM’s Globally Integrated Enterprise (GIE) vision
Blue Harmony will enable the GIE by …
•
Integrating core processes that are now business unit and regionspecific, such as Quote to Cash, horizontally across the enterprise
on a common systems platform – SAP
•
More effectively integrating the global processes that are already
in place
―For example, Blue Harmony will integrate financial reporting with
business reporting, providing real time access to consolidated
data
The significance of the Blue Harmony name
•
The name reflects how we are harmonizing process, technology
and people across the IBM enterprise
© 2009 IBM Corporation
Blue Harmony: Main Hardware Components
Xseries
z10
p595
p595
z10
2498-M48
pair of 2027-140
Ficon directors
2498-M48
2109-m48
2109-m48
DS8300B
DS8300A
Metro mirror
Tape
Tape
dcx-4s
channel extenders
z/OS Global Mirror
subnet
Cisco 3750E-48T
CD 6504
Poughkeepsie, N.Y.
ATTWAN
1200 GSR
oc48
oc48
Portsmouth, U.K.
© 2009 IBM Corporation
Blue Harmony Mainframe Configuration
Hardware – z/10 - 56-way – 7,674 MIPS on general CPUs
• 11 General Purpose CPUs (CPs) for z/OS
• 6 Internal Coupling Facilities (ICFs) for sysplex sharing
• 6 Integrated Information Processor (zIIPs) for DB2
• 17 Integrated Facility for Linux (IFLs)
Connectivity
• 10 OSA 1000 Base T (4 ports)
• 10 OSA 10GbE (2 ports)
• 2 ESCON (16 port)
• 7 ESCON Channel Ports
• 2 FICON Express4 (SX, 4 ports)
• 36 FICON Express4 (4KM, LX, 4 ports)
• 8 ICB4 (data rate of 2GBps)
© 2009 IBM Corporation
Blue Harmony US Implementation
FIRST Z/10
SECOND Z/10
BlueZone
© 2009 IBM Corporation
What is z/Linux at IBM
What is z/Linux
•
•
Provisioned underneath a hypervisor
Provisioned barebones (not supported
within my current architecture)
Currently supported are:
z/ECM: The Enterprise Computing Model
(ECM) is an IBM initiative to virtualize 25
percent of our own distributed IT infrastructure
onto System z mainframes. ECM is also an
IBM showcase for Project Big Green and the
New Enterprise Data Center.
Blue Harmony: IBM's major transformation
program to radically simplify the design and
operation of three Globally Integrated Support
Processes - Finance, Opportunity To Order,
and Order To Cash
24
IBM Confidential
© 2009 IBM Corporation
What is z/Linux the OS
z/VM is an operating system that runs on the z/Series
mainframes. It functions both as a fully functional
operating system as well as as hypervisor, much like
VMWARE.
z/VM builds on more than three decades of
virtualization experience. z/VM’s CP (Control Program)
is a hypervisor which has been built with isolation in
mind and extended this to in-memory network
virtualization.
The System z based workload and as such Linux on z
based services are ideal for low utilized services which
exploit the fast context switching capabilities provided
by the hardware, the z/VM hypervisor and the many
hardware assist functions.
z/VM allows the virtualization of memory, CPU cycles,
storage, consoles, and even can emulate non-existing
hardware like for example a virtual zAAP processor to
speed up JAVA workload.
The IO subsystem provides many advantages to high
volume IO workloads, and for LINUX on z SCSI
devices can be assigned as well as the mainframe
typical CKD devices.
From ACS, “Mainframe on the Cloud”
25
© 2009 IBM Corporation
IBM’s Enterprise Architecture for ECM: Security Zones
Possible Architecture Path for Single LPAR with Multiple Security Zones and Multiple
Clients is represents a possible solution to have multiple security zones and multiple
customers (MSZ) on the same LPAR.
This would be used for supporting small and medium business and Clients who deploy
multi-network-zone tiered architectures where firewall separation of zones happens
externally.
This architecture is currently in place today supporting IBM internal businesses.
© 2009 IBM Corporation
Enterprise Architecture
VLzH Operational Model- Conceptual
Internet
Proxy
Load Balancer
VPN
Firewalls
Security zone 1
tion
Vlan
l i za
Build
Server
Build
Server
)
k Vi
Tivoli
Gateway
LPAR
Net
wor
BuildServer
Server
Build
Z System
rtua
Security zone 2
Vlan
backup
SANs
LPAR
LPAR
Security Zone 3
VLAN
Dynamic
Resource
Prov
Load Balancer
Single LPAR on z
Provisioning
Server
BuildServer
Server
Build
Tivoli
Gateway
backup
Proxy
Possible design path for using a single LPAR with multiple
security zones or multiple Clients. This involves network
design and security approval
Operations team
Tools Network
Internal Network
System
Administrators
IIP
27
SRM
GCMS
Other Config
Mgt
TLM?
eESM
Tivoli TEC&
TMR
Hardware
Asset Mgt
Security
Management
© 2009 IBM Corporation
How to manage and access the Linux guests
The four major components for managing and access the Linux guests are:
Hardware Monitor Console (HMC)
Integrated Communications Console (ICC)
VSWITCH
z/VM Management
The z/VM management for the LPARS can share the same physical ports while
the VSWITCH(s) have dedicated ports.
http://news.cnet.com/8301-19413_3-10237274-240.html
© 2009 IBM Corporation
How to manage and access the Linux guests: HMC
Hardware Monitor Console (HMC)
The HMC is the lowest level of management connectivity of a physical zSeries
processor. Through the HMC hardware components are mapped and
assigned to various parts of the box and defines the bounds of the different
logical partitions inside the box. Operations use the HMC to implement the
configuration of the box and IPL LPARs.
The HMC is physically connected to the processor by two Service Element
ports.
These HMCs provide the interface to configure the processor. An administrator
can physically access the HMC to configure the box or remotely connect to
the HMC via an internal web service.
Hardware Master Console
(Linux or OS/2)
Rest of Network
Private Network
Service
Element
Service
Element
© 2009 IBM Corporation
How to manage and access the Linux guests: ICC
Integrated Communications Console (ICC)
Operator management of the zSeries processor can also
be made through the Integrated Communications
Controller (ICC). The two ICC_OSA ports are
connected to two different LAN segments for
redundancy in case there are network issues
accessing one of the ports. The port has an IP
address that is routed in the network for AF/Remote
Server connectivity can be established. OSA-ICC’s
are shared between LPARS.
The AF/Remote Server communicates to the ICC_OSA
via TN3270 traffic. Administrators utilize a software
client to connect to the AF Remote Server which
then downloads a list of LPARs the administrator is
able to control. When the administrator opens a
connection to the LPAR, the connection is proxied
through the AF Remote Server before being passed
to the processor.
AF Remote Server
SA IOM
(backup)
AF Remote Server
SA IOM
(Windows Server)
Access to ICC
Copper Only
OSA pair
TN3270 traffic
ICC_OSA
ICC_OSA
The client on the administrator’s workstation
communicates to the AF Remote Server over port
1035 (default). From there, the AF Remote Server
communicates with the ICC_OSA over port 1024
(default).
© 2009 IBM Corporation
How to manage and access the Linux guests: VSWITCH
The four major components for managing and access the Linux guests are:
VSWITCH
The z/VM management for the LPARS can share the same physical ports while
the VSWITCH(s) have dedicated ports. External connectivity to a network from
inside an LPAR can be established via a Layer 2 VLAN aware VSWITCH.
The OSA ports plug directly into a Layer 2 switched distribution box. This
utilizes the VSWITCH as the access switch in a multi-tiered architecture. The
deployments in questions however vary on how the VLANs are deployed to
the given LPARs.
Service Core
Si
Si
SD Block
Si
Si
Backup OSA
HSRP Active
zSeries
Processor
Port 1
OSA
VSwitch
Port 50
VM
Controller
Linux
Port 2
Port 51
Linux
OSA
OSA
Port 1
Port 52
Linux
VSwitch
Port 50
VM
Controller
zSeries
Processor
OSA
Linux
Port 2
Port 51
Linux
Port 1
Port 52
Linux
OSA
VSwitch
Port 50
VM
Controller
Linux
Port 2
Port 51
Linux
OSA
OSA
Port 1
Port 52
Linux
VSwitch
Port 50
VM
Controller
zSeries
Processor
OSA
Linux
Port 2
Port 51
Linux
Port 1
Port 52
Linux
OSA
VSwitch
Port 50
VM
Controller
Linux
Port 2
Port 51
Linux
OSA
OSA
Port 1
Port 52
Linux
OSA
VSwitch
Port 50
VM
Controller
Linux
zVM
zVM
zVM
zVM
zVM
zVM
LPAR
LPAR
LPAR
LPAR
LPAR
LPAR
Port 2
Port 51
Linux
Port 52
Linux
© 2009 IBM Corporation
How to manage and access the Linux guests: z/VM TCPIP
The four major components for managing and access the Linux guests are:
z/VM Management TCPIP
Once an LPAR is created on the processor, with a TCP/IP Stack for CP/CMS access, an
administrator can logon as a controller session to handle the creation and management
of Linux guests inside the LPAR. The controller only manages inside the LPAR and is
separated from the production traffic of the Linux Guests. The TCP/IP stack will have a
VIPA IP address with two OSA interfaces for redundancy.
The first option shows OSAs pairs that are used for shared management only, across all
z/VM instances that exist on a single processor. It is defined by sharing the OSA across
all LPARs which is separate from production traffic. Access to these management OSAs
need connectivity to the network segments where administrators sit.
OSA
Port 1
OSA
OSA
VSwitch Port 2
Port 1
Port 49 Port 50 Port 51 Port 52
VM
Linux
Controller
zVM
Linux
Linux
Group 1
OSA
OSA
VSwitch Port 2
Port 1
Port 49 Port 50 Port 51 Port 52
VM
Linux
Controller
zVM
LPAR
Linux
Group 2
LPAR
OSA
Linux
OSA
VSwitch Port 2
Port 49 Port 50 Port 51 Port 52
VM
Linux
Controller
zVM
Linux
Linux
Group 3
LPAR
OSA
zSeries
Processor
© 2009 IBM Corporation
Performance and Capacity Tooling
To the right, you can see the overlap of
capacity planning and performance.
VMPTK does provide several linux level
reports lxmemory and lxcpu logs that show
memory and cpu usage at the linux level.
(VM- Performance Tool Kit).
WAS
SRM (example)
Smoothstart
Software Requirements
LINUX6
SLES10 64bit
LINUX5
PTK
Omegamon XE
ELAPSE
DB2
LINUX4
Omegamon XE
SAR
vmstat
ps
Nmon
ITM
Other/Application
Specific
Application
LINUX3
Work is ongoing to load the VMPTK
hypervisor data into SRM to correct the
CPU utilization data.
Other/Application
Specific
Capacity
LINUX2
SRM data utilizes standard SUSE10
monitors which are used illegally (as the
CPU time is incorrect due to the
virtualization). However, midrange people
prefer the interface.
Performance
LINUX1
Note that both team rely on the VMACCT
data. Although Omegamon has an agent, it
is not providing good response time in our
environment.
zRACE 1080
z/VM Capacity Planner
SCPON (internal)
LINUX1
Z9 EC
S38/512GB
Reduce Agents on Guest
Each agent consume 2% MIPS
© 2009 IBM Corporation
Performance and Capacity: Separation of duties
Note the different responsibilities
here between Capacity, z/VM,
Performance, and z/Linux.
Responsibilities
Application Developers
Application
WAS
DB2
Middleware Competency
SLES10 64bit
zLinux Performance
LINUX1
LINUX6
LINUX5
LINUX4
LINUX3
LINUX2
z/VM Performance
LINUX1
The interesting thing here was that
z/VM retained their z/VM skills and
z/Linux moved to folks who had
previous mid-range UNIX skills.
Capacity Planning
LINUX2
ZOS1
Z9 EC
S38/512GB
© 2009 IBM Corporation
Tooling for Alerting
Z/VM LPAR Native Console and
Hardware Alerting
HMF, WWAO
Native Console alerts can be filtered and acted upon by HMF
and ultimately focal pointed to another Z/OS system via PSM
with WWAO.
Z/VM LPAR CPU
PTK and Omegamon XE
Performance Toolkit and/or Omegamon XE provide the ability
to proactively monitor CPU utilization.
Linux Guest Linux Middleware - DB2 / ITM for Databases
"ITM for xxxx" are monitoring extensions whch are used for
MQ Series
*Omegamon XE for Messaging (ITM
either standard middleware like Lotus DB2, MQ, Oracle - or
6.1)
industry wide applications like Notes, Siebel, SAP. A monitoring
extension will contain both resource monitors checking for
something specific (i.e a DB2 table space becoming full) or
checking logs (i.e the MQ Series error log)
Linux Guest OS - Linux
IBM Tivoli Monitoring Base - OS
This is primarily handled through some form of OS Agent
Agent ( Disk, CPU, OS Processes)
monitoring. The key advantage is that no matter how many
*Logfile adapters / Log Monitors for OS occurrences of a problem you get, there will be one alert
Logs
generated until the problem is resolved. A logfile adapter can
also be utilized as an adhoc monitoring aid where NO existing
agent can satisfy the requirement.
© 2009 IBM Corporation
ITM6 Tooling Integration
IBM Locations
(IBM Intranet / SNI Network)
Customer Locations
(Customer Premise/Extended
Premise)
Service Management
Operations
w
Vie
nt
itorin
g
po
ta
rts
Da
Mo n
rtal Da
ta
g u re
e
Ev
Confi
View P
o
ta
ta
Da
Da
l
rta
Po
l
rta
Po
View Portal Data
w
Vie
w
Vie
Future Integration
System Admins
Re
Operations
Intranet User
Systems
Management
Infrastructure
Vi
ew
Internet User
Fo
c
al
zOS Focal
Update Portal Status
T/EC
Receive Event / Perf Data
on
s
z/VM LPAR
Situation Data
Problem Management
Ag
e
nt
Websphere,
Exchange
Agent Configuration,
Policy & Situations
ati
&S
itu
s
cy
cy
on
ta
oli
,P
ati
,P
oli
gu
ion
ra t
itu
&S
ra t
ion
nfi
ta
gu
Co
Da
nfi
Ag
t
en
on
Da
DB, SQL, Oracle
oli
n, P
ns
Co
ti
ua
Sit
atio
atio
Situ
cy &
at
ion
tC
ur
onfig
ta
Si
tu
n
Age
n Da
Po
in
Omegamon /
ITM 6
Infrastructure
TDW
tio
Situa
tin
g
Long Term Reporting
Middleware and
Applications
Server
(SuSE, Redhat, etc.)
Linux Customer
Systems
z/VM
z/VM
(VM Lpars)
Mainframe Customer Systems
© 2009 IBM Corporation
Tooling: Tivoli
•
The Tivoli Enterprise Management Server (TEMS) can be seen as the central repository for all monitoring data
and resources. In particularly it stores the definitions for conditions that indicate a problem with a particular
resource, these are known as situations. It also controls the security for the monitoring solution. Each enterprise
monitoring solution must contain one Hub TEMS, and can include multiple Remote TEMS. Unlike previous Tivoli
Monitoring products where the lower levels of the Architecture (Endpoint or Gateway) interface directly with
event management, it’s the HUB TEMS that controls forwarding of Events to the T/EC Component.
•
The Tivoli Enterprise Portal Server (TEPS) functions as a graphical access server that displays near time and
real time status indications on the health of resources in the enterprise. It also serves as a repository for all user
data, such as the user IDs, user access control for the monitoring data, meaning what data each user will be
able to access, and how it is displayed. It connects to the Hub TEMS and can be accessed by the TEP clients
who are either desktop clients or browser clients. Data is only available in the TEPS for a maximum of 24 hours,
and the component should not be considered a historical reporting capability. The TEPS Database is also used
to store internal data used by the TEPS (TEP user IDs, queries, and navigator views).
•
Tivoli Enterprise Console Server provides a centralized location for the management of events. The server
creates an entry in the event database for each incoming event and then evaluates these events against a set of
rules to determine if the event server should automatically perform any predefined tasks or modify the event. If
human intervention is required, the event server notifies the appropriate operator. The operator performs the
required tasks and then notifies the event server when the condition that caused the event has been resolved. In
addition, there is the option of enabling a form of automated trouble ticketing to a desired Problem Management
system.
•
Tivoli Enterprise Monitoring Agents (TEMA) are the data collectors for the ITM monitoring solution. They are
installed to gather data from one or multiple systems that are critical to the client business.
•
Data that is collected by the agents is sent to the central Tivoli Enterprise Monitoring Server. Agents are typespecific (I.e. Unix, Linux, SAP, DB2, Omegamon XE on z/VM and Linux). The Omegamon XE on z/VM and
Linux agent will be an integral part of this solution and helps address performance bottlenecks on z/VM
resources. The three main areas where monitoring can be beneficial are:
• Monitoring processor usage at the system level
• Monitoring paging and spooling activity
• Monitoring resources and workloads across LPARs
© 2009 IBM Corporation
Monitoring Diagram for Existing ECM Farm
Mainframe Operator
Workstation
ECMz Monitoring
Architectural
Overview
Diagram
Existing Auto Ticketing
Distributed Operator
Workstation
Existing Auto Ticketing
Problem Ticketing system
Existing Auto Ticketing
Existing Tivoli Enterprise
Console
Tivoli Enterprise Portal
(TEC)
Server
(TEPS)
Existing WWAO z/OS Mainframe
Focal Point
Tivoli Enterprise
Management Server
(TEMS)
Tivoli Enterprise
Monitoring Agent
(TEMA)
Existing z/os Mainframe
Existing z/VM Mainframe
New ECMz Mainframe
New ECMz Mainframe
Performance Monitoring data
Existing z/os Mainframe
Existing Problem Management
Tivoli Enterprise
Monitoring Agent
(TEMA)
Existing Customer
Servers
Existing Distributed Operator alerting and command/response
Existing Mainframe Opertor alerting and command/response
Version 2.0 26/11/07 Richard Short
Version 3.0 March 10, 2008 Gene Capano
Linux Guest Alerting
z/VM Alerting
© 2009 IBM Corporation
Tivoli Monitoring Flow
Key:
TEP Browser or Desktop
Monitoring
Infrastructure
TEP – Tivoli Enterprise Portal
TEPS – Tivoli Enterprise Portal Server
TDW – Tivoli Data Warehouse
TEMS – Tivoli Enterprise Monitoring Server
TEMA – Tivoli Enterprise Monitoring Agent
TEPS DB
TEPS
Warehouse
TEMS Hub
TEMS Data (Built in operational DB)
TEMS Data
Remote TEMS
Local Cache
Local Cache
Agents
Agents
© 2009 IBM Corporation
z/VM and z/Linux Monitoring Flow
Here is the monitoring
flow from the VMPTK
into and the previous
Omegamon XE solution
(now lxmemory and
lxcpu from VMPTK).
MONWRITE Utility
History
Raw
Files
Monwrite
3270
Performance
Our data ends up in a
DB2 DB on z/OS.
TCP/IP
Network
Toolkit
Not much has really
changed in 20 years!
MONDCSS
Segment
VMRM
DCSS
SEGOUT
Browser
HTTP
TEMS
VM
IRA
*MONITOR System Service
LINUX
Guest
VM Events
OMEGAMON XE
CP Control
Application
Blocks
Data
© 2009 IBM Corporation
How Tivoli interacts with the z/Linux Guests
• This is a logical
diagram, not a
physical one.
TEP
• Note that all data
flows from the
z/VM host.
• Internal monitors
are supported by
SRM, but not by
Tivoli.
TEPS
Tivoli
Management
Services
z/VM command
and response
flows
TEMS
Data and synchronization flows
VM Linux
IRA IRA
Extensions
CMD
(Guest)
TDW
Mon
PTK
DCSS
Linux
(Guest)
Linux
IRA
Linux
(Guest)
Linux
IRA
Linux
(Guest)
CP
z/VM
© 2009 IBM Corporation
Autoprovisioning using IBM Director
Strategic Solution
Tactical Approach
TPM Server
1
Cloning/Breeder Solution provide
subset of the strategic solution
NIC
1
TPM for Software
Fusion
OSA
IBM Director
Administrator
Console
z/VM Center
Console extension
Linux SuSE
CIMServer
NIC
3
MAPSERVE
NIC
1
LAN
MAPLAN
Remote Access to IBM
Director and other
Consoles
(via Remotely Anywhere)
NIC
1
Director Server
2
SBLIM Client
OSA
Cloning Solution
(Canada)
(IGA + Commercial)
VSWITCH
KickStart Server
2
DIRMAINT
NIC
z/VM Center Server
extension
OS Images
Software
Packages
NIC
1
4
1
IBM Director Mgmt
Server
NIC
2
z/VM MAP
Director Console
Global Delivery
GlobalCenters
Delivery
GlobalCenters
Delivery
Centers
Linux SuSE
VMRMSVM
NIC
1
NIC
1
Linux
Linux
Linux
RH
RH
RH
Linux
Linux
Linux
S10
S10
S10
Breeder Solution
(EMEA)
(IGA + Commercial)
VSMSERV
TCPIP
PORTMAP
CIMServer
NIC
z/VM
IBM Director SWD
Virtual Machine
Manager
z/VM Mgmt Access
Point Agent
extension
© 2009 IBM Corporation
Autoprovisioning: The way it really works (cloning)
© 2009 IBM Corporation
Autoprovisioning: Why
• z/VM Center manifests itself inside the z/VM hypervisor as a zLinux service
machine, making available the CIM Instrumentation (CIM is also known as
Pegasus, and stands for Common Information Model) for z/VM, and creates
the API by means of the z/VM MAP function. As seen above, both, IBM
DIRECTOR Virtualization Manager and the IBM TIVOLI Provisioning Manager
exploit the same API and CIM as both products have been designed to be
platform (Operating System) agnostic.
• The flexibility by separating the GUI from the automation and the hardware
line it runs on makes it also inflexible for zLinux deployments. So far it is not
clear how much it can be tweaked to fulfill the above mentioned target
architecture requirements. Testing needs to be performed and with a focus
on exploiting Exits rather than changing the product.
© 2009 IBM Corporation
Backup Methodology.
zLinux Operating System (OS) image and essential
filesystems and data to the zLinux OS will be maintained
on FICON/ESCON attached disk. OS data and images are
not backed up using the same technology as Application
Data. The primary technology deployed for backup of
Application Data on zLinux guests is the IBM Tivoli
Storage Manager (TSM) using Automated Tape Libraries
(ATLs) or newer Virtal Tape Libraries (VTSs) to store
backup data.
45
IBM Confidential
© 2009 IBM Corporation
Backup and Recovery Model
Storage Services for ECM/z:
Architecture Overview Diagram
Security Zone 3
Security Zone 2
Security Zone 1
FICO
N
MSS YZ SAs &
SAN Designers
ECM System Z
Servers
TSM Backup
Servers
FCP
FC-SAN
FCP
Security Zone 2
Security
Zone 3
LAN (DMZ)
DS8300
Unix/AIX
Servers
LAN
LTO drives in ATLs
NETWORK FIREWALL
Security Zone 4
Security
Zone 4
FC
P
LAN
FCP
Unix/AIX
Servers
FCP
P
FC
P
FC
TSM Backup
Servers
NETWORK FIREWALL
Security Zone 1
(Internet)
359x Drives in VTSs
FCP
ECM System Z
Servers
Unix/AIX
Servers
ECM/z RZ (Internet)
Application User (AU)
FI
CO
N
P
FC
ECM/z RZ
AU
NETWORK FIREWALL
FCP
FC
P
ECM/z BZ AU
P
FC
LTO drives
in ATLs
FC-SAN
TSM Backup
Servers FCP
FCP
ON
FIC
DS8300
LTO drives in ATLs
FICON
Datacentre Perimeter
ECM System Z
Servers
359x Drives in VTSs
ECM/z BZ
SAs
MSmith IBM GTS Stor Comp 2007-11-30
46
© 2009 IBM Corporation
Backup
47
© 2009 IBM Corporation
Defining cloud computing.
Cloud computing is a new consumption
and delivery model inspired by consumer
Internet services. Cloud computing exhibits
the following 5 key characteristics:
•On-demand self-service
•Ubiquitous network access
•Location independent resource pooling
•Rapid elasticity
•Pay per use
48
IBM Confidential
© 2009 IBM Corporation
What does it really mean?

On-demand self-service. A consumer can unilaterally provision computing capabilities, such as server
time and network storage, as needed without requiring human interaction with each service's provider.

Ubiquitous network access. Capabilities are available over the network and accessed through standard
mechanisms that promote use by heterogeneous thin or thick client platforms (e.g., mobile phones,
laptops, and PDAs).

Location independent resource pooling. The provider's computing resources are pooled to serve all
consumers using a multi-tenant model, with different physical and virtual resources dynamically
assigned and reassigned according to consumer demand. The customer generally has no control or
knowledge over the exact location of the provided resources. Examples of resources include storage,
processing, memory, network bandwidth, and virtual machines.

Rapid elasticity. Capabilities can be rapidly and elastically provisioned to quickly scale up and rapidly
released to quickly scale down. To the consumer, the capabilities available for rent often appear to be
infinite and can be purchased in any quantity at any time.

Pay per use. Capabilities are charged using a metered, fee-for-service, or advertising based billing
model to promote optimization of resource use. Examples are measuring the storage, bandwidth, and
computing resources consumed and charging for the number of active user accounts per month.
Clouds within an organization accrue cost between business units and may or may not use actual
currency.
http://news.cnet.com/8301-19413_3-10237274-240.html
© 2009 IBM Corporation
Pay no attention to the man (person) behind the curtain!
© 2009 IBM Corporation
How does this apply to Z/Linux?
 Provide on demand service levels.
 Fast service provisioning.
 Capacity Planning
 SO must prepare to react on IBM’s marketing strategies (Cloud and
Virtualization)
 Cost:




Datacenter
Hardware
Management
Hypervisor Management
Author: Gerhard Widmayer
© 2009 IBM Corporation
How does this apply to Z/Linux? (Service)
 Provide on demand service levels.
On demand service levels are for example “zero planned outage”, “scale-up”
and “scale-out”, or “pay per usage”. The service levels can only be provided
in a strict virtualized server resource pools infrastructure. SMB customers can
not afford the hardware price for a cluster which they would utilize only at a
small percentage.
Author: Gerhard Widmayer
© 2009 IBM Corporation
How does this apply to Z/Linux? (Provisioning using the free
pool)
 Fast service provisioning

Server ordering and infrastructure adjustments is making provisioning a week
long activity at best. This is also true because SO must create a new design
for each service request rather than establishing a service inside a predefined
infrastructure unless there is an enterprise architecture, such as ODCS.
Competition is delivering on demand basic server services within 24 hours
after accepting the request.

Using a virtualized free pool there will always be some resource available in
the headroom in an appropriate resource pool (selected target location) if it is
built to be multi-tenant enabled. Any service establishment is just a set of
commands away, true for servers, storage, but also the real network.
Increase of the resource server pool is only needed when thresholds are
reached.

Patent 7685283 shows how to size for the free pool, “METHOD FOR
MODELING ON DEMAND FREE POOL OF RESOURCES”
Author: Gerhard Widmayer
with adds by Clea
© 2009 IBM Corporation
How does this apply to Z/Linux? (Capacity Planning)
 Capacity Planning –

Behind the curtain, define servers by actual consumption rather than by
peak workload plus uplift.

Today’s distributed servers face the problem of hardware underutilization.
Normal utilization values seen are 8% average in the distributed world.

Currently, the customer pays what was ordered regardless of usage. At the
ODCS, there was an LPAR charge and a utilization charge.

In order to establish “on demand” pricing there is a need to put multiple
services onto a single box, exploiting virtualization, like the ODCS. With the
ever increasing compute power year by year the “boxes” became that big
that utilizing single boxes by single network zones or SMB customers is
very unlikely.

The ability to do coadunation is the ability to put each virtualized LPAR in
it’s mathematically correct spot.
Author: Gerhard Widmayer
with adds by Clea
© 2009 IBM Corporation
Coadunation Example
Mixing workload shares headroom but you pay in response time at low
utilization....workload management shifts peaks based on business priorities to use
"white space" but response time of lower priority work is traded off...
© 2009 IBM Corporation
How does this apply to Z/Linux? (Moolah)
Datacenter Cost



With the Green IT there is a drive to move
towards larger boxes for network, storage,
and servers in order to optimize the
capacity per watt. For optimization of
environmental savings (space, power,
cooling) these large boxes need to perform
at a high utilization level, which requires a
very flexible setup with many isolated
network zones on it.
Migrating pSeries hosts into z/Linux is
proving to be quite cost-effective. The ECM
consolidated environment will use 80%
less energy and 85% less floor space.
Specific energy savings follows.
Hardware Cost

Potential Savings: Categories as a
% of Gross Savings
HW Acquisition*
17%
HW Maintenance
11%
Facilities
6%
Labor
19%
Software
45%
Connectivity
2%
HW acquisition and maintenance cost goes
down from a distributed environment to a
z/Series environment. Further, SPOF
failover is cheaper as well in the mainframe
world.
Author: Gerhard Widmayer
with adds by Clea
© 2009 IBM Corporation
Ubiquitous network access: Based on the vSWITCH
Author: Gerhard Widmayer
with adds by Clea
© 2009 IBM Corporation
VSWITCHES and OSAs
From a security risk mitigation
perspective a distinct VSWITCH and
dedicated physical OSA ports and real
switches is worth a consideration.
However the basic design should
verify that the VSWITCH and VLAN isolation
if bundled together with customer and
application traffic is secure. In such a case
the (S)CF firewall becomes a PF (public
firewall), and the 3/4-NIC design would be
preserved.
Author: Gerhard Widmayer
with adds by Clea
© 2009 IBM Corporation
VSwitch, Continued
VSWITCH is a z/VM established virtualization of a real
switch. In this deployment guidance the VSWITCH is
replacing the real access layer switch from an Ethernet
hierarchy. For each of the three (or four) NIC’s in the NIC
virtual server design a distinct VSWITCH is built. Each
customer receives its own VLAN number, and each network
zone receives also its own VLAN number. This VLAN
assignment is true for all three or four NICs, so in effect
always a triple or quad of VLAN numbers are assigned to
each single network zone.
All VLAN’s of a type, like all Extended Premises, are
combined in a single VSWITCH. Because the design is for
SMB environments, this means that multiple customers
and/or multiple network zones share a VSWITCH. The
VSWITCH function must therefore provide the VLAN
isolation and must disallow any VLAN cross-over
interaction.
Remember that the basic design keeps the security control
for allowed interaction flows in the firewalls outside the
System z box, where it exists today.
Author: Gerhard Widmayer
with adds by Clea
© 2009 IBM Corporation
How does this apply to Z/OS?
This page intentionally left blank.
© 2009 IBM Corporation
Questions?
© 2009 IBM Corporation