XP 102 – MetaFrame XP in the Wild: Notes from the Field
Download
Report
Transcript XP 102 – MetaFrame XP in the Wild: Notes from the Field
Citrix
MetaFrame XP
for Windows
Agenda
Intro to MetaFrame XP
What is MetaFrame XP?
What’s New in MetaFrame XP?
How is MetaFrame XP Packaged?
New Terms and Architectural Concepts
MetaFrame XP Features (technically speaking)
MetaFrame XP Advanced Management
Time permitting:
Migrating to MetaFrame XP
Useful Command Line Utilities
Intro to
MetaFrame XP
What is MetaFrame XP?
The next generation of Citrix’s MetaFrame application serving
software.
The product of a ground up reassessment by our engineers
coupled with your input on what enterprise class server-based
computing should be.
Built to eliminate current and future obstacles to speed,
performance and control while maintaining backward compatibility
for ease of migration.
Everything you have seen in MetaFrame 1.8/FR1 and MORE.
What’s New in MetaFrame XP?
Citrix Management Console
Application-based Load Management
System Monitoring & Analysis
Application Packaging & Delivery
Network Management
Printer management
Shadowing enhancements
Client time zone support
Active Directory support
NFuse ready
Enhanced scalability
Up to 1,000 servers in server farm!
What’s New in MetaFrame XP?
Reduced IT administration
Reduced network traffic
Centralized license management
Enterprise-wide license pooling
Enhanced license availability
Citrix administrator accounts
Read/Write & Read Only access
ICA client extensibility
MetaFrame and WinFrame interoperability
‘Mixed’ or ‘Interoperability’ mode
How is it Packaged?
The enterprise application serving infrastructure for Netbased Windows 2000 environments requiring extensive
scalability, rapid application delivery and robust
management—enabling unparalleled command and control
The advanced application serving platform for Windows 2000
servers and beyond designed for growing organizations that
need to maximize application availability and manageability
across the Net—all from a single point
The rapid application serving system designed for to extend
the reach of Windows 2000 Server to any device, any
departmental workgroup connection—wired, wireless, Web
MetaFrame XP Family Comparison
Functionality
Enterprise Application Serving Infrastructure
Corporate-wide deployment
20-1000+ servers in a farm
Advanced Application Serving Platform
Multiple departments and applications
2-100 servers in a farm
Base Application Serving System
Workgroup or specific application
Individual and non-load balanced servers
Size and Scope of Installation
How is it Packaged?
MetaFrame XPs
Base MetaFrame XP functionality
MetaFrame XPa
Base MetaFrame XP functionality
Load Management
MetaFrame XPe
Base MetaFrame XP functionality
Load Management
System Monitoring & Analysis
Application Packaging & Delivery
Network Management
Licensed per Connection!
Deploy as many servers as you need…
New Terms and
Architectural
Concepts
New Terms
IMA: Independent Management Architecture
Data Store: Central configuration database
LHC: Local Host Cache (Persistent data cache that exists on each
server)
Data Collector: Manages dynamic data and client
enumeration/resolution (replaces ICA Master Browser)
Zone: Deliberate grouping of MetaFrame XP servers, each with its
own Data Collector
CMC: Citrix Management Console (replaces MetaFrame 1.8
administration tools)
What is IMA? Why is it important?
IMA…
Is a TCP-based, event driven messaging bus, used by MetaFrame
XP servers.
Is a modular and easily extensible subsystem capable of
supporting current and future MetaFrame products and tools.
Overcomes the scalability constraints of the MetaFrame 1.8
platform, allowing MetaFrame XP to scale environments to new
levels.
Provides capability to administer any farm from a central tool
(CMC) that doesn’t have to run on a MetaFrame server.
Independent Management Architecture
Citrix Management
Console
MetaFrame XP
NT 4.0 TSE
MetaFrame XP
Windows 2000
Independent Management Architecture (IMA)
DB
Central Data Store
•SQL, Oracle, Access
Load
Management
Application
Packaging &
Delivery
System
Monitoring &
Analysis
MetaFrame Server Farms
MetaFrame 1.8:
Server Farms in MetaFrame 1.8 are a collection of servers on a
given broadcast segment that are managed as a single unit.
Server Farms in MetaFrame 1.8 may also be defined by sharing a
common ‘Application Set’.
MetaFrame XP:
The Server Farm in MetaFrame XP defines the scope of
management as well as the ‘Application Set’.
Server Farms in MetaFrame XP are designed to operate across
segments and are managed through the Citrix Management
Console.
MetaFrame 1.8/ICA Browser
MetaFrame 1.8/ICA Browser Attributes
Server Farms cannot span segments.
Each segment has ONE ICA Master Browser.
ICA Master Browser stores dynamic data for the segment and
handles Enumeration/Resolution for ICA clients.
Persistent data stored in registry (farm membership, licenses,
published applications, etc.).
ICA Master
Browser
MFAdmin,
PAM, etc.
Segment 1
10.1.1.x
Farm 1 (2, 3)
ICA Master
Browser
MFAdmin,
PAM, etc.
Segment 2
10.1.2.x
Farm 4 (5,
6)
MetaFrame 1.8/ICA Browser
MetaFrame 1.8/ICA Browser Attributes
Persistent data read by ICA browser/PN Service at startup.
Cross server configuration tools read/write to registry on all
servers.
Servers communicate via UDP broadcasts, remote REG calls,
RPCs, etc.
ICA Master
Browser
MFAdmin,
PAM, etc.
Segment 1
10.1.1.x
Farm 1 (2, 3)
ICA Master
Browser
MFAdmin,
PAM, etc.
Segment 2
10.1.2.x
Farm 4 (5,
6)
MetaFrame XP/IMA
MetaFrame XP/IMA Attributes
Server farms can span segments, can contain multiple zones.
Each zone has ONE Data Collector.
Persistent farm data stored in shared, persistent Data Store.
Server
Farm
CMC
DC
DC
DS
Zone 2
LHC
LHC
LHC
LHC
Zone 1
LHC
Data Collectors store dynamic data and handle
Enumeration/Resolution for ICA clients.
LHC
MetaFrame XP/IMA
MetaFrame XP/IMA Attributes
Servers communicate via IMA (TCP).
Server
Farm
CMC
DC
DC
DS
Zone 2
LHC
LHC
LHC
Zone 1
LHC
Management tool communicates via IMA to Data Store and member
servers.
LHC
Persistent data read from DS at startup, cached in Local Host
Cache.
LHC
Data Store
Attributes of the MetaFrame XP Data Store (DS)
The DS is a repository (database) which contains persistent, farmwide data, such as member servers, licenses in farm, zone configs,
printers/drivers, published apps, load evaluators, trust
relationships, etc.
Each MetaFrame XP farm shares one Data Store.
All information in the DS is stored in an encrypted binary format
(except indexes).
A farm can operate for 48 hours if DS is unavailable, then licenses
time out and no new users can connect.
A DS can be an Access, MS SQL, or Oracle database.
A DS can be configured for either ‘Direct’ or ‘Indirect’ access.
Data Store in ‘Direct’ Mode
Attributes of Direct Mode
Uses Microsoft SQL 7/2000 or Oracle 7.3.4/8.0.6/8.1.6 database.
Servers initialize directly from the DS via ODBC.
Servers maintain an open connection to the database for
consistency checks.
LHC
LHC
LHC
LHC
LHC
DS
LHC
Data Store in ‘Indirect’ Mode
Attributes of Indirect Mode
If using JET database, MF20.MDB lives on the ‘IMA host’ server.
DS
LHC
LHC
LHC
IMA Host
(indirect mode)
DC
LHC
Member servers communicate via through ‘IMA host’ server to
read/write to data store.
LHC
Uses JET 4.x, Microsoft SQL 7/2000 or Oracle 7.3.4/8.0.6/8.1.6
database.
LHC
Local Host Cache (LHC)
Attributes of the Local Host Cache
A subset of the Data Store, stored on each individual server
(IMALHC.MDB).
Contains basic info about servers in farm, pub. apps and
properties, trust relationships, server specific configs (product
code, SNMP settings, load evaluators, etc.).
Used for initialization if DS is down.
Used for ICA client application Enumeration.
Server
Farm
CMC
DC
DC
DS
Zone 2
LHC
LHC
LHC
LHC
LHC
LHC
Zone 1
Data Collectors
Attributes of Data Collectors
There is a DC for each Zone.
Server
Farm
CMC
DC
DC
DS
Zone 2
LHC
LHC
LHC
LHC
Zone 1
LHC
A DC stores dynamic information about a farm, such as servers
up/down, logons/logoffs, disconnect/reconnect, license in
use/released, server/application load, etc.
LHC
Data Collectors
Attributes of Data Collectors
DC’s distribute most persistent data changes to member servers
for LHC update.
Server
Farm
CMC
DC
DC
DS
Zone 2
LHC
LHC
LHC
LHC
Zone 1
LHC
DC’s handle all ICA client Resolution activity, should handle all
Enumeration activity. ANY DC can Resolve ANY app for ANY client
(DC’s are peers in a multi-zone implementation).
LHC
Zones
Attributes of Zones
Logical, centrally configurable grouping of MetaFrame XP servers.
Each Zone has one Data Collector (DC).
Can span IP networks (LAN, WAN).
Aren’t necessarily tied to an IP segment (only by default).
Server
Farm
CMC
DC
DC
DS
Zone 2
LHC
LHC
LHC
LHC
LHC
LHC
Zone 1
Zones
Attributes of Zones
Are useful for partitioning/controlling persistent data update traffic
and for distributing ICA client Enumeration/Resolution traffic.
A Zone can contain up to 256 hosts without a registry modification.
In most cases, fewer zones are better!
Server
Farm
CMC
DC
DC
DS
Zone 2
LHC
LHC
LHC
LHC
LHC
LHC
Zone 1
Citrix Management Console (CMC)
Attributes of the CMC
Central management tool where 98% of farm
configuration/maintenance occurs.
Extensible framework that allows different tools to ‘snap in’.
Doesn’t need to run on a MetaFrame server.
Server
Farm
CMC
DC
DC
DS
Zone 2
LHC
LHC
LHC
LHC
LHC
LHC
Zone 1
Citrix Management Console (CMC)
Attributes of the CMC
Works through the IMA service (dest. port 2513) to access DS, DC,
and member servers.
Should be run through a DC that has local access to the DS.
Is the most read/write intensive usage of the DS.
Server
Farm
CMC
DC
DC
DS
Zone 2
LHC
LHC
LHC
LHC
LHC
LHC
Zone 1
Demonstration:
CMC in Action
MetaFrame XP’s Communication
Communication ‘Layers’ (5 of them)
IMA (server to server) Communication
Persistent Data Events (1)
Dynamic Data Events (2)
Printer Management Events/Processes (3)
ICA Session (client to server) Communication
Client Enumeration/Resolution (4)
ICA Client to Server (5)
IMA Persistent Data Events (1)
Communication Events
IMA Service Initialization
Periodic Consistency Check (configurable timing)
Key: HKLM\Software\Citrix\IMA\DCNChangePollingInterval
Default value: 600000 milliseconds REG_DWORD: 0x927C0
DS
LHC
LHC
LHC
IMA Host
(indirect mode)
DC
LHC
LHC
Read heavy initialization/validation of the LHC
LHC
IMA Persistent Data Events (1)
Communication Events
Farm modification through CMC
Modifications happen through a 2-phase process:
1: CMC commits the change to the DS.
If member servers are unavailable, they receive the change
during the periodic LHC consistency check.
LHC
DC
LHC
CMC
LHC
DC
DS
LHC
2: CMC/IMA packages and distributes changes <10k to the DCs,
who then distribute it to member servers. If the change is >10k,
it distributes a change notification and servers perform
consistency check on LHC.
LHC
LHC
IMA Dynamic Data Events (2)
Communication Events
Member server notifies it’s DC of the change.
The member server’s DC notifies ALL other DC’s of the change.
LHC
DC
LHC
LHC
DC
LHC
Note: DC’s have a peer-to-peer relationship. Every DC knows what
every other DC knows.
LHC
Any state change on server (logon/logoff, disconnect/reconnect,
load change) triggers a dynamic data update.
LHC
IMA Dynamic Data Events (2)
Communication Events
Key: HKLM\Software\Citrix\IMA\Runtime\KeepAliveInterval
Default value: 60000 milliseconds REG_DWORD: 0xEA60
DC to DC consistency check.
LHC
LHC
DC
LHC
Default value: 300000 milliseconds REG_DWORD: 0x493E0
DC
Key:
HKLM\Software\Citrix\IMA\Runtime\Gateway\ValidationInterval
LHC
LHC
Member server to zone DC heartbeat check.
LHC
IMA Printer Management Events (3)
Communication Events
Printer Management has a relatively substantial impact upon IMA
traffic.
ICA Session Communication (4)
Client Enumeration/Resolution
Client to MetaFrame XP: Client asking, server answering ‘what
apps can I run?’ and ‘where do I go for this app?’
DC
Resolution (where do I go…): Client to MetaFrame XP DC,
TCP80 (default with TCP+HTTP server location), or UDP1604
(with TCP/IP server location).
LHC
Enumeration (what apps…): Client to MetaFrame XP server,
TCP80 (default with TCP+HTTP server location), or UDP1604
(with TCP/IP server location), enumerated from LHC on
MetaFrame server.
LHC
LHC
ICA Session Communication (4)
Client Enumeration/Resolution
NFuse to MetaFrame XP: NFuse asking (on behalf of the client) and
MetaFrame XP answering the above questions.
Resolution (where do I go…): NFuse to MetaFrame XP DC,
TCP80 (default), or SSL, builds ICA file with resulting info.
DC
Enumeration (what apps…): NFuse to MetaFrame XP server,
TCP80 (default), or SSL, enumerated from LHC on MetaFrame
server, presented to web browser as hyperlinks.
NFuse
LHC
Client web browser talking HTTPS to NFuse server for both
processes.
LHC
LHC
ICA Session Communication (5)
ICA Client to Server
Actual ICA session stream from the ‘Connector’ (ICA client) to a
MetaFrame XP server, destination port TCP1494 (default).
LHC
LHC
DC
(launched
from
either UI)
LHC
IMA In Depth
Hardware and Software Configuration
Load up on processors and memory
Have home directories on separate server
Roaming profiles in multi-server enviroments
Q161334-Guide to Windows NT 4.0 Profiles
and Policies
NTFS partitions only ( at least 4096 cluster )
Install only required network components and
protocols
Change drive letters at installation time only
Hardware and Software Configuration
For 4 and 8 processors systems, use one
controller for OS and one for applications and
temporary files.
Dedicate a drive for page file for best
performance.
Increase Maximum Registry Size to 100 MB.
See MF Install and Tuning Tips for more info.
Selecting a Data Store
Direct Mode
IMA directly querying the database
• Microsoft SQL Server 7 or 2000
• Oracle 7.3.4, 8.0.6, or 8.1.6
Indirect Mode
IMA requesting another server to query the database
on its behalf
• Gathering its DS information indirectly from another
server who is accessing the DS directly
Data Store Info
Indirect Mode
Select Use
a local database as the data store to enable
Indirect mode to Access (Direct Mode is not available for
Access) on the first server installed. All subsequent servers
joining the farm must be installed with the Connect to a data
store set up locally on another server option.
First server will be Zone DC by default.
Server hosting the Access DS will be the only server to write
to the Access database.
Server hosting the DS in Access acts as proxy for all other
servers.
Overcomes the file locking and corruption problems common
with Access.
Data Store Info
Using Access
Approximately
20 MB of disk space should be available for
every 100 servers in the farm.
32 MB of additional RAM is recommended if the MetaFrame
XP server will also host connections.
Need MDAC 2.5 SP 1 installed on TSE. Stop TS Licensing
Service before Installing MDAC. Reboot.
%ProgramFiles%\Citrix\Independent Management
Architecture\MF20.MDB ( System must have read/write
access)
The default user name/password is citrix/citrix. To change the
password on the database, use the dsmaint config
/pwd:newpassword command with the IMA service running.
Data Store Info
Using Access
Each
time the IMA service is stopped gracefully, the existing
mf20.mdb file is backed up, compacted, and copied as
mf20.unk. Each time the IMA service starts successfully, it
deletes any existing instance of mf20.bak and then renames the
mf20.unk file to mf20.bak. This file is used when the dsmaint
recover command is executed.
If the server runs out of disk space on the drive where the
mf20.mdb file is stored, the automatic backup stops
functioning. Always ensure there is enough disk space to hold
3 times the size of the mf20.mdb.
Perform backup of DS with DSMAINT BACKUP
Data Store Info
Using SQL
Approximately
20 MB of disk space for every 100 servers in
the farm. The disk space used may increase if there are a large
number of published applications in the farm.
The temp database should be set to Auto Grow on a partition
with at least 1 GB of free space (4 GB is recommended if it is a
large farm with multiple print drivers).
Verify that enough disk space exists on the server to support
growth of both the temp database and the farm database.
Use MDAC 2.5 SP1 on TSE. Do not use MDAC 2.6 with SQL
2000. Known bug.
Data Store Info
Using SQL
When
using Microsoft SQL Server in a replicated environment,
be sure to use the same user account on each Microsoft SQL
Server for the DS.
Each MetaFrame XP farm requires a dedicated database.
However, multiple databases may be running on a single
Microsoft SQL Server.
The MetaFrame XP farm should not be installed in a
database that is shared with any other client-server applications.
Databases should have the Truncate log on Checkpoint
option set to keep log space controlled.
Ensure DS is backed up whenever a change is made via CMC.
Data Store Info
Using SQL
For
high security environments, Citrix recommends using NT
Authentication only.
The account used for the DS connection should have
db_owner (DBO) rights on the database that is being used for
the DS.
If tighter security is required, after the initial installation of the
database as DBO, the user permissions may be modified to be
read/write only.
If installing more than 256 servers in a farm, increase number
of worker threads available for database.
Data Store Info
Using Oracle
Approximately
20 MB of disk space for every 100 servers in
the farm. The space used may increase if there are a large
number of published applications in the farm.
The Oracle Client (version 8.1.55 or 8.1.6) must be installed on
the terminal server prior to the installation of MetaFrame XP.
The 8.1.5 and 8.1.7 clients are not supported with MetaFrame
XP.
The server should be rebooted after installation of the Oracle
Client, or the MetaFrame XP installation fails to connect to the
DS.
Data Store Info
Using Oracle
Oracle8i version
8.1.6 or later is recommended. However,
Oracle7 (7.3.4) and Oracle8 (8.0.6) are supported for the
MetaFrame XP platform.
Creating a separate tablespace for the DS simplifies backup
and restoration operations.
A small amount of data is written to the system tablespace. If
experiencing installation problems, verify that the system
tablespace is not full.
Using Shared/MTS (Multi-Threaded Server) mode may reduce
the number of processes in farms over 200 servers. Consult the
Oracle documentation on configuring the database to run in
MTS mode.
Data Store Info
Using Oracle
Oracle
for Solaris supports Oracle authentication only.
Oracle user account must the the same for every server in the
farm because all servers share a common schema.
This account needs the following permissions:
Connect
Resource
Dedicating a server for Indirect Mode
May
be necessary when the following occurs:
Delays in using CMC
Increased IMA service start times due high CPU utilization on
server hosting DS.
Cut maximum users to one half to two thirds of full load to
improve performance.
Bandwidth Requirements
In
a single server configuration, a single server reads
approximately 275 KB of data from the DS. The amount of data
read is a function of the number of published applications in the
farm, the number of servers in the farm, and the number of
printers in the farm. The number of kilobytes read from the DS
during startup can be approximated by the following formula:
KB Read = 275 + 5*Srvs + 0.5*Apps + 92*PrintD
Where:
Srvs = Number of servers in the farm
Apps = Number of published applications in the farm
PrintD = Number of print drivers installed on the member
server
Data Store Info
High
Latency WAN Concerns
Without use of replicated databases, may create
situations where DS is locked for extensive periods
of time when performing maintenance
A high latency situation reads should not adversely
affect any local connections, but the remote site may
experience slow performance.
Replicated
Databases
Speed up performance if there is enough
MetaFrame servers to justify the cost
Database replication will consume bandwidth but is
controlled through the database chosen, not
MetaFrame
Data Store Info
Access
is best used for centralized farms.
Access supports only indirect mode for other servers, and as
such will have slower performance then a direct mode DS on
large farms.
Database replication is not supported with Access.
Databases supporting replication should be used when
deploying large farms across a WAN.
Server farms with over 100 servers should use SQL or Oracle
to remain at acceptable performance levels.
Data Store Info
Farms
using excessive printer drivers and scheduled
replication should use SQL or Oracle.
Farms that cycle boot large groups of servers simultaneously
should use SQL or Oracle in direct mode to minimize the IMA
service start time.
Both Microsoft SQL Server and Oracle are very similar in
performance. In the Citrix Test eLabs both database servers
performed similarly with large farms. The choice between the
two should be based on feature sets of the databases, in-house
expertise, management tools, and licensing costs rather than
performance numbers
Use Microsoft Clustering Services with SQL or Oracle
Parallel Server with Oracle for fault tolerance.
Data Store Info
DS
Query Interval
• Key: HKLM\Software\Citrix\IMA\DCNChangePollingInterval
• Default value: 600000 milliseconds REG_DWORD: 0x927C0
If
a member server is unable to contact the data store
for 48 hours, licensing will stop functioning on the
member server
CMC always connects directly to the DC
Change > 10K in size, all member servers in the
farm will be sent a change notification and query the
DS for the change
Data Distribution with Data Collectors
Server 1 writes information to the DS
Server 1 sends change notification to its zone
DC
Zone DC distributes change notification to all
member servers in its zone
Other zone DC’s receive notification and
distribute it to all member servers within their
respective zones
All member servers receive the notification and
update their LHC as requested
Data Distribution with Data Collectors
Inter-zone connection formula
• N * (N-1)/2, where N is the number of zones in the farm
IMA ping configuration parameter
• Key: HKLM\Software\Citrix\IMA\Runtime\KeepAliveInterval
• Default value: 60000 milliseconds REG_DWORD: 0xEA60
Zone DC synchronization parameter
• Key:HKLM\Software\Citrix\IMA\Runtime\Gateway\ValidationInterval
• Default value: 300000 milliseconds REG_DWORD: 0x493E0
Inter-zone connection formula
•
Key:HKLM\Software\Citrix\IMA\Runtime\MaxHostAddressCacheEntriesl
• Default Value: 256 Entries REG_DWORD: 0x100
Data Distribution with Data Collectors
Bandwidth requirements between zones
Connect: ~3Kb
Disconnect: ~2.25Kb
Reconnect: ~2.91Kb
Logoff: ~1.50Kb
CMC: ~2.23
Application Publishing: ~9.07
Data Collector Elections
Each zone is responsible for electing its own data
collector (DC). By default, the first server in the
farm becomes the DC and is set to Most Preferred.
If the setting is changed from Most Preferred,
another election will take place. DC elections are
won based on the following criteria:
1. Highest Master Version Number (1 for all
MetaFrame XP 1.0 servers)
2. Lowest Master Ranking (1=Most Preferred –
4=Not Preferred)
3. Highest Host ID (0-65536 randomly assigned at
installation)
Data Collector Elections
To view server’s ranking, use Queryhr ( copy from
support\debug\i386 on CD
DC elections are triggered in the following
situations:
A member server loses contact with the DC.
The DC goes offline.
A farm server is brought online.
The querydc -e command is executed to force
an election.
Zone configurations are changed (i.e. zone
name, election preference, adding or removing
servers)
Data Collector Elections
When a new DC is elected, all servers in the zone send a
complete update to the new DC. The following formula
can be used to approximate the amount of data in bytes
sent by all servers in the zone to the new zone DC:
Bytes = (11000 + (1000 * Con) + (600 * Discon) + (350
* Apps)) * (Srvs - 1)
Where:
Con = Average number of connected sessions per server
Discon = Average number of disconnected sessions per
server
Apps = Number of published applications in the farm
Srvs = Number of servers in the zone
Local Host Cache
Attributes of the Local Host Cache
A subset of the Data Store, stored on each
individual server (IMALHC.MDB).
Contains basic info about servers in farm,
pub. apps and properties, trust relationships,
server specific configs (product code, SNMP
settings, load evaluators, etc.).
Used for initialization if DS is down.
Used for ICA client application
Enumeration.
Local Host Cache
On
the first startup of the member server, the LHC is populated
with a subset of information from the DS. From then on, the
IMA service is responsible for keeping the LHC synchronized
with the DS. The IMA service performs this task through change
notifications and periodic polling of the DS.
In
the event the DS is unreachable, the LHC contains enough
information about the farm to allow normal operations for up to
48 hours.
During this “grace” period, the server continues to service
requests while the IMA service attempts to connect to the DS
periodically (based on the DS query interval as described in the
Data Store Activity section of the IMA Architecture chapter of
this document). If the DS is unreachable for 48 hours, the
licensing subsystem fails to verify licensing and the server stops
taking incoming connections.
Local Host Cache
Because
the LHC holds a copy of the published applications
and NT trust relationships, ICA Client application
enumeration requests can be resolved locally by the LHC.
This provides a faster response to the ICA Client for
application enumerations because the local server does not
have to contact other member servers or the zone DC. The
member server must still contact the zone DC for LM
resolutions.
If
the IMA service is currently running, but information in
the CMC appears to beincorrect, a refresh of the LHC can be
manually forced by executing dsmaint refreshlhc from the
command prompt of the affected server. This action forces
the LHC to read all changes immediately from the DS.
Local Host Cache
If the IMA service does not start, it may be caused by
a corrupt LHC.
.
1. Verify the DS is available before continuing because
this procedure causes the LHC to be reloaded directly
from the DS.
2. Stop the IMA service on the MetaFrame server.
3. Launch the ODBC Data Source Administrator. On
Windows 2000, choose Control Panel |
Administrative Tools | Data Sources (ODBC). On
TSE choose Control Panel | ODBC Data Sources.
4. Select the File DSN tab.
Local Host Cache
5. Open the imalhc.dsn file located in
%ProgramFiles%\Citrix\IndependentManagement
Architecture by default.
6. Once that file is selected, click on Create from the
ODBC Setup screen.
7. Enter in any name besides imalhc for the new LHC
database. Optionally, rename the old imalhc and reuse
the name.
8. Exit the ODBC Data Source Administrator.
Local Host Cache
8. Exit the ODBC Data Source Administrator.
9. Modify the following registry value:
Key: HKEY_LOCAL_MACHINE\SOFTWARE\Citrix\
IMA\RUNTIME
Value: PSRequired REG_DWORD: 0x1
10. Restart the IMA service.
Note: The DS server must be available for this
procedure to work. If the DS is not available, the IMA
service fails to start until the DS is available.
Security
Considerations
IMA Security
Always
install on NTFS partitions
Make sure the LHC is secure
•%SystemDrive%\Program Files\Citrix\Independent Management Architecture
•Give access to “System” and “Administrators” group with Full Control only
Run
the CMC from the console only
Run the CMC as a published application if wanting
to run on a separate machine
Make sure encryption is used for traffic between DS
and MetaFrame servers
Run the MFCfg.exe utility and remove the
“Everyone” group from each of the listeners
IMA Security
When
using the “Local Database” the MS Access
username/password is citrix/citrix. This should be
changed using DSMAINT.
The user account used to access the SQL Server
database must have “public” and “db_owner” roles
on the database that houses the DS. Do not grant
user account access to this.
SA accounts are not needed for DS access with SQL
Server. If using Oracle, do not use the SYSTEM or
SYS account.
IMA Security
A domain
user group should be used to administer
MetaFrame servers:
In the accounts domain, create a global group called
“MFAdmins”
Add domain users who will have administrative
privileges to the MFAdmins global group
Add the MFAdmins global group to each MetaFrame
server’s local administrators group
Whenever a new user account needs to be configured for
MetaFrame admin privileges, add the new account to the
MFAdmins global group
Active Directory
Single AD use a Domain Local Group
Farms that span a forest, use a universal group
Optimizations
Disk Subsystem
Disk
Caching
Lazy writes occur when data is cached instead of immediately
written to disk. If data is being sent across the network or the
server has a caching controller card, disabling lazy writes
improves performance. Network and local lazy writes can be
disabled by modifying the following registry settings:
Key: HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\
Services\Lanman Server\Parameters
Value: IRPStackSize REG_DWORD: 0x6
Key: HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\
Service\Lanman WorkStation\Parameters
Value: UtilizeNtCaching REG_DWORD: 0x0
Disk Subsystem
I/O
Locks
The registry setting IoPageLockLimit specifies the limit of the
number of bytes that can be locked for I/O operations. Since
RAM is being sacrificed for increased disk performance, the
optimal setting for this value should be determined through
pilot tests. Changing this setting from the default can speed
up file system activity. Use the table below as a guide for
changing the registry setting.
Server RAM (MB) IoPageLockLimit (Decimal) IoPageLockLimit (Hex)
64-128
256
512
1024+
4096
8192
16384
65536
1000
2000
4000
10000
Disk Subsystem
The
registry setting can be modified as follows Key:
HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\
Control\Session Manager\Memory Management
Value: IoPageLockLimit REG_DWORD
Default: 0 (512 KB is used)
For additional information on the IoPageLockLimit registry
setting, refer to the Microsoft Knowledge Base articles
Q121965 and Q102985.
Disk Subsystem
Last Access Update
The NTFS file system stores the last time a file is
accessed, whether it is viewed in a directory listing,
searched, or opened. In a multi-user environment, this
updating can cause a small performance decrease.
Modifying the following registry setting disables this
feature:
Key:
HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\
Control\FileSystem
Value: NtfsDisableLastAccessUpdate REG_DWORD: 1
Memory Subsystem
The paging file should be placed on its own disk
controller or on a partition that is separate from the OS,
application, and user data files. If the paging file must
share a partition or disk, place it on the partition or disk
with the least amount of activity.
Always set the paging file initial size to be the same
as the maximum size to prevent disk fragmentation
of the paging file. The optimal size of a paging file is
best determined by monitoring the server under a
peak load. Set the paging file to be 3-5 times the
physical RAM, and then stress the server while
observing the size of the paging file. To conserve
resources, the paging file should then be set to a
value slightly larger than the maximum utilized while
Memory Subsystem
Single-server scalability may be improved by
manually adjusting the page table entries (PTE) in
the registry. The NT kernel uses PTE values to
allocate physical RAM between two pools of
memory. By manually setting the maximum space
allocated to the System PTE, the remaining space
may be used to increase the number of users
supported on the server. Determining the optimal
configuration for PTE values is a complex task. For
detailed information see the Microsoft Knowledge
Base article Q247904. A Kernel Tuning Assistant for
Windows 2000 servers is also available from
Microsoft.
Network Subsystem
Most 10/100-based network cards auto-sense the network
speed by default. Manually setting these cards prevents the
auto-sensing process from interfering with communication
and forces the desired speed.
If working in a mixed Windows 2000 and TSE environment,
additional performance can be gained by modifying the
network request buffer size on the TSE servers. Increasing
this value to 65536 bytes, from the default of 4356 bytes,
significantly improves LAN Manager file writes. For more
information, seeMicrosoft Knowledge Base article Q279282.
Key:
HKEY_LOCAL_MACHINE\SYSTEM\CurrentContolSet\
Services\LanmanServer\Parameters
Value: SizReqBuf REG_DWORD: 65536
Range: 512 bytes to 65536 bytes
Network Subsystem
The server may refuse connections due to self-imposed limits
specified by the MaxMpxCt and MaxWorkItem registry
values. The users may see the following errors:
“System could not log you on because domain <domainname>
is not available”
“You do not have access to logon to this session”
Before changing these values, read the Microsoft Knowledge
Base article Q232476. When modifying these registry
settings, be sure that the MaxWorkItems value is always 4
times the MaxMpxCt value. Suggested new values for
MaxMpxCt and MaxWorkItems are 1024 and 4096
respectively.
Key:
HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\
Services\LanmanServer\Parameters
Network Subsystem
Key:
HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\
Services\LanmanServer\Parameters
Value: MaxWorkItems REG_DWORD: 4096
Value: MaxMpxCt REG_DWORD: 1024
To ensure that a host server is quickly aware of dropped
sessions, the two TCP registry settings listed below can
be modified with the following moderately aggressive
values:
Key:
HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\
Services\Tcpip\Parameters
Value: KeepAliveTime REG_DWORD: 0000ea60
Value: KeepAliveInterval REG_DWORD: 000003e8
Network Subsystem
Caution: Aggressive parameters may cause TCP/IPbased
communications to time out prematurely. These
parameters
should be adjusted as necessary to prevent this behavior.
For more information, see CTX708444: How to Set TCP
Keep Alives so TCP/IP Users Go To Disconnected State
in the Citrix Knowledge Base.
Server Configuration
In some instances, modifying the NT application
performance setting may provide an additional
performance boost. Set the Application performance
slider to None.
When opening remote procedure call (RPC) aware
applications such as Windows Explorer and Control
Panel, delays of several minutes may be the result of
incorrect service startup settings. Verify that the RPC
service Startup type is set to Automatic and the RPC
Locator service Startup type is set to Manual.
Set Server Service to Maximize throughput for Network
Applications.
Disable AutoGrammar for Microsoft Word.
MetaFrame XP
Features Revealed
MetaFrame XP Management
Centralized Administration
Single Point Command and Control
All administration, configuration, monitoring and control of the
Citrix Server Farm is managed centrally.
Independent Management Architecture
IMA-compliant servers and management products share a common
and extensible management infrastructure.
Unified Management Console
The Citrix Management Console communicates across a single
Management Scope of the server farm using the IMA protocol.
Central Data Store
Configuration information for the Server Farm is stored centrally in
the Citrix Data Store.
Citrix Management Console
Centralized License Management
Description: Licenses are installed into the Central Data Store and
managed centrally from the Citrix Management Console.
Benefit: Administrators can easily manage all of the licenses for
the Citrix Server Farm from a single point.
License Pooling Across Subnets
Description: ICA Connection licenses for client connections
can be pooled across the entire server farm regardless of
whether the server farm crosses network subnets.
Benefit: Enables pooling of ICA connection licenses across
the extended enterprise network within the MetaFrame Server
Farm.
Data Center 1
Data Center 2
Subnet 10.1.X.X
Subnet 10.2.X.X
Data Center 3
Subnet 10.3.X.X
Only 1 ICA License
Is utilized
License Fail Over and Redundancy
Description: MetaFrame XP allows licenses to be available
for fail over and redundancy. In the case of a server failure
the server and connection license are still available.
Benefit: Administrators have the ability to bring up “warm”
backup servers in the event of a server failure without the
need to re-install and re-activate their server licenses.
MetaFrame XP License Management
Centralized License Management
Flexible licensing for emerging business models
Increased flexibility to support Citrix Licensing Programs (Shrink
Wrap, CLP, ELP, and iLicense).
Single point of license installation and activation
License installation and activation can be done centrally via the
Citrix Management Console.
Support for multiple server/product platforms
The new licensing system supports all MetaFrame XP server and
connection licenses.
Active Directory Support
Application Publishing
Enables application publishing to users and groups in Active
Directory.
Account Authority Access
Utilizes native Active Directory Interfaces to access the Active
Directory.
User Principal Names
Allows users to logon to the MetaFrame XP server using User
Principal Names: i.e. [email protected].
NFuse and Program Neighborhood
Enables users to utilize their Active Directory accounts to access
MetaFrame XP applications via NFuse and Program Neighborhood.
Active Directory Support
Description: Applications can be published on MetaFrame
XP servers and assigned to users from Active Directory.
Benefit: MetaFrame XP integrates and fully supports
Application Publishing in a native Active Directory
environment.
Publish
Application
for ADS Users
from the CMC
ADS
Enumerate User and
Group Accounts
from Active Directory
Citrix Management
Console
Printer Management
Printer Management
Printer Driver Replication
Printer Mapping
Network Printer Auto-Creation
Printer Compatibility
Printer Bandwidth Control
Terminal Printer Auto-Creation
Client Printer Creation Logging
Printer Management
Printer Mapping
Ability to create mappings for Windows 9X client printers on the
MetaFrame XP server and automatically distribute to the server
farm.
Printer Bandwidth Control
Allows the administrator to specify the amount of bandwidth that
can be used by printing over the client connection.
Setting used for all users over all connections for a given
MetaFrame XP server.
Terminal Printer Auto-Creation
Ability for the administrator to setup auto-creation of printers for
ICA DOS and WinCE Terminal Devices.
Client Printer Creation Logging
Logs all information about auto-creation of client printers, allowing
the administrator to proactively detect printer issues and resolve
them with the required information.
Printer Driver Replication
Description: Through the Citrix Management Console, printer
drivers can be managed across the entire server farm.
Benefit: Gives administrators the power and control to manage and
distribute printer drivers to all of the MetaFrame servers in the server
farm, providing a consistent printing environment for all users.
Install New Printer
Driver on 1 Server
Use CMC to Manage and
Distribute New Printer
Driver to entire Server
Farm
CMC
Printer Driver Mapping
Description: Printer drivers on different platforms often have
differing names which can interfere with client printer creation.
Printer driver mapping enables administrators to control
differing printer drivers.
Benefit: Allows the administrator to specify mappings of
printer driver names from one platform to another (ie.
Windows 95 to Windows 2000)
Printer Compatibility
Description: Printer compatibility allows the administrator to specify
client printers that can be used in the MetaFrame environment or
specify printers that can never be used.
Benefit: Gives administrators the power to determine and control
the types of client printers that can be utilized on the MetaFrame XP
servers enabling them to ensure a consistent and stable computing
environment.
ICA Client
Printer is on the
restricted list
ABC Printer
Printer
Creation is
Disabled
ICA Connection
Restricted
•XYZ Printer
•ABC Printer
Printer Bandwidth Control
Description: Bandwidth limits can be specified for printing
from an ICA Client.
Benefit: Allows the administrator to control and specify the
amount of bandwidth that can be used for printing in the
MetaFrame XP server farm.
Terminal Printer Auto-Creation
Description: Printers connected to ICA DOS and WinCE
terminal devices can be pre-defined for auto-creation from the
CMC.
Benefit: When users login to MetaFrame from the Terminal
devices, the pre-defined printer will be auto-created without
any user interaction.
Client Printer Creation Logging
Description: All information related to client printer creation
is logged in the system event log.
Benefit: Gives administrators the power and information to
proactively detect and resolve client printer issues.
Printer Management Recommendations
Recommendations
Printer drivers can only be replicated to the servers of the same OS
as the source server.
Install drivers on the source server and select any available port on
the server.
If installing for the sole purpose of replication there is no need to
share the printers or set them as default.
Can be very CPU intensive on the source server so avoid
replicating drivers while the source server has a heavy load.
Printer Queue Management
#QueueEntries = [#Drivers] * [#Servers]
Every driver/server combination creates a queue item in the
printer replication queue.
Should not exceed 1500 entries in length.
Eg. 30 drivers to 50 servers
QPRINTER Utility
Not installed by default.
\support\debug\i386
QPRINTER /REPLICA
Expected Performance
Handled by IMA Service at very low priority.
Depends on network traffic and server load.
Shadowing Enhancements
Shadowing Installation Option
Ability to select whether shadowing is
available.
Lock down the shadowing configuration to
avoid changes.
Allows administrators flexibility with privacy
and security issues involving shadowing.
Shadow Indicator
Notifies users that shadowing is in progress.
Provides users with a “cancel” button to end
the shadow.
Shadow Activity Logging
Logs all session and user information during a
shadow.
Enables the creation of a shadow “audit log”.
ICA Client Enhancements
Published Application Parameter Support
Enables the MetaFrame server to accept published application
parameters provided by a client, and the client to pass published
application parameters to the server.
ICA Client Object Interface
A framework that exposes the functionality of the Citrix ICA Win32
Client to other objects or applications. Allows any application that
supports embedding of objects, to interface with and pass
instructions to the ICA Client.
Per Session Time Zone Support
Ability to run applications on the MetaFrame server in the context
of the users local time zone. The MetaFrame Server can support
different users running applications at different time zones on the
same server.
ICA Client Object
Provides a programmable interface for
integration of ICA Clients into
Portals
Dashboards
Vertical market applications, etc.
Supports major web browsers
Internet Explorer 4.0 and greater
Netscape 4 and greater
Supports ActiveX “containers”
MS Office, MS Visual Studio tools, Borland Delphi, etc.
Per Session Time Zone Support
Redmond
Time Zone
GMT -8
ICA Client
Ft. Lauderdale
Time Zone
GMT -7
MetaFrame XP Server
Salt Lake City
Time Zone
GMT -5
ICA Client
Published Applications
Run in context of
User’s Local Time Zone
London
Time Zone
GMT
ICA Client
NFuse Ready
NFuse Ready
NFuse now integrated into MetaFrame XP.
NFuse install option if IIS detected.
Sets up default web and startup page.
In short—you can now ACCIDENTLY deploy NFuse!
MetaFrame XP
Advanced
Management
Load
Management
Load Management (MetaFrame XPa and XPe)
Load Management
Configuration of application load balancing.
Monitoring of application and server load.
Dynamic adjustment of load balancing criteria.
Citrix Load Management replaces Load Balancing Services in
MetaFrame 1.8.
Load Management utilizes IMA for communication.
Provides the ability to create criteria for servers and applications.
Load Management code built into MetaFrame XPa/XPe (no separate
CD-ROM).
Major Components
Major Components of Load Management
Rules
Measure statistics for high or low loads on servers.
Lower a rule, reach a threshold more quickly.
Elevate a rule, a threshold is hard to reach.
Load Evaluators
Used to configure server load measurements.
Use Default for Citrix provided load evaluators.
Use Advanced to create your own.
Can vary on each server and/or application.
Can use any combination of rule and load evaluators per server
across the farm.
Load Management Criteria
Load Management Criteria
CPU Utilization
Memory Usage
Page Swap
Page Fault
Server User Load
IP Range New
Scheduling New
Context Switches New
Disk Data I/O New
License Threshold New
Application User Load New
Disk Operations New
Load Management Criteria
IP Range
Using the IP Range rule, an administrator can specify a distinct
address or set of addresses that can access the published
application.
Scheduling
Using the Scheduling rule, administrators can create a Load
Evaluator that allows access to a specific application or server only
during specified days and times.
Load Monitoring
Load Monitoring
Load management provides monitoring capabilities that allow
extended analysis of how load evaluation criteria is affected in the
enterprise. With monitoring capabilities and trend graphs,
evaluation criteria can be monitored and adjusted over time.
Pre-Configured Load Evaluators
Default
Rule represents the number of users logged onto a MetaFrame XP
server.
Contains one rule, Server User Load, that reports a full load when
100 users log on to the attached server.
Advanced
The rules in this load evaluator represent server performance
using:
Disk I/O
CPU Utilization
Disk Operation
Memory Usage
System
Monitoring and
Analysis
System Monitoring & Analysis (MetaFrame XPe)
System Monitoring & Analysis
Uses utility called “Resource Manager.”
Ground up re-write!
Integrates with the Citrix Management Console.
Adds tabs to the CMC.
Can control summary data in the CMC.
Configure Alert recipients in the CMC.
Adds counters to each server for monitoring, can manage several
servers in the CMC.
Monitors application usage by published applications.
Watcher Window requires the CMC to monitor servers.
Located on “System Monitoring & Analysis” CD-ROM.
Feature Categories
Real-time server monitoring
Receive real-time notification of server problems such as memory
shortage, hard disk space or CPU utilization.
Real-time application monitoring
See at a glance exactly how many application licenses are being
used farm wide.
Alerting
Receive notification via the CMC, SMS message to mobile phones,
SNMP or e-mail.
Commonly Used Terms
Local database: A database created on every MetaFrame XPe
server for storing the real-time information
Farm metric server: Interprets farm-wide data and deals with alerts
Metric: A trackable item that Resourse Manager measures for
servers or applications (e.g., memory available bytes)
Resource Manager Application: An application which may or may
not have been published by MetaFrame, but which you have set up
to be monitored by MetaFrame
CMC Integration
Description. System Monitoring and Analysis is integrated
into the Citrix Management Console.
Benefit. Provides a single point of control for monitoring and
managing your application serving environment.
Application Server Farm Monitoring
Description. Track applications to determine when new
licenses are needed. Configure options for servers—either
individually or farm-wide—to trigger alarms when various
thresholds are reached.
Benefit. Monitors applications and server farms as the
enterprise grows.
Real-time Graphing and Alerting
Description. Monitor the health and performance of
application servers in real time while simultaneously receiving
a wide selection of alerts, including on-screen notifications,
email messages sent to mobile phones and SNMP traps.
Benefit. Detects and resolves potential performance
bottlenecks before they become system failures.
Watcher Window
Description. Monitors parameters through a small watcher
window in the corner of the screen. If an alert is raised,
simply double click on the alert icon to access all relevant
information using the CMC interface.
Benefit. Allows for constant monitoring of a server farm while
working in other applications.
System Scalability
Description. System Monitoring and Analysis is designed to
scale with your MetaFrame XPe environment.
Benefit. Expanded management as servers and server farms
grow.
Intuitive User Interface
Description. Click on an application in the CMC to bring up a
menu of functions, including snooze, sleep, real-time graph,
properties and alarm options. Add and remove alarms with a
few clicks of the mouse.
Benefit. Allows easy and quick set up of alarms and monitors
performance by application.
Simplified Setup
Description. No database setup required—works out of the
box with MSDE.
Benefit. Simplifies installation and eliminates errors with
creating a database. Allows instant access to all features
across the environment.
Server and Application Groups
Description. Create server or application groups consisting
of specific users by department or location.
Benefit. Ease viewing and management of many servers.
Server Reboot Support
Feature Description. All managed servers can be scheduled
to reboot at specific times.
Benefit. Eliminate the tedious, time-consuming task of
manual reboots.
Application
Packaging and
Delivery
Application Packaging & Delivery
(MetaFrame XPe)
Application Packaging & Delivery
Uses utility called “Installation Manager.”
Nearly a ground up re-write!
Integrates with the CMC.
Configure Network account to be used by the installer service to
install packages.
Can select to reboot servers post installation.
Define how often to expire and remove “jobs”.
Define server groups and application packages.
Status can be checked in Job properties.
Located on “Application Packaging & Delivery” CD-ROM.
How Does It Work?
Package, Deliver, and Publish
Using the Packager, software replication packages are
automatically created and prepared for distribution.
Packages are then scheduled for delivery to the targeted servers
via the Citrix Management Console.
Once delivered, applications can be published to provide instant
access through Citrix Program Neighborhood and NFuse.
Applications can also be automatically uninstalled with a few clicks
of the mouse.
Installation Management
Key Elements
Packaging. Configure once, automatically deliver to all – fast and
accurately.
Delivery. Choose the where, when and how of delivery and
installation with complete confidence. And, verify the results, too.
Application Packaging
Description. Include changes to applications in an
installation script that can be automatically replicated on
multiple application servers. This includes unattended
installations where there is no need for application recording.
Benefit. Improves productivity by reducing the time and
effort to manage change in the application server
environment.
Service Pack Packaging
Description. Facilitates maintenance of application serving
environments by enabling the packaging, delivery and
installation of service packs and patches.
Benefit. Maintain applications and distribution of service
packs and patches quickly and easily with central
management, reducing need for additional IT staff.
File Packaging
Description. Packages individual files or groups of files for
distribution.
Benefit. Enables administrators to distribute company
templates and documents associated with a particular
application. They can also use this feature to distribute a
system profile to be used by everyone accessing the
application.
Improved User Interface
Description. Provides a greater degree of control over the
installation package, plus more feedback about the status of
the application being delivered.
Benefit. Gain and enhance user experience and additional
feedback about the status of the application being delivered.
Project Details
Description. View all project settings, including file details,
registry changes, and target directories within the project.
Benefit. Customize and plan during the creation of a
package.
Rollback
Description. Quickly and easily “wipe the slate clean” on the
packaging server following package creation to prepare for
new installations.
Benefit. Spend less time restoring the packaging
environment and more time deploying applications and
supporting users.
CMC Integration
Description. Access Installation Manager from the Citrix
Management Console.
Benefit. Enjoy a single point of control for managing the
entire server farm.
Package Delivery
Description. Deliver a package of applications, files and/or
service packs to multiple servers from a central point in
minutes instead of days or weeks.
Benefit. Save time and improve productivity by ensuring
rapid time-to-value for new or updated applications.
Scheduling
Description. Set up installations to occur automatically
during off-peak hours or on weekends.
Benefit. Conserve bandwidth and minimize user disruption.
Server Groups
Description. Create server groups based on different
categories, such as operating system, geographic location,
department or other user-defined criteria.
Benefit. Precisely target application delivery to the desired
servers.
Server Reboot Support
Description. Support applications that require server
rebooting upon completion of installation. Users connected to
the application server will automatically be notified prior to
server rebooting.
Benefit. Eliminate the tedious, time-consuming task of
manual reboots.
MSI Support
Description. Deploy any application that provides a
Microsoft Installer Package (MSI) without the need for
repackaging.
Benefit. Enjoy all the benefits of this common, industry
standard such as, self-healing, install on demand and DLL
resolution.
Delivery Verification
Description. Status of application delivery to target servers
can be easily verified through Installation Manager.
Benefit. Enjoy added confidence when centrally delivering
applications.
Inventory
Description. Allows administrators to easily inventory all
applications delivered to a server using Installation Manager.
Benefit. Simplify the process of tracking software deployed
in large-scale, multi-application environments.
Network
Management
Network Management (MetaFrame XPe)
Network Management
Network Management is an SNMP agent that runs on your
MetaFrame XP servers.
It can be managed with any SNMP management service or utility.
SNMP agent automatically installed with MetaFrame XPe.
Console plug-ins are available for:
Tivoli NetView (v. 5.1.2 and above)
HP OpenView (v6.0 only)
Plug-ins are located on “Network Management” CD-ROM.
How Does It Work?
Simple Network Management Protocol (SNMP)
Allows network devices to be monitored
and managed from a central location
SNMP
Manager
SNMP Managers are applications that collect
SNMP data and receive SNMP events (traps)
Standard SNMP Support
Description. Citrix now supports the most widely used
network management protocol, SNMP
Benefit. Now conveniently integrates with a huge body of
existing software and hardware tools based on SNMP.
SNMP
SNMP
SNMP Management Console
Integration with Market Leaders
Tivoli NetView with the MetaFrame XPe Plug-in
Monitor & Control MetaFrame Servers
Description. Discovers, monitors and controls MetaFrame
XPe servers in single or multiple farms.
Benefit. Conveniently monitor and control common
MetaFrame session and user status information across
multiple farms from a single console.
Monitor & Control MetaFrame Servers
Disconnect session, send message, logoff user,
and reboot server
Migrating to
MetaFrame XP
Why Migrate to MetaFrame XP?
Increased farm scalability and stability
Easier to manage with CMC
Integrated advanced management capabilities
Simplified license management and activation
Printer management
Enhanced NFuse integration
Active Directory User Principal Name support
Client time zone support
Less server-to-server network traffic
MetaFrame 1.8 and Feature Release 1 enhancements integrated and
available to more clients
Mixed Mode Is…
Mixed Mode is designed to facilitate migration to
MetaFrame XP with little or no end user
disruption
Provides support for:
Published application migration
Application load balancing
Subnet license pooling
Existing NFuse, Program Neighborhood, and Custom ICA
connections
Mixed Mode Is Not…
Mixed Mode is NOT designed to be a permanent
solution
Interoperability is achieved by emulating the
services and communication mechanisms used
by MetaFrame 1.8
Mixed Mode – Architecture Comparison
MetaFrame 1.8
Other
PN Srvrs
NFuse
(named pipe)
HTTP
HTTP &
XML
ICA Client
XML Svc
PN Named
Pipe
PN Virtual
Channel
UDP 1604
QServer
UDP 1604
PN Svc
ICA Browser
Svc
Local
NT
Registry
AppCfg
MFAdmin
WinstationAPI
(RPC)
NT Reg
Termsrv
(remote
regedt)
Mixed Mode – Architecture Comparison
MetaFrame XP
(Native Mode)
Other
IMA Srvrs
NFuse
(TCP)
HTTP
HTTP &
XML
ICA Client
XML Svc
TCP
PN Virtual
Channel
CMC
IMA
Svc
TCP
Mixed Mode – Architecture Comparison
MetaFrame XP
(Interoperability Mode)
NFuse
Other
IMA Srvrs
Other
PN Srvrs
(TCP)
(named pipe)
HTTP
HTTP &
XML
ICA Client
XML Svc
PN Named
Pipe
PN Virtual
Channel
UDP 1604
QServer
UDP 1604
PN Svc
ICA Browser
Svc
Local
NT
Registry
AppCfg
MFAdmin
WinstationAPI
(RPC)
NT Reg
Termsrv
(remote
regedt)
IMA
Svc
Mixed Mode
Until you get to Native Mode, you can’t take full
advantage of:
Increased farm scalability and stability
Advanced printer management
Active Directory UPN support
Simplified license management and activation
MetaFrame 1.8 license gateways are not supported.
MetaFrame connection licenses are equally distributed among
subnets.
CMC/Farm/Properties/Interoperability can change licenses
assigned to each subnet.
Mixed Mode
Running in Mixed Mode
On first MetaFrame XP install, if MetaFrame 1.8 is detected on the
segment, it will offer to run in Mixed Mode.
Administrators must use two sets of tools to manage a mixed farm.
If yes, legacy tools are automatically installed.
appcfg shipped on MetaFrame XP is same as MetaFrame 1.8
SP2. Older versions may not be able to manage applications
published with newer versions.
Applications may be published on MetaFrame 1.8, then MetaFrame
XP--not the reverse.
Migration
Strategies
Migration Strategies – Flash Upgrade
All servers are upgraded to MetaFrame XP during
scheduled network maintenance window
Consider this for highly centralized and/or cloned server
environments.
Citrix now supports both unattended and cloned installs for all but
the first server in a MetaFrame XP farm. See specific
documentation in Admin Guide.
Note: Repeated licenses will give an error upon migration to
IMA Data Store.
Migration Strategies – Parallel
MetaFrame XP servers built in native mode.
MetaFrame 1.8 and XP servers do not
communicate with each other.
Consider this for fast growing installs, new
Windows 2000 rollouts, or multi-site scenarios:
Requires additional hardware and licenses.
Alternately, users may be manually migrated in proportion to
servers.
MetaFrame XP apps are published manually rather than migrated.
Publish MetaFrame XP and 1.8 apps to distinct user groups to
prevent redundant icons.
Migration Strategies – Mixed Mode
Rolling upgrade of existing MetaFrame servers
Set during install of first server in the farm.
MetaFrame XP and 1.8 farm names must match.
MetaFrame XP server will win ICA browser election.
(except MetaFrame 1.8 SP1 MB hardcode)
Mixed Mode applies to all MetaFrame XP servers in the farm.
Starts PN and ICA browser services on MetaFrame XP servers.
Existing apps are migrated to IMA data store (1 time).
Any appcfg changes made to MetaFrame 1.8 apps after
migration are not updated to the data store.
NFuse as a Bridging Technology
NFuse allows administrators to hide complexity
from the end user. For migration, it can be used
to present applications from an arbitrary number
of farms.
MetaFrame XP for Windows
MetaFrame 1.8 for Windows
MetaFrame 1.1 for UNIX
Multiple sites
Migration
Scenarios
Scenario 1: Single Site, Single Farm
Migration
High level steps (‘rolling’ upgrade):
1. Create IMA Data Store if necessary.
2. Upgrade a MetaFrame 1.8 server other than the ICA Master
Browser.
Install in Interoperability Mode when prompted.
Apply upgrade licenses to MetaFrame XP farm.
3. Upgrade remaining servers.
4. Switch to MetaFrame XP Native Mode.
5. Enable/disable UDP browsing as needed.
Scenario 1: Single Site, Single Farm
Migration
Single farm migration tips:
Avoid publishing new apps or changing app configuration while in
MetaFrame XP Interoperability Mode. If necessary, create/modify
apps in MetaFrame1.8 first, then MetaFrame XP.
Use NFuse and/or auto-client update to distribute new ICA clients.
Scenario 2: Multi-Farm Consolidation
High level steps:
Ensure IMA server-to-server communication (default TCP 2512).
Upgrade first farm (including switch to Native Mode) or build new
Enterprise MetaFrame XP farm (in Native Mode).
Perform upgrades of other MetaFrame 1.8 servers (one farm at a
time) joining them to the ‘Master’ MetaFrame XP farm.
Some manual cleanup of duplicate app names may be necessary.
Scenario 2: Multi-Farm Consolidation
Multi-farm consolidation tips:
Key: managing user connectivity
If possible, use an NFuse portal pointing to multiple farms.
NFuse can play a HUGE role here!
If using PN, add/change Application Set objects and server
location/browser type.
If using single published app, may need to modify server
location/browser type.
If using ICA file(s), may need to modify server location/browser
type.
Use NFuse and/or auto-client update to distribute new clients.
Useful Command
Line Utilities
Useful Command Line Utilities
QUERY FARM (QFARM, replaces QSERVER)
/APP Display app names and server load.
/DISC Display disconnected session data.
/LOAD Display server load.
/PROCESS Display active processes.
/ADDR Display address data on selected server.
/TCP, /IPX, /NETBIOS Display protocol data.
CLICENSE.EXE: Built in, useful for querying licensing information
on the farm.
Add_and_activiate
Enumerate
In_use
Servers_using
Useful Command Line Utilities
QUERYHR.EXE: From Support directory on MetaFrame XP CD,
useful for querying zone/DC info on the farm.
-z Show all the available zones
-h <zone name> Show all the hosts in a specified zone
-l Show the Local Host Record
-n <host name> Show the specified Host Record given a host
name
-I <Host ID> Show the specified Host Record given a host ID
-N Show the farm name
-d <Host ID> Delete an IMA Host Entry
Useful Command Line Utilities
QUERYDS.EXE: From Support directory on MetaFrame XP CD,
useful in determining what servers are currently alive in a server
farm.
Usage:
Queryds /table:<tablename>[/query:<querystring>]
Table names:
SubscriptionTable
ServiceTable
PN_Table
Conn_Sessions
Disc_Sessions
Useful Command Line Utilities
QUERYDC.EXE: From Support directory on MetaFrame XP CD,
useful for querying DC info and forcing ‘elections’.
-z <zone name> Show Data Collector name
-e Force Election
-a Show data collectors for all zones
QPRINTER.EXE: From Support directory on MetaFrame XP CD,
useful for viewing printer replication queue and importing mapping
files into the DS.
/REPLICA Display info about printer replication queue
/IMAPRMAPPING <file name> Import mapping file into DS.