SlideShare a Scribd company logo
1 of 27
Download to read offline
WHITE PAPER | Citrix XenServer
Citrix XenDesktop on XenServer
and IBM Hardware
Sizing Baseline and Reference Configuration
www.citrix.com
Page 2
Contents
Overview.............................................................................................................................................................3
Executive Summary...........................................................................................................................................3
Reference Architecture......................................................................................................................................4
Hardware Infrastructure Components ......................................................................................................................5
XenDesktop 5 Components .......................................................................................................................................7
XenServer Hosts.........................................................................................................................................................7
XenDesktop Desktop Delivery Controllers (DDCs).....................................................................................................8
Provisioning Services (PVS)........................................................................................................................................8
Citrix Web Interface...................................................................................................................................................9
Citrix Licensing...........................................................................................................................................................9
Virtual Desktops ......................................................................................................................................................10
Login Consultants VSI ..............................................................................................................................................11
Methodology.....................................................................................................................................................12
Pool Spin Up.............................................................................................................................................................12
User Workload.........................................................................................................................................................12
Findings.............................................................................................................................................................14
Success Criteria........................................................................................................................................................15
Test Results..............................................................................................................................................................16
References.........................................................................................................................................................26
Revision Description Managed By Date
1.0 Document created Roberto C. Moreno 3/14/2011
1.1 Document revised Ricardo Colon 3/17/2011
1.2 Document edited Roberto C. Moreno 3/17/2011
1.3 Document revised Zaid Haiman 3/18/2011
1.4 Document edited Roberto C. Moreno 3/18/2011
1.5 Document edited Roberto C. Moreno 3/19/2011
1.6 Document revised Zaid Haiman 3/21/2011
2.0 Final Draft Zaid Haiman 4/6/2011
Page 3
Overview
The overall goal of this White paper is to provide sizing baselines for the deployment of Citrix
XenDesktop on IBM BladeCenter Servers running Citrix XenServer.
A laboratory environment was created to test and validate the different components of the IBM
hardware and the XenDesktop solution. The main focus of all tests conducted in the laboratory was
to define the optimum number of virtual desktops that can be hosted on a single IBM BladeCenter
HX5 server.
Reference configuration of the environment, methodology for conducting tests, and the findings
from tests performed are presented in this document.
This document assumes that the reader has an architectural understanding of blade server hardware,
storage area networks, and the Citrix XenDesktop Platinum components.
Executive Summary
The laboratory environment used for conducting all tests was based on the IBM BladeCenter HX5
Blade Server and the IBM System Storage N5200.
A total of two BladeCenter HX5 servers were used. Each featured dual 1.86GHz Nehalem-EX
processors with Intel Hyper-Threading enabled. Both servers were equipped with 50GB solid state
drives (SSD), dual port Gigabit NICs, and Qlogic Fiber Channel adapters. Memory configurations
included 64GB of DDR3 RAM for the blade hosting virtual infrastructure servers and 128GB for
the blade hosting the virtual desktops.
The IBM System Storage N5200 featured a cluster of two controllers and the IBM EXN4000 disk
expansion unit with 14 FC 15K HDDs. The N5200 was attached to the HX5 Blade servers via
2Gbit Fiber Channel.
The latest release of the Citrix XenDesktop 5 components available at the time of testing were used
to build the laboratory virtual desktop infrastructure. The XenDesktop 5 Site consisted of:
 Two (2) Desktop Delivery Controllers for brokering and managing the virtual desktops,
 Two (2) Provisioning Servers for OS provisioning.
Citrix Licensing and Web Interface servers were also deployed as part of the environment. All of the
XenDesktop 5 infrastructure servers were virtualized on XenServer 5.6 FP1.
Tests conducted within the laboratory environment validated the Citrix XenDesktop 5 hosted VDI
solution running on the XenServer platform with IBM BladeCenter HX5 servers. Single server
scalability tests show a single IBM BladeCenter HX5 server being capable of supporting an optimum
number of 80 virtual desktops with a “medium user” workload using Microsoft Office 2010 on
Microsoft Windows 7 Professional.
Page 4
Memory utilization was determined to be the main bottleneck limiting the server from effectively
hosting additional desktops with the XenServer virtual desktop host reporting 98% of RAM
utilization with 80 active desktops.
A combination of tools was used to simulate user load, monitor resource utilization, and gather data
in order to validate the reference configuration. The latest version of the Login Virtual Session
Indexer (VSI) 3 from Login Consultants was used for load testing and capturing system response
time data. Other tools used for benchmarking and monitoring the environment included Citrix
XenCenter, Citrix VM Performance Utility, Windows Performance Monitor, and NetApp’s
ONTAP.
Reference Architecture
This section details the various components used in the laboratory to configure the environment.
Configurations and specifications of the hardware infrastructure, XenDesktop 5, and LoginVSI
components is presented in the following subsections. This configuration is only provided for
reference purposes; specific configurations for production environments will vary based on each
implementation’s needs.
Figure 1 – Laboratory Components
Page 5
Environment Summary:
2 XenServer Hosts Hosting of infrastructure VMs and virtual desktops
2 Desktop Deliver Controllers Brokering and managing virtual desktops
2 Provisioning Servers Virtual desktop OS provisioning
1 Citrix Web Interface / AD Domain Controller Provides access to XenDesktop resources and
directory services
1 Citrix Licensing Server Licensing for all XenDesktop components
5 LoginVSI Launchers Initiates user sessions
Table 1
Hardware Infrastructure Components
All hardware used to build the laboratory environment was provided by IBM. The environment
consisted of IBM BladeCenter HX5 servers and N Series System Storage.
The blade chassis provided was the IBM BladeCenter S equipped with Ethernet and Fiber Channel
switch modules. Installed on the chassis were two BladeCenter HX5 blade servers: one for hosting
infrastructure servers and the other for the hosting of virtual desktops. Each server was configured
with dual Intel Nehalem-EX processors, DDR3 VLP RAM, Solid State Drives, Qlogic Fiber
Channel adapters, and onboard Broadcom network interface cards.
Storage was provisioned by IBM System Storage N5200 controllers and the EXN4000 disk
expansion unit connected to the servers via Fiber Channel. The aggregate used for the test
environment consisted of six FC 15K 300 GB disks in a RAID 4 configuration. This provided the
best performance and capacity combination based on the amount of disks available and the
requirements of the environment.
Additional details of the configuration of all hardware components used is presented tables 2 – 4
below.
Figure 2 - IBM BladeCenter S
Page 6
Component LAB Configuration
IBM BladeCenter S Two BladeCenter HX5 Blades
Four power supply modules
Four cooling modules
One management module
One BNT Layer 2/3 Copper Gigabit Ethernet Switch Module
One Brocade 8Gigabit Fiber Channel Switch Module
Table 2
Figure 3 - IBM BladeCenter HX5
Component LAB Configuration
IBM BladeCenter HX5 XenServer Virtual Desktop Host
 Two Intel Xeon Processors L7555
 Intel Hyper-Threading Technology enabled
 128 GB RAM (16 x 8GB DDR3 SDRAM)
 50GB SSD Drive (SATA 1.5GB)
 Broadcom 5709S onboard NIC with dual Gigabit Ethernet ports
 Qlogic 8Gb Fiber Channel Expansion Car
IBM BladeCenter HX5 XenServer Management Server Host
 Two Intel Xeon Processors L7555
 Intel Hyper-Threading Technology enabled
 64 GB RAM (16 x 4GB DDR3 SDRAM)
 50GB SSD Drive (SATA 1.5GB)
 Broadcom 5709S onboard NIC with dual Gigabit Ethernet ports
 Qlogic 8Gb Fiber Channel Expansion Card
Table 3
Page 7
Figure 4 - IBM N5200 Figure – 5 IBM EXN4000
Component LAB Configuration
IBM System Storage N
Series
N5200 (2864-A20) Modular Disk Storage System
Dual Controller Active/Active Configuration
DATA ONTAP 7.3.4
2 x Intel Xeon @ 2.8 GHz
4 GB RAM
1 GB Nonvolatile Memory
8 x 2Gb Fiber Channel Ports
8 x 1Gb Ethernet Ports
Disk Expansion Unit EXN4000
4 x 2/4Gb Fiber Channel Ports
14 x 300 GB 15K RPM FC Disk Drives
Table 4
XenDesktop 5 Components
The XenDesktop 5 solution included the XenServer Hypervisor, XenDesktop Desktop Delivery
Controllers (DDC), Provisioning Servers (PVS), Citrix Web Interface and License servers.
All of the XenDesktop infrastructure servers were virtualized on XenServer. With a proper design
and best practices applied none of these components require the use of physical servers.
Both the XenDesktop DDCs and Provisioning Servers where deployed in pairs to provide
redundancy and high availability. No performance gains from load balancing were expected as a
single instance of either of these components could adequately handle the load from the single
server scalability tests.
XenServer Hosts
Two XenServer hosts where deployed, each dedicated to a particular function; the management
server host was used for all infrastructure servers, and the virtual desktop host was used for all
virtual desktops. XenServer 5.6 FP1, the latest release at the time of testing, was used for both
hosts.
NOTE: It was necessary to enable “Remote Console” under Console Redirection Settings in the
BladeCenter HX5 server BIOS in order to install XenServer 5.6 FP1. Not doing so resulted in the
Installation process hanging.
Page 8
For the installation of XenServer, the locally attached solid state drives (SSD) of the host were used.
While XenServer 5.6 FP1 supports boot-from-SAN configurations, it was decided to use local
storage to simplify configurations and dedicate the storage array to virtual machine disk operations.
Storage repositories for virtual machines were created using provisioned LUNs over Hardware
HBA.
To handle network traffic a NIC bond was created using the two available NICs on each host. This
bond was used for both XenServer management traffic and virtual machine network traffic.
To better handle the number of virtual machines, the amount of RAM assigned to Dom0 on the
virtual desktop host was increased to 2.94 GB from the default 752 MB, as recommended by Citrix.
XenDesktop Desktop Delivery Controllers (DDCs)
The laboratory XenDesktop Site consisted of two DDCs. The site’s database was configured using
the locally installed copy of SQL 2008 Express on the first DDC deployed.
In XenDesktop 5, collections of virtual machines (VMs) or physical computers are managed as a
single entity called a catalog. A catalog is a collection of machines of the same type. The machine
type defines the type of hosting infrastructure used for desktops and the level of control that users
have over their desktop environment.
For the tests conducted in this laboratory, the catalog was created using the streamed machine type.
The streamed machine type enables you to deliver desktops to VMs and blade PCs that have been
configured to load the operating system over the network leveraging Citrix Provisioning services.
NOTE: For the purpose of this laboratory, Citrix Provisioning Services was used for provisioning all
virtual desktops used in tests. Citrix Provisioning Services technology has been validated and proven
effective in large scale deployments of XenDesktop. The XenDesktop 5 Machine creation services
(MCS) was not used during testing.
As with previous versions, XenDesktop 5 uses desktop groups to allocate virtual desktops to users
and groups. A single desktop group was created from the streamed machine catalog for the
laboratory tests.
Provisioning Services (PVS)
The Provisioning Services farm consisted of two Provisioning servers. The farm’s database was
created on the same SQL 2008 Express instance used for the XenDesktop Site.
The farm was configured with a single site and device collection for all target devices. Target
devices were created using Provisioning Server’s Auto add feature.
A single vDisk store was configured to create the master virtual desktop image (vDisk) that was used
for all target devices. For the laboratory tests the store was configured using the local hard disk
subsystem of the Provisioning Servers to store the vDisks. This provides the easiest way of
implementing vDisk high availability, suitable for environments with few Provisioning servers such
as this. The vDisks were manually synchronized between the Provisioning Servers.
Page 9
Provisioning Services PXE and TFTP services were used to allow the target devices to boot from
the network. The environment’s DHCP service scope options were configured to provide the target
devices with the required information to connect with the Provisioning Servers and obtain their
assigned vDisk.
Citrix Web Interface
The Citrix Web Interface 5.4 is included with XenDesktop 5 and it is installed by default on each of
the DDCs during the XenDesktop setup wizard. Although both of the DDCs Web Interface sites
were functional, they were not used in any of the tests performed.
For testing purposes, version 5.2 of the Web Interface was used in order to match the LoginVSI
reference configuration provided by Citrix and described in section 3.4.
Citrix Licensing
The functionality of all the components required that valid licenses specific to each product were
available. For the purpose of this laboratory, a XenDesktop Platinum License file was used. This file
included individual licensing for each of the components used.
Additional details of the configuration of the XenDesktop components used are presented on table
5 below.
Component LAB Configuration
Management Server Host XenServer 5.6 FP1
2 x 8 Core Intel L7555 @ 1.86 GHz (32 Logical Cores Total)
64 GB RAM @ 1067 MHz
50 GB SSD Local Storage
550 GB Storage Repository (Hardware HBA)
NIC Bond (2 x 1Gb NICs)
Virtual Desktop Host XenServer 5.6 FP1
2 x 8 Core Intel L7555 @ 1.86 GHz (32 Logical Cores Total)
128 GB RAM @ 1067 MHz
50 GB SSD Local Storage
450 GB Storage Repository (Hardware HBA)
NIC Bond (2 x 1Gb NICs)
XenDesktop – Desktop
Delivery Controllers (DDC)
Windows Server 2008 R2 Standard Edition
XenServer 5.6 VM
2 vCPU
4 GB RAM
24 GB VHD
XenDesktop 5 Controller
XenDesktop 5 Desktop Studio
Citrix Web Interface 5.4
*DDC 1 – SQL 2008 Express (XD and PVS Data Stores)
Provisioning Services / PXE
/ TFTP
Windows Server 2008 R2 Standard Edition
XenServer 5.6 VM
4 vCPU
8 GB RAM
24 GB VHD
30 GB VHD (vDisk Store)
Provisioning Server 5.6 SP1
Page 10
Component LAB Configuration
Citrix Web Interface /
Domain Controller / DNS
/ DHCP
Windows Server 2008 R2 Standard Edition
XenServer 5.6 VM
2 vCPU
4 GB RAM
40 GB VHD
Citrix Web Interface 5.2
Citrix Licensing Server Windows Server 2008 R2 Standard Edition
XenServer 5.6 VM
1 vCPU
1 GB RAM
24 GB VHD
Citrix Licensing Server 11.6.1
XenDesktop Platinum License
Table 5
Virtual Desktops
All of the virtual desktops deployed for testing were based on the Windows 7 Professional operating
system. Desktop virtual machines were configured for a normal user workload following Citrix
guidelines in “XenDesktop Planning Guide: Hosted VM-Based Resource Allocation.” Each virtual
desktop was configured 1 vCPU, 1.5GB of RAM, and a 4GB locally attached VHD for write cache.
To prepare the master image, a new virtual machine was created on the XenServer virtual desktop
host. Windows 7 was installed and updated with all available security patches and updates available
from Microsoft. For the purposes of this test all applications required for testing were installed
locally, not streamed or hosted.
The following additional configurations were also implemented for increased performance:
 All user profiles for the test user accounts were created locally on the master image
according to the “Login VSI 3.0 Admin Guide.”
 All Windows 7 optimizations for XenDesktop and Provisioning were configured following
the “Windows 7 Optimization Guide” provided by Citrix.
 All virtual desktops were configured with a locally attached VHD for PVS Write Cache
 Following Login Consultants and Citrix recommendations, page file settings were modified
to use a fixed size equal to the amount of memory allocated to the virtual machine, and the
location of the page file was set to use the locally attached write cache VHD.
The virtual desktop image was then converted to a Provisioning Services vDisk with the use of the
PVS Imaging Wizard.
Page 11
Table 6 below details the configuration used for all virtual desktops.
Component LAB Configuration
Virtual Desktop Windows 7 Professional Edition with SP1
XenServer 5.6 VM
1 vCPU
1.5 GB RAM
25 GB vDisk
4 GB VHD (Write Cache)
MS Office 2010 Professional Plus
Virtual Desktop Agent 5.0.0.148
Provisioning Services Target Device 5.6.1
Adobe Flash Player 10
Adobe Reader 9.1
Adobe Shockwave Player 10.2.0.023
Bullzip PDF Printer 6.0.0.865
FreeMind 0.8.1
Java 6 Update 21
GPL Ghostscript Lite 8.64
Kid-Key-Lock 1.2.1.0
Table 6
Login Consultants VSI
Login Consultants VSI is a tool designed for benchmarking Server Based Computing (SBC) and
Desktop Virtualization (VDI) solutions including Citrix XenDesktop. The free version, VSI 3.0
Express, was used for all tests performed.
VSI consists of 4 components:
 AD Domain controller for user accounts and standard policies
 A file share for central configuration and logging
 Launcher workstations to initiate the sessions
 Target platform (VDI or SBC) where the user load scripts are installed and performed
Domain user accounts and standard policies were automatically configured using the AD
Deployment setup included with VSI 3.0. As part of the setup a new root OU is created, test users
are created and GPOs are imported and linked.
The file share and related permissions were configured manually following the “Login VSI 3.0
Admin Guide.”
A total of five launchers were deployed with capacity configured for twenty sessions each. The
launchers were prepared for integration with XenDesktop following the “Benchmarking Citrix
XenDesktop 4 using Login Consultants VSI 2.1” white paper from Citrix. While some of the
product versions described on the white paper are different from the versions used in this
laboratory, the configurations are still valid.
Page 12
Table 7 below details the configuration of the VSI Launchers.
Component LAB Configuration
VSI Launcher Windows Server 2003 R2 32-Bit Standard Edition
XenServer 5.6 VM
2 vCPU
2 GB RAM
8 GB VHD
Login VSI Launcher 3.0
Login VSI Analyzer 3.0
Citrix Online plug-in 12.1.0.30
Python 2.6.150
Python 2.6 pywin32-214
Table 7
Methodology
All tests were performed using the laboratory environment and reference configuration described on
section 3 of this document. All testing concentrated in determining the optimum amount of
desktops that could be hosted on a single IBM BladeCenter HX5 server. Testing was done in two
stages: pool spin up and user workload. During both stages, data was gathered from the XenServer
hypervisor, XenDesktop infrastructure servers, virtual desktops, and storage to analyze for
performance and resource utilization.
Tests were performed several times to ensure results were consistent from one test to the next.
Once it was determined that results were in fact consistent, a total of three official tests were
executed. The highest and lowest scoring tests were then discarded and the remaining test result
was selected to present the findings in this document.
Pool Spin Up
The pool spin up stage consisted of the process involved in getting all virtual desktops within the
test desktop group available for use from a halted state. In preparation for the test, the desktop
group is put on maintenance mode and all hosted virtual desktops are kept on a halted state within
the XenServer host. To start the test, maintenance mode is disabled for the test desktop group and
all virtual desktops are then started via XenDesktop 5 desktop group power management settings.
As the virtual desktops boot using the provisioned Windows 7 image and register with the Desktop
Delivery Controllers, they become available to users. The test is considered complete once all
virtual desktops within the test desktop group show as available on the XenDesktop 5 desktop
studio management console.
User Workload
The user workload was the most significant stage in validating the XenDesktop 5 solution with the
reference configuration included in this document. The test consisted in generating simulated user
load and measuring response time of the various tasks performed during the simulation. The
process of simulating user load involved all phases of a user session including user logon, work time,
and user logoff.
Page 13
User workload was simulated using the Login Virtual Session Indexer (VSI 3.0) from Login
Consultants. VSI is a benchmarking methodology based on the amount of simultaneous sessions
that can be run on a single machine running either bare metal or virtualized operating systems. VSI
is the standardized benchmark for Server Based Computing (SBC) and Virtual Desktop
Infrastructure (VDI) environments and used by key vendors in the desktop virtualization space.
The tool is 100% platform and protocol independent which allows for easy replication of results
across most environments. The Express version of the VSI 3.0 tool was selected for all testing
performed.
VSI simulates realistic user workloads. The medium workload used for testing simulates an office
worker using generic applications like Office, IE (including Flash applets) and Adobe Acrobat
Reader. Like real users, the scripted session will leave multiple applications open at the same time.
Every session will average about 20% minimal user activity, similar to real world usage.
The medium workload is the only workload available in VSI Express and it is the default workload
in VSI. The following list outlines the automated workflow performed by the medium workload:
 This workload emulated a medium workload user using Office 2010, IE 8 and PDF.
 Once a session has been started the medium workload will repeat every 12 minutes.
 During each loop the response time is measured every 2 minutes.
 The medium workload opens up to 5 applications simultaneously.
 The type rate is 160ms for each character.
 Approximately 2 minutes of idle time is included to simulate real-world users.
Each loop will open and use:
 Outlook 2010 - browse 10 messages.
 Internet Explorer - one instance is left open (BBC.co.uk), one instance is browsed to
Wired.com, Lonelyplanet.com and heavy flash application gettheglass.com.
 Word 2010 - one instance to measure response time, one instance to review and edit
document.
 Bullzip PDF Printer & Acrobat Reader - the Word document is printed and reviewed to
PDF.
 Excel 2010 - a very large randomized sheet is opened.
 PowerPoint 2010 - a presentation is reviewed and edited.
 7-zip - using the command line version the output of the session is zipped.
Sessions were launched at 30 second intervals until the total amount of sessions configured were
started. The test was then allowed to run for about 1 hour to allow each session to go through a
minimum of 2 loops before logoff.
In addition to the user workload simulated by the Login VSI tool, a virtual desktop was given to a
real person for use during testing. This was of great help in getting a feel for what system response
time was like during tests.
Page 14
Monitoring of Solution Components
During all stages of testing data was gathered and analyzed to obtain performance and resource
utilization results of the various components involved.
The following set of tools was used to capture and analyze data:
 Login Consultants VSI 3.0 Express – For capturing user experience data based on system
response time.
 Citrix XenCenter 5.6 FP1 – For real time performance monitoring of the XenServer host.
 Windows Performance Monitor – For monitoring and capturing performance data of the
Desktop Deliver Controllers and Provisioning Servers.
 Netapp DATA ONTAP 7.3.4 – For monitoring and capturing performance data of the
SAN.
 Citrix Virtual Machine Performance Utility (StormVM) – For Disk I/O performance
monitoring of the XenServer host storage repository used for virtual desktop write cache
disks.
Findings
This section describes the findings of the tests performed to validate the XenDesktop 5 solution and
defines the success criteria for a successful test run. Data from test results is presented in graphs
detailing performance and resource utilization of the main components during each of the testing
stages.
Tests results show that with the architecture described in section 3 of this document it was possible
to run up to eighty (80) Windows 7 virtual desktops with 1.5 GB of RAM on a single IBM
BladeCenter HX5 server without scarifying performance and user experience. Pool spin up test
results show that all 80 virtual desktops could be started and made available to users in 12 minutes.
The Intel Nehalem-EX processors proved to be highly efficient under this type of load. Test results
show average CPU utilization between 20% and 40% with 5 virtual desktops per core.
Windows 7 virtual desktops
1 vCPU
1.5 GB RAM
25GB vDisk
4 GB PVS Write Cache
XenServer 5.6 FP1
VMs/Host VMs/Core
IBM BladeCenter HX5
Dual 8 Core (1.86GHz L7555)
128 GB RAM
80 5
Table 8
The factor limiting additional density on the server was available RAM. XenServer reported 98% of
total memory utilization when the 80 virtual desktops where active. The BladeCenter HX5 Server
available at the time of this writing had a maximum capacity of 128GB of RAM per single-wide
blade. However, IBM is set to release a new version of the HX5 which will support up to 256 GB
of RAM and Dual 10 Core processors. These additional resources should allow for greater virtual
desktop density on a server with the same form factor.
Page 15
The IBM System Storage N5200 effectively processed all I/O operations requested by the
XenServer hosts. Citrix guidelines state that a typical virtual desktop on a working state would
generate a load of about 14 IOPS1
and require 125kB/s of disk throughput. Based on these
guidelines, the N5200 was expected to manage a load of approximately 640 (8 x 80) IOPS with 80%
being random write operations and 10MB/s (125kB/s x 80) of disk throughput during the user
workload stage. Test results show the N5200 handling Fiber Channel Protocol (FCP) OPS in the
500-600 range, and disk throughput near 10MB/s, very close to expectations. Results from
additional stress tests conducted, show the N5200 supporting a high of 878 FCP OPS and 75MB/s
FCP write throughput.
As expected, none of the XenDesktop components showed significant resource utilization during
any of the tests conducted. Critical components such as the Desktop Delivery Controllers and
Provisioning Servers have been proven to work under greater loads.
Success Criteria
Success criteria to determine if a test cycle passed or failed was based on the VSI Max score. The
Login VSI Max evaluates the user response time during increasing user load. VSI focuses on how
many users can run on the system before it saturates.
The operations executed by VSI to simulate user load are designed to measure system response time
by completing activities that leverage resources such as CPU, memory, disk, the OS in general,
printing and GDI, among others. These operations are specifically short by nature. When such
operations are consistently long the system is saturated because of excessive queuing on any kind of
resource. As a result, the average response times escalate. This effect is clearly visible to end-users.
When such operations consistently consume multiple seconds the user will regard the system as slow
and unresponsive.
Response time for all tasks performed during the simulated user workload is measured in
milliseconds. The Login VSI 3.0 Express medium workload used for testing has a pre-defined
threshold of 4000ms to determine the VSI Max Pass or Fail rating. Tests that resulted in user
response times above 4000ms were considered failed.
While the VSI Max score was the single factor considered for the pass or fail of any given test, real
user experience was also helpful in validating VSI results. During all tests conducted, a real user was
given access to a virtual desktop in the same manner that the VSI Launchers established
connections. User feedback regarding system response time was then evaluated and used to confirm
the pass or fail rating given by the VSI Max score. While user feedback is a very subjective measure
it is generally used in production environments to determine if the system is performing according
to expectations.
1
During certain events, such as logon storms, IOPS will be higher than these estimated values.
Page 16
Test Results
The following graphs demonstrate results from all data gathered during tests. Results from both
stages of testing, pool spin up and user workload, are presented for the main components involved.
Login VSI
Graph 1 below demonstrates a successful test run where user experience was not affected by the 80
concurrent sessions running on the same host. All response times are below the 4000ms threshold
defined by VSI 3.0 Express.
Graph 1 - Login VSImax / 80 Sessions
Vertical axis: Response Time in milliseconds
Horizontal axis: Total Active Sessions
 Red line: Maximum Response (worst response time of an individual measurement within a
single session)
 Orange line: Average Response Time within for each level of active sessions
 Blue line: The VSImax average (Typically VSImax is reached when CPU Utilization is
around 100%)
 Green line: Minimum Response (best response time of an individual measurement within a
single session)
XenServer
Figure 6 below shows 98% of RAM utilization by the virtual desktop host with 80 active VMs. This
was the limiting factor in the amount of virtual desktops that could be hosted on a single server with
the specified configuration. Graphs 2 and 3 below demonstrate resource utilization by the virtual
desktop host during the Pool Spin Up and User Workload stage. Graph 2 shows cores 0-3 of the
host running at about 80%, these cores are used by XenServer and were expected to be highly
utilized during the startup of all VMs.
Page 17
Figure 6 - XenServer Virtual Desktop Host / Memory Utilization
Graph 2 - XenServer Virtual Desktop Host / Pool Spin Up
Page 18
Graph 3 - XenServer Virtual Desktop Host / User Workload
Page 19
IBM System Storage N5200
The N5200 proved to be more than adequate for a workload of 80 virtual desktops. Test data
confirms that write operations account for approximately 80% of the total IOPS generated by VDIs.
Graphs 4 and 5 below present results from the XenServer Virtual Machine Performance Utility
(StormVM) disk I/O benchmark. The StormVM Disk I/O utility was used to conduct a stress test
consisting of random write operations on the storage repository used for virtual desktop write cache.
The stress test was conducted with virtual desktops being idle and during the user workload stage.
The two spikes seen in graphs 7, 9, 11, and 13 represent the impact of the stress test performed
during the user workload stage.
Graph 4 - StormVM 1MB Random Writes / 80 Desktops IDLE
Graph 5 - StormVM 1MB Random Writes / 80 Desktops Active
Page 20
Graph 6 - Pool Spin UP / N5200 FCP OPS Graph 7 – User Workload / N5200 FCP OPS
Graph 8 - Pool Spin UP / N5200 CPU-Disk Util Graph 9 – User Workload / N5200 CPU-Disk Util
0
50
100
150
200
250
300
350
400
450
500
13:04:00
13:05:45
13:07:30
13:09:15
13:11:00
13:12:45
13:14:30
FCP OPS
0
100
200
300
400
500
600
700
800
900
1000
15:02:00
15:13:30
15:25:00
15:36:30
15:48:00
15:59:30
16:11:00
FCP OPS
0%
2%
4%
6%
8%
10%
12%
13:04:00
13:05:45
13:07:30
13:09:15
13:11:00
13:12:45
13:14:30
CPU Util
Disk Util
0%
10%
20%
30%
40%
50%
60%
70%
80%
15:02:00
15:12:00
15:22:00
15:32:00
15:42:00
15:52:00
16:02:00
16:12:00
CPU Util
Disk Util
Page 21
Graph 10 - Pool Spin UP /
N5200 FCP Throughput
Graph 11 – User Workload /
N5200 FCP Throughput
Graph 12 - Pool Spin UP /
N5200 Disk R/W
Graph 13 – User Workload /
N5200 Disk R/W
0
500
1000
1500
2000
2500
3000
3500
4000
4500
5000
13:04:00
13:06:15
13:08:30
13:10:45
13:13:00
13:15:15
FCP kB/s
in
FCP kB/s
out
0
10000
20000
30000
40000
50000
60000
70000
80000
15:02:00
15:15:30
15:29:00
15:42:30
15:56:00
16:09:30
FCP kB/s in
FCP kB/s
out
0
1000
2000
3000
4000
5000
6000
7000
8000
13:04:00
13:05:45
13:07:30
13:09:15
13:11:00
13:12:45
13:14:30
Disk kB/s
read
Disk kB/s
write
0
10000
20000
30000
40000
50000
60000
70000
80000
90000
100000
15:02:00
15:15:30
15:29:00
15:42:30
15:56:00
16:09:30
Disk kB/s
read
Disk kB/s
write
Page 22
Desktop Delivery Controller
As both DDCs test results were nearly identical, only the data from DDC1 is presented in this
document. Data shows DDC resource utilization was very similar during both tests.
Graph 14 - Pool Spin UP / DDC1 CPU Graph 15 – User Workload / DDC1 CPU
Graph 16 - Pool Spin UP / DDC1 RAM Graph 17 – User Workload / DDC 1 RAM
0
10
20
30
40
50
60
70
80
90
100
%
Processor
Time:_Tot
al (2vCPU)
0
10
20
30
40
50
60
70
80
90
100
%
Processor
Time:_Tot
al (2vCPU)
0
500
1000
1500
2000
2500
3000
3500
4000
MemoryC
ommitted
Megabytes
0
500
1000
1500
2000
2500
3000
3500
4000
MemoryC
ommitted
Megabytes
Page 23
Graph 18 - Pool Spin UP / DDC1 Network Graph 19 - User Workload / DDC1 Network
0
20
40
60
80
100
120
140
Network
Interface:
kB/s
Received
Network
Interface:
kB/s Sent
0
10
20
30
40
50
60
70
Network
Interface:
kB/s
Received
Network
Interface:
kB/s Sent
Page 24
Provisioning Server
As both PVS test results were nearly identical, only the data from PVS1 is presented in this
document. As expected, the Provisioning server CPU and network utilization (Graph 20, 24) was
the highest during the Pool Spin UP test. Grpahs 22 and 23 show that RAM utilization remained
the same throughout both tests indicating the vDisk being cached in system RAM. This is further
confirmed on graphs 26 and 27 where disk utilization remained idle nearly 100% of the time during
both tests.
Graph 20 - Pool Spin UP / PVS1 CPU Graph 21 – User Workload / PVS1 CPU
Graph 22 - Pool Spin UP / PVS1 RAM Graph 23 – User Workload / PVS1 RAM
0
10
20
30
40
50
60
70
80
90
100
13:04:05
13:06:05
13:08:05
13:10:05
13:12:05
13:14:05
13:16:05
%
Processor
Time:_Tot
al (4vCPU)
0
10
20
30
40
50
60
70
80
90
100
%
Processor
Time:_Tot
al (4vCPU)
0
1000
2000
3000
4000
5000
6000
7000
8000
13:04:05
13:06:20
13:08:35
13:10:50
13:13:05
13:15:20
MemoryC
ommitted
Megabytes
0
1000
2000
3000
4000
5000
6000
7000
8000
MemoryC
ommitted
Megabytes
Page 25
Graph 24 - Pool Spin UP / PVS1 Network Graph 25 - User Workload / PVS1 Network
Graph 26 - Pool Spin UP / PVS1 Disk Graph 27 – User Workload / PVS 1 Disk
0
5000
10000
15000
20000
25000
30000
35000
13:04:05
13:06:20
13:08:35
13:10:50
13:13:05
13:15:20
Network
Interface:
kB/s
Received
Network
Interface:
kB/s Sent
0
2000
4000
6000
8000
10000
12000
14000
16000
18000
20000
Network
Interface:
kB/s
Received
Network
Interface:
kB/s Sent
0
10
20
30
40
50
60
70
80
90
100
13:04:05
13:06:05
13:08:05
13:10:05
13:12:05
13:14:05
13:16:05
PhysicalDis
k: % Idle
Time
0
10
20
30
40
50
60
70
80
90
100
PhysicalDis
k: % Idle
Time
Page 26
References
Reference Title
http://support.citrix.com/article/CTX125630 Citrix XenDesktop with Login
Consultants VSI
http://support.citrix.com/article/CTX127050 XenDesktop - Windows 7
Optimization Guide
http://support.citrix.com/article/CTX127277 XenDesktop Planning Guide: Hosted
VM-Based Resource Allocation
http://support.citrix.com/article/CTX128645 Design Considerations for
Virtualizing Provisioning Services
Included with Login VSI 3.0 Express LoginVSI 3.0 Admin Guide
http://support.citrix.com/article/ctx127065 XenServer Virtual Machine
Performance Utility
http://support.citrix.com/proddocs/index.jsp Citrix eDocs
http://support.citrix.com/article/CTX126531 Configuring Dom0 Memory in
XenServer 5.6
.
Page 27
About Citrix
Citrix Systems, Inc. (NASDAQ:CTXS) is the leading provider of virtualization, networking and software as a service
technologies for more than 230,000 organizations worldwide. Its Citrix Delivery Center, Citrix Cloud Center (C3)
and Citrix Online Services product families radically simplify computing for millions of users, delivering applications
as an on-demand service to any user, in any location on any device. Citrix customers include the world’s largest
Internet companies, 99 percent of Fortune Global 500 enterprises, and hundreds of thousands of small businesses
and prosumers worldwide. Citrix partners with over 10,000 companies worldwide in more than 100 countries.
Founded in 1989, annual revenue in 2008 was $1.6 billion.
©2009 Citrix Systems, Inc. All rights reserved. Citrix®, Access Gateway™, Branch Repeater™, Citrix Repeater™,
HDX™, XenServer™, XenApp™, XenDesktop™ and Citrix Delivery Center™ are trademarks of Citrix Systems, Inc.
and/or one or more of its subsidiaries, and may be registered in the United States Patent and Trademark Office
and in other countries. All other trademarks and registered trademarks are property of their respective owners.

More Related Content

What's hot

Monitoring Windows Server Systems Demo Setup
Monitoring Windows Server Systems Demo SetupMonitoring Windows Server Systems Demo Setup
Monitoring Windows Server Systems Demo SetupMicrosoft TechNet
 
50,000-seat_VMware_view_deployment
50,000-seat_VMware_view_deployment50,000-seat_VMware_view_deployment
50,000-seat_VMware_view_deploymentAccenture
 
VDI performance comparison: Dell PowerEdge FX2 and FC430 servers with VMware ...
VDI performance comparison: Dell PowerEdge FX2 and FC430 servers with VMware ...VDI performance comparison: Dell PowerEdge FX2 and FC430 servers with VMware ...
VDI performance comparison: Dell PowerEdge FX2 and FC430 servers with VMware ...Principled Technologies
 
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...Better email response time using Microsoft Exchange 2013 with the Dell PowerE...
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...Principled Technologies
 
Oracle Exadata 1Z0-485 Certification
Oracle Exadata 1Z0-485 CertificationOracle Exadata 1Z0-485 Certification
Oracle Exadata 1Z0-485 CertificationExadatadba
 
50,000 Seat V Mware View Deployment
50,000 Seat V Mware View Deployment50,000 Seat V Mware View Deployment
50,000 Seat V Mware View DeploymentMichael Hudak
 
Windows server 2012 r2 Hyper-v Component architecture
Windows server 2012 r2 Hyper-v Component architecture Windows server 2012 r2 Hyper-v Component architecture
Windows server 2012 r2 Hyper-v Component architecture Tũi Wichets
 
Consolidating older database servers onto Dell PowerEdge FX2 with FC830 serve...
Consolidating older database servers onto Dell PowerEdge FX2 with FC830 serve...Consolidating older database servers onto Dell PowerEdge FX2 with FC830 serve...
Consolidating older database servers onto Dell PowerEdge FX2 with FC830 serve...Principled Technologies
 
Introduction to storage technologies
Introduction to storage technologiesIntroduction to storage technologies
Introduction to storage technologiesNuno Alves
 
Exadata Patching Demystified
Exadata Patching DemystifiedExadata Patching Demystified
Exadata Patching DemystifiedEnkitec
 
Hyper v and failover clustering mini poster
Hyper v and failover clustering mini posterHyper v and failover clustering mini poster
Hyper v and failover clustering mini posterChawin Patimapornchai
 
Xen server storage Overview
Xen server storage OverviewXen server storage Overview
Xen server storage OverviewNuno Alves
 
Vsphere esxi-vcenter-server-55-setup-mscs
Vsphere esxi-vcenter-server-55-setup-mscsVsphere esxi-vcenter-server-55-setup-mscs
Vsphere esxi-vcenter-server-55-setup-mscsDhymas Mahendra
 
download it from here
download it from heredownload it from here
download it from herewebhostingguy
 
Data Protector 9.07 what is new
Data Protector 9.07 what is new Data Protector 9.07 what is new
Data Protector 9.07 what is new Andrey Karpov
 
Optimize Oracle On VMware (Sep 2011)
Optimize Oracle On VMware (Sep 2011)Optimize Oracle On VMware (Sep 2011)
Optimize Oracle On VMware (Sep 2011)Guy Harrison
 
Optimize oracle on VMware (April 2011)
Optimize oracle on VMware (April 2011)Optimize oracle on VMware (April 2011)
Optimize oracle on VMware (April 2011)Guy Harrison
 
vSAN Beyond The Basics
vSAN Beyond The BasicsvSAN Beyond The Basics
vSAN Beyond The BasicsSumit Lahiri
 
Cascade lake-advanced-performance-press-deck
Cascade lake-advanced-performance-press-deckCascade lake-advanced-performance-press-deck
Cascade lake-advanced-performance-press-deckDESMOND YUEN
 

What's hot (20)

Monitoring Windows Server Systems Demo Setup
Monitoring Windows Server Systems Demo SetupMonitoring Windows Server Systems Demo Setup
Monitoring Windows Server Systems Demo Setup
 
50,000-seat_VMware_view_deployment
50,000-seat_VMware_view_deployment50,000-seat_VMware_view_deployment
50,000-seat_VMware_view_deployment
 
VDI performance comparison: Dell PowerEdge FX2 and FC430 servers with VMware ...
VDI performance comparison: Dell PowerEdge FX2 and FC430 servers with VMware ...VDI performance comparison: Dell PowerEdge FX2 and FC430 servers with VMware ...
VDI performance comparison: Dell PowerEdge FX2 and FC430 servers with VMware ...
 
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...Better email response time using Microsoft Exchange 2013 with the Dell PowerE...
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...
 
Oracle Exadata 1Z0-485 Certification
Oracle Exadata 1Z0-485 CertificationOracle Exadata 1Z0-485 Certification
Oracle Exadata 1Z0-485 Certification
 
50,000 Seat V Mware View Deployment
50,000 Seat V Mware View Deployment50,000 Seat V Mware View Deployment
50,000 Seat V Mware View Deployment
 
Windows server 2012 r2 Hyper-v Component architecture
Windows server 2012 r2 Hyper-v Component architecture Windows server 2012 r2 Hyper-v Component architecture
Windows server 2012 r2 Hyper-v Component architecture
 
Consolidating older database servers onto Dell PowerEdge FX2 with FC830 serve...
Consolidating older database servers onto Dell PowerEdge FX2 with FC830 serve...Consolidating older database servers onto Dell PowerEdge FX2 with FC830 serve...
Consolidating older database servers onto Dell PowerEdge FX2 with FC830 serve...
 
Introduction to storage technologies
Introduction to storage technologiesIntroduction to storage technologies
Introduction to storage technologies
 
Exadata Patching Demystified
Exadata Patching DemystifiedExadata Patching Demystified
Exadata Patching Demystified
 
Hyper v and failover clustering mini poster
Hyper v and failover clustering mini posterHyper v and failover clustering mini poster
Hyper v and failover clustering mini poster
 
Xen server storage Overview
Xen server storage OverviewXen server storage Overview
Xen server storage Overview
 
Vsphere esxi-vcenter-server-55-setup-mscs
Vsphere esxi-vcenter-server-55-setup-mscsVsphere esxi-vcenter-server-55-setup-mscs
Vsphere esxi-vcenter-server-55-setup-mscs
 
distcom-short-20140112-1600
distcom-short-20140112-1600distcom-short-20140112-1600
distcom-short-20140112-1600
 
download it from here
download it from heredownload it from here
download it from here
 
Data Protector 9.07 what is new
Data Protector 9.07 what is new Data Protector 9.07 what is new
Data Protector 9.07 what is new
 
Optimize Oracle On VMware (Sep 2011)
Optimize Oracle On VMware (Sep 2011)Optimize Oracle On VMware (Sep 2011)
Optimize Oracle On VMware (Sep 2011)
 
Optimize oracle on VMware (April 2011)
Optimize oracle on VMware (April 2011)Optimize oracle on VMware (April 2011)
Optimize oracle on VMware (April 2011)
 
vSAN Beyond The Basics
vSAN Beyond The BasicsvSAN Beyond The Basics
vSAN Beyond The Basics
 
Cascade lake-advanced-performance-press-deck
Cascade lake-advanced-performance-press-deckCascade lake-advanced-performance-press-deck
Cascade lake-advanced-performance-press-deck
 

Viewers also liked

Introducing Xen Server
Introducing Xen ServerIntroducing Xen Server
Introducing Xen ServerStephenRice86
 
Xen server 6.1 technical sales presentation
Xen server 6.1 technical sales presentationXen server 6.1 technical sales presentation
Xen server 6.1 technical sales presentationNuno Alves
 
vRealize Operations (vROps) Management Pack for Citrix XenDesktop Installatio...
vRealize Operations (vROps) Management Pack for Citrix XenDesktop Installatio...vRealize Operations (vROps) Management Pack for Citrix XenDesktop Installatio...
vRealize Operations (vROps) Management Pack for Citrix XenDesktop Installatio...Blue Medora
 
Citrix Store front planning guide
Citrix Store front planning guideCitrix Store front planning guide
Citrix Store front planning guideNuno Alves
 
Citrix Xen Server 5 5
Citrix Xen Server 5 5Citrix Xen Server 5 5
Citrix Xen Server 5 5anterope
 
Citrix XenServer 5.5 Troubleshooting
Citrix XenServer 5.5 TroubleshootingCitrix XenServer 5.5 Troubleshooting
Citrix XenServer 5.5 TroubleshootingThomas Krampe
 
XenServer, Hyper-V, and ESXi - Architecture, API, and Coding
XenServer, Hyper-V, and ESXi -  Architecture, API, and CodingXenServer, Hyper-V, and ESXi -  Architecture, API, and Coding
XenServer, Hyper-V, and ESXi - Architecture, API, and Coding_Humair_Ahmed_
 
Hypervisors and Virtualization - VMware, Hyper-V, XenServer, and KVM
Hypervisors and Virtualization - VMware, Hyper-V, XenServer, and KVMHypervisors and Virtualization - VMware, Hyper-V, XenServer, and KVM
Hypervisors and Virtualization - VMware, Hyper-V, XenServer, and KVMvwchu
 
Slideshare Project
Slideshare ProjectSlideshare Project
Slideshare Projectsuzrocky
 
Complete Guide to the Citrix Logon Process
Complete Guide to the Citrix Logon ProcessComplete Guide to the Citrix Logon Process
Complete Guide to the Citrix Logon ProcessJohn Grant
 

Viewers also liked (10)

Introducing Xen Server
Introducing Xen ServerIntroducing Xen Server
Introducing Xen Server
 
Xen server 6.1 technical sales presentation
Xen server 6.1 technical sales presentationXen server 6.1 technical sales presentation
Xen server 6.1 technical sales presentation
 
vRealize Operations (vROps) Management Pack for Citrix XenDesktop Installatio...
vRealize Operations (vROps) Management Pack for Citrix XenDesktop Installatio...vRealize Operations (vROps) Management Pack for Citrix XenDesktop Installatio...
vRealize Operations (vROps) Management Pack for Citrix XenDesktop Installatio...
 
Citrix Store front planning guide
Citrix Store front planning guideCitrix Store front planning guide
Citrix Store front planning guide
 
Citrix Xen Server 5 5
Citrix Xen Server 5 5Citrix Xen Server 5 5
Citrix Xen Server 5 5
 
Citrix XenServer 5.5 Troubleshooting
Citrix XenServer 5.5 TroubleshootingCitrix XenServer 5.5 Troubleshooting
Citrix XenServer 5.5 Troubleshooting
 
XenServer, Hyper-V, and ESXi - Architecture, API, and Coding
XenServer, Hyper-V, and ESXi -  Architecture, API, and CodingXenServer, Hyper-V, and ESXi -  Architecture, API, and Coding
XenServer, Hyper-V, and ESXi - Architecture, API, and Coding
 
Hypervisors and Virtualization - VMware, Hyper-V, XenServer, and KVM
Hypervisors and Virtualization - VMware, Hyper-V, XenServer, and KVMHypervisors and Virtualization - VMware, Hyper-V, XenServer, and KVM
Hypervisors and Virtualization - VMware, Hyper-V, XenServer, and KVM
 
Slideshare Project
Slideshare ProjectSlideshare Project
Slideshare Project
 
Complete Guide to the Citrix Logon Process
Complete Guide to the Citrix Logon ProcessComplete Guide to the Citrix Logon Process
Complete Guide to the Citrix Logon Process
 

Similar to Citrix XenDesktop on XenServer and IBM Hardware Sizing Baseline and Reference Configuration

Hitachi whitepaper-protect-ucp-hc-v240-with-vmware-vsphere-hdid
Hitachi whitepaper-protect-ucp-hc-v240-with-vmware-vsphere-hdidHitachi whitepaper-protect-ucp-hc-v240-with-vmware-vsphere-hdid
Hitachi whitepaper-protect-ucp-hc-v240-with-vmware-vsphere-hdidChetan Gabhane
 
Diretrizes para Implementação do Citrix XenServer 6.2.0 em Servidores HP Prol...
Diretrizes para Implementação do Citrix XenServer 6.2.0 em Servidores HP Prol...Diretrizes para Implementação do Citrix XenServer 6.2.0 em Servidores HP Prol...
Diretrizes para Implementação do Citrix XenServer 6.2.0 em Servidores HP Prol...Lorscheider Santiago
 
Windows Server 2008 R2 Dev Session 01
Windows Server 2008 R2 Dev Session 01Windows Server 2008 R2 Dev Session 01
Windows Server 2008 R2 Dev Session 01Clint Edmonson
 
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SAN
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SANVMmark virtualization performance of Micron Enterprise PCIe SSD-based SAN
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SANPrincipled Technologies
 
WebSphere Portal Version 6.0 Web Content Management and DB2 Tuning Guide
WebSphere Portal Version 6.0 Web Content Management and DB2 Tuning GuideWebSphere Portal Version 6.0 Web Content Management and DB2 Tuning Guide
WebSphere Portal Version 6.0 Web Content Management and DB2 Tuning GuideTan Nguyen Phi
 
Xendesktop to_branch_offices_v2 (1)
Xendesktop to_branch_offices_v2 (1)Xendesktop to_branch_offices_v2 (1)
Xendesktop to_branch_offices_v2 (1)Rajnish kumar
 
Red hat enterprise_linux-5.5-release_notes-en-us
Red hat enterprise_linux-5.5-release_notes-en-usRed hat enterprise_linux-5.5-release_notes-en-us
Red hat enterprise_linux-5.5-release_notes-en-usDuong Hieu
 
Reference architecture dir and es - final
Reference architecture   dir and es - finalReference architecture   dir and es - final
Reference architecture dir and es - finalNuno Alves
 
Achieving maximum performance in microsoft vdi environments - Jeff Stokes
Achieving maximum performance in microsoft vdi environments - Jeff StokesAchieving maximum performance in microsoft vdi environments - Jeff Stokes
Achieving maximum performance in microsoft vdi environments - Jeff StokesJeff Stokes
 
IBM SmartCloud Desktop Infrastructure with Citrix XenDesktop
IBM SmartCloud Desktop Infrastructure with Citrix XenDesktopIBM SmartCloud Desktop Infrastructure with Citrix XenDesktop
IBM SmartCloud Desktop Infrastructure with Citrix XenDesktopIBM India Smarter Computing
 
Red hat Enterprise Linux 6.4 for IBM System z Technical Highlights
Red hat Enterprise Linux 6.4 for IBM System z Technical HighlightsRed hat Enterprise Linux 6.4 for IBM System z Technical Highlights
Red hat Enterprise Linux 6.4 for IBM System z Technical HighlightsFilipe Miranda
 
V2V migration to Red Hat Enterprise Virtualization on the Dell PowerEdge R820
V2V migration to Red Hat Enterprise Virtualization on the Dell PowerEdge R820V2V migration to Red Hat Enterprise Virtualization on the Dell PowerEdge R820
V2V migration to Red Hat Enterprise Virtualization on the Dell PowerEdge R820Principled Technologies
 
Microsoft (Virtualization 2008)
Microsoft (Virtualization 2008)Microsoft (Virtualization 2008)
Microsoft (Virtualization 2008)Vinayak Hegde
 
Creating Virtual Infrastructure
Creating Virtual InfrastructureCreating Virtual Infrastructure
Creating Virtual InfrastructureJake Weston
 
Accelerating Virtual Machine Access with the Storage Performance Development ...
Accelerating Virtual Machine Access with the Storage Performance Development ...Accelerating Virtual Machine Access with the Storage Performance Development ...
Accelerating Virtual Machine Access with the Storage Performance Development ...Michelle Holley
 
IBM Flex System Solution for Microsoft Hyper-V (2-node) Reference Architecture
IBM Flex System Solution for Microsoft Hyper-V (2-node) Reference ArchitectureIBM Flex System Solution for Microsoft Hyper-V (2-node) Reference Architecture
IBM Flex System Solution for Microsoft Hyper-V (2-node) Reference ArchitectureIBM India Smarter Computing
 
M|18 Intel and MariaDB: Strategic Collaboration to Enhance MariaDB Functional...
M|18 Intel and MariaDB: Strategic Collaboration to Enhance MariaDB Functional...M|18 Intel and MariaDB: Strategic Collaboration to Enhance MariaDB Functional...
M|18 Intel and MariaDB: Strategic Collaboration to Enhance MariaDB Functional...MariaDB plc
 
Virtual Design Master Challenge 1 - Akmal
Virtual Design Master Challenge 1  - AkmalVirtual Design Master Challenge 1  - Akmal
Virtual Design Master Challenge 1 - Akmaltovmug
 
Private cloud server virtualization
Private cloud server virtualization Private cloud server virtualization
Private cloud server virtualization Pierre-Juan Labeyrie
 

Similar to Citrix XenDesktop on XenServer and IBM Hardware Sizing Baseline and Reference Configuration (20)

Hitachi whitepaper-protect-ucp-hc-v240-with-vmware-vsphere-hdid
Hitachi whitepaper-protect-ucp-hc-v240-with-vmware-vsphere-hdidHitachi whitepaper-protect-ucp-hc-v240-with-vmware-vsphere-hdid
Hitachi whitepaper-protect-ucp-hc-v240-with-vmware-vsphere-hdid
 
Diretrizes para Implementação do Citrix XenServer 6.2.0 em Servidores HP Prol...
Diretrizes para Implementação do Citrix XenServer 6.2.0 em Servidores HP Prol...Diretrizes para Implementação do Citrix XenServer 6.2.0 em Servidores HP Prol...
Diretrizes para Implementação do Citrix XenServer 6.2.0 em Servidores HP Prol...
 
Windows Server 2008 R2 Dev Session 01
Windows Server 2008 R2 Dev Session 01Windows Server 2008 R2 Dev Session 01
Windows Server 2008 R2 Dev Session 01
 
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SAN
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SANVMmark virtualization performance of Micron Enterprise PCIe SSD-based SAN
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SAN
 
WebSphere Portal Version 6.0 Web Content Management and DB2 Tuning Guide
WebSphere Portal Version 6.0 Web Content Management and DB2 Tuning GuideWebSphere Portal Version 6.0 Web Content Management and DB2 Tuning Guide
WebSphere Portal Version 6.0 Web Content Management and DB2 Tuning Guide
 
Xendesktop to_branch_offices_v2 (1)
Xendesktop to_branch_offices_v2 (1)Xendesktop to_branch_offices_v2 (1)
Xendesktop to_branch_offices_v2 (1)
 
Red hat enterprise_linux-5.5-release_notes-en-us
Red hat enterprise_linux-5.5-release_notes-en-usRed hat enterprise_linux-5.5-release_notes-en-us
Red hat enterprise_linux-5.5-release_notes-en-us
 
Reference architecture dir and es - final
Reference architecture   dir and es - finalReference architecture   dir and es - final
Reference architecture dir and es - final
 
Achieving maximum performance in microsoft vdi environments - Jeff Stokes
Achieving maximum performance in microsoft vdi environments - Jeff StokesAchieving maximum performance in microsoft vdi environments - Jeff Stokes
Achieving maximum performance in microsoft vdi environments - Jeff Stokes
 
IBM SmartCloud Desktop Infrastructure with Citrix XenDesktop
IBM SmartCloud Desktop Infrastructure with Citrix XenDesktopIBM SmartCloud Desktop Infrastructure with Citrix XenDesktop
IBM SmartCloud Desktop Infrastructure with Citrix XenDesktop
 
Red hat Enterprise Linux 6.4 for IBM System z Technical Highlights
Red hat Enterprise Linux 6.4 for IBM System z Technical HighlightsRed hat Enterprise Linux 6.4 for IBM System z Technical Highlights
Red hat Enterprise Linux 6.4 for IBM System z Technical Highlights
 
V2V migration to Red Hat Enterprise Virtualization on the Dell PowerEdge R820
V2V migration to Red Hat Enterprise Virtualization on the Dell PowerEdge R820V2V migration to Red Hat Enterprise Virtualization on the Dell PowerEdge R820
V2V migration to Red Hat Enterprise Virtualization on the Dell PowerEdge R820
 
Microsoft (Virtualization 2008)
Microsoft (Virtualization 2008)Microsoft (Virtualization 2008)
Microsoft (Virtualization 2008)
 
IBM Flex System x240 Compute Node (E5-2600 v2)
IBM Flex System x240 Compute Node (E5-2600 v2)IBM Flex System x240 Compute Node (E5-2600 v2)
IBM Flex System x240 Compute Node (E5-2600 v2)
 
Creating Virtual Infrastructure
Creating Virtual InfrastructureCreating Virtual Infrastructure
Creating Virtual Infrastructure
 
Accelerating Virtual Machine Access with the Storage Performance Development ...
Accelerating Virtual Machine Access with the Storage Performance Development ...Accelerating Virtual Machine Access with the Storage Performance Development ...
Accelerating Virtual Machine Access with the Storage Performance Development ...
 
IBM Flex System Solution for Microsoft Hyper-V (2-node) Reference Architecture
IBM Flex System Solution for Microsoft Hyper-V (2-node) Reference ArchitectureIBM Flex System Solution for Microsoft Hyper-V (2-node) Reference Architecture
IBM Flex System Solution for Microsoft Hyper-V (2-node) Reference Architecture
 
M|18 Intel and MariaDB: Strategic Collaboration to Enhance MariaDB Functional...
M|18 Intel and MariaDB: Strategic Collaboration to Enhance MariaDB Functional...M|18 Intel and MariaDB: Strategic Collaboration to Enhance MariaDB Functional...
M|18 Intel and MariaDB: Strategic Collaboration to Enhance MariaDB Functional...
 
Virtual Design Master Challenge 1 - Akmal
Virtual Design Master Challenge 1  - AkmalVirtual Design Master Challenge 1  - Akmal
Virtual Design Master Challenge 1 - Akmal
 
Private cloud server virtualization
Private cloud server virtualization Private cloud server virtualization
Private cloud server virtualization
 

More from IBM India Smarter Computing

Using the IBM XIV Storage System in OpenStack Cloud Environments
Using the IBM XIV Storage System in OpenStack Cloud Environments Using the IBM XIV Storage System in OpenStack Cloud Environments
Using the IBM XIV Storage System in OpenStack Cloud Environments IBM India Smarter Computing
 
TSL03104USEN Exploring VMware vSphere Storage API for Array Integration on th...
TSL03104USEN Exploring VMware vSphere Storage API for Array Integration on th...TSL03104USEN Exploring VMware vSphere Storage API for Array Integration on th...
TSL03104USEN Exploring VMware vSphere Storage API for Array Integration on th...IBM India Smarter Computing
 
A Comparison of PowerVM and Vmware Virtualization Performance
A Comparison of PowerVM and Vmware Virtualization PerformanceA Comparison of PowerVM and Vmware Virtualization Performance
A Comparison of PowerVM and Vmware Virtualization PerformanceIBM India Smarter Computing
 
IBM pureflex system and vmware vcloud enterprise suite reference architecture
IBM pureflex system and vmware vcloud enterprise suite reference architectureIBM pureflex system and vmware vcloud enterprise suite reference architecture
IBM pureflex system and vmware vcloud enterprise suite reference architectureIBM India Smarter Computing
 

More from IBM India Smarter Computing (20)

Using the IBM XIV Storage System in OpenStack Cloud Environments
Using the IBM XIV Storage System in OpenStack Cloud Environments Using the IBM XIV Storage System in OpenStack Cloud Environments
Using the IBM XIV Storage System in OpenStack Cloud Environments
 
All-flash Needs End to End Storage Efficiency
All-flash Needs End to End Storage EfficiencyAll-flash Needs End to End Storage Efficiency
All-flash Needs End to End Storage Efficiency
 
TSL03104USEN Exploring VMware vSphere Storage API for Array Integration on th...
TSL03104USEN Exploring VMware vSphere Storage API for Array Integration on th...TSL03104USEN Exploring VMware vSphere Storage API for Array Integration on th...
TSL03104USEN Exploring VMware vSphere Storage API for Array Integration on th...
 
IBM FlashSystem 840 Product Guide
IBM FlashSystem 840 Product GuideIBM FlashSystem 840 Product Guide
IBM FlashSystem 840 Product Guide
 
IBM System x3250 M5
IBM System x3250 M5IBM System x3250 M5
IBM System x3250 M5
 
IBM NeXtScale nx360 M4
IBM NeXtScale nx360 M4IBM NeXtScale nx360 M4
IBM NeXtScale nx360 M4
 
IBM System x3650 M4 HD
IBM System x3650 M4 HDIBM System x3650 M4 HD
IBM System x3650 M4 HD
 
IBM System x3300 M4
IBM System x3300 M4IBM System x3300 M4
IBM System x3300 M4
 
IBM System x iDataPlex dx360 M4
IBM System x iDataPlex dx360 M4IBM System x iDataPlex dx360 M4
IBM System x iDataPlex dx360 M4
 
IBM System x3500 M4
IBM System x3500 M4IBM System x3500 M4
IBM System x3500 M4
 
IBM System x3550 M4
IBM System x3550 M4IBM System x3550 M4
IBM System x3550 M4
 
IBM System x3650 M4
IBM System x3650 M4IBM System x3650 M4
IBM System x3650 M4
 
IBM System x3500 M3
IBM System x3500 M3IBM System x3500 M3
IBM System x3500 M3
 
IBM System x3400 M3
IBM System x3400 M3IBM System x3400 M3
IBM System x3400 M3
 
IBM System x3250 M3
IBM System x3250 M3IBM System x3250 M3
IBM System x3250 M3
 
IBM System x3200 M3
IBM System x3200 M3IBM System x3200 M3
IBM System x3200 M3
 
IBM PowerVC Introduction and Configuration
IBM PowerVC Introduction and ConfigurationIBM PowerVC Introduction and Configuration
IBM PowerVC Introduction and Configuration
 
A Comparison of PowerVM and Vmware Virtualization Performance
A Comparison of PowerVM and Vmware Virtualization PerformanceA Comparison of PowerVM and Vmware Virtualization Performance
A Comparison of PowerVM and Vmware Virtualization Performance
 
IBM pureflex system and vmware vcloud enterprise suite reference architecture
IBM pureflex system and vmware vcloud enterprise suite reference architectureIBM pureflex system and vmware vcloud enterprise suite reference architecture
IBM pureflex system and vmware vcloud enterprise suite reference architecture
 
X6: The sixth generation of EXA Technology
X6: The sixth generation of EXA TechnologyX6: The sixth generation of EXA Technology
X6: The sixth generation of EXA Technology
 

Recently uploaded

Computer 10: Lesson 10 - Online Crimes and Hazards
Computer 10: Lesson 10 - Online Crimes and HazardsComputer 10: Lesson 10 - Online Crimes and Hazards
Computer 10: Lesson 10 - Online Crimes and HazardsSeth Reyes
 
IaC & GitOps in a Nutshell - a FridayInANuthshell Episode.pdf
IaC & GitOps in a Nutshell - a FridayInANuthshell Episode.pdfIaC & GitOps in a Nutshell - a FridayInANuthshell Episode.pdf
IaC & GitOps in a Nutshell - a FridayInANuthshell Episode.pdfDaniel Santiago Silva Capera
 
AI You Can Trust - Ensuring Success with Data Integrity Webinar
AI You Can Trust - Ensuring Success with Data Integrity WebinarAI You Can Trust - Ensuring Success with Data Integrity Webinar
AI You Can Trust - Ensuring Success with Data Integrity WebinarPrecisely
 
Igniting Next Level Productivity with AI-Infused Data Integration Workflows
Igniting Next Level Productivity with AI-Infused Data Integration WorkflowsIgniting Next Level Productivity with AI-Infused Data Integration Workflows
Igniting Next Level Productivity with AI-Infused Data Integration WorkflowsSafe Software
 
Building Your Own AI Instance (TBLC AI )
Building Your Own AI Instance (TBLC AI )Building Your Own AI Instance (TBLC AI )
Building Your Own AI Instance (TBLC AI )Brian Pichman
 
Comparing Sidecar-less Service Mesh from Cilium and Istio
Comparing Sidecar-less Service Mesh from Cilium and IstioComparing Sidecar-less Service Mesh from Cilium and Istio
Comparing Sidecar-less Service Mesh from Cilium and IstioChristian Posta
 
NIST Cybersecurity Framework (CSF) 2.0 Workshop
NIST Cybersecurity Framework (CSF) 2.0 WorkshopNIST Cybersecurity Framework (CSF) 2.0 Workshop
NIST Cybersecurity Framework (CSF) 2.0 WorkshopBachir Benyammi
 
UiPath Solutions Management Preview - Northern CA Chapter - March 22.pdf
UiPath Solutions Management Preview - Northern CA Chapter - March 22.pdfUiPath Solutions Management Preview - Northern CA Chapter - March 22.pdf
UiPath Solutions Management Preview - Northern CA Chapter - March 22.pdfDianaGray10
 
Salesforce Miami User Group Event - 1st Quarter 2024
Salesforce Miami User Group Event - 1st Quarter 2024Salesforce Miami User Group Event - 1st Quarter 2024
Salesforce Miami User Group Event - 1st Quarter 2024SkyPlanner
 
9 Steps For Building Winning Founding Team
9 Steps For Building Winning Founding Team9 Steps For Building Winning Founding Team
9 Steps For Building Winning Founding TeamAdam Moalla
 
Basic Building Blocks of Internet of Things.
Basic Building Blocks of Internet of Things.Basic Building Blocks of Internet of Things.
Basic Building Blocks of Internet of Things.YounusS2
 
UiPath Studio Web workshop series - Day 6
UiPath Studio Web workshop series - Day 6UiPath Studio Web workshop series - Day 6
UiPath Studio Web workshop series - Day 6DianaGray10
 
UiPath Studio Web workshop series - Day 7
UiPath Studio Web workshop series - Day 7UiPath Studio Web workshop series - Day 7
UiPath Studio Web workshop series - Day 7DianaGray10
 
Meet the new FSP 3000 M-Flex800™
Meet the new FSP 3000 M-Flex800™Meet the new FSP 3000 M-Flex800™
Meet the new FSP 3000 M-Flex800™Adtran
 
Linked Data in Production: Moving Beyond Ontologies
Linked Data in Production: Moving Beyond OntologiesLinked Data in Production: Moving Beyond Ontologies
Linked Data in Production: Moving Beyond OntologiesDavid Newbury
 
Building AI-Driven Apps Using Semantic Kernel.pptx
Building AI-Driven Apps Using Semantic Kernel.pptxBuilding AI-Driven Apps Using Semantic Kernel.pptx
Building AI-Driven Apps Using Semantic Kernel.pptxUdaiappa Ramachandran
 
activity_diagram_combine_v4_20190827.pdfactivity_diagram_combine_v4_20190827.pdf
activity_diagram_combine_v4_20190827.pdfactivity_diagram_combine_v4_20190827.pdfactivity_diagram_combine_v4_20190827.pdfactivity_diagram_combine_v4_20190827.pdf
activity_diagram_combine_v4_20190827.pdfactivity_diagram_combine_v4_20190827.pdfJamie (Taka) Wang
 

Recently uploaded (20)

Computer 10: Lesson 10 - Online Crimes and Hazards
Computer 10: Lesson 10 - Online Crimes and HazardsComputer 10: Lesson 10 - Online Crimes and Hazards
Computer 10: Lesson 10 - Online Crimes and Hazards
 
IaC & GitOps in a Nutshell - a FridayInANuthshell Episode.pdf
IaC & GitOps in a Nutshell - a FridayInANuthshell Episode.pdfIaC & GitOps in a Nutshell - a FridayInANuthshell Episode.pdf
IaC & GitOps in a Nutshell - a FridayInANuthshell Episode.pdf
 
AI You Can Trust - Ensuring Success with Data Integrity Webinar
AI You Can Trust - Ensuring Success with Data Integrity WebinarAI You Can Trust - Ensuring Success with Data Integrity Webinar
AI You Can Trust - Ensuring Success with Data Integrity Webinar
 
Igniting Next Level Productivity with AI-Infused Data Integration Workflows
Igniting Next Level Productivity with AI-Infused Data Integration WorkflowsIgniting Next Level Productivity with AI-Infused Data Integration Workflows
Igniting Next Level Productivity with AI-Infused Data Integration Workflows
 
Building Your Own AI Instance (TBLC AI )
Building Your Own AI Instance (TBLC AI )Building Your Own AI Instance (TBLC AI )
Building Your Own AI Instance (TBLC AI )
 
Comparing Sidecar-less Service Mesh from Cilium and Istio
Comparing Sidecar-less Service Mesh from Cilium and IstioComparing Sidecar-less Service Mesh from Cilium and Istio
Comparing Sidecar-less Service Mesh from Cilium and Istio
 
NIST Cybersecurity Framework (CSF) 2.0 Workshop
NIST Cybersecurity Framework (CSF) 2.0 WorkshopNIST Cybersecurity Framework (CSF) 2.0 Workshop
NIST Cybersecurity Framework (CSF) 2.0 Workshop
 
UiPath Solutions Management Preview - Northern CA Chapter - March 22.pdf
UiPath Solutions Management Preview - Northern CA Chapter - March 22.pdfUiPath Solutions Management Preview - Northern CA Chapter - March 22.pdf
UiPath Solutions Management Preview - Northern CA Chapter - March 22.pdf
 
Salesforce Miami User Group Event - 1st Quarter 2024
Salesforce Miami User Group Event - 1st Quarter 2024Salesforce Miami User Group Event - 1st Quarter 2024
Salesforce Miami User Group Event - 1st Quarter 2024
 
201610817 - edge part1
201610817 - edge part1201610817 - edge part1
201610817 - edge part1
 
9 Steps For Building Winning Founding Team
9 Steps For Building Winning Founding Team9 Steps For Building Winning Founding Team
9 Steps For Building Winning Founding Team
 
Basic Building Blocks of Internet of Things.
Basic Building Blocks of Internet of Things.Basic Building Blocks of Internet of Things.
Basic Building Blocks of Internet of Things.
 
UiPath Studio Web workshop series - Day 6
UiPath Studio Web workshop series - Day 6UiPath Studio Web workshop series - Day 6
UiPath Studio Web workshop series - Day 6
 
UiPath Studio Web workshop series - Day 7
UiPath Studio Web workshop series - Day 7UiPath Studio Web workshop series - Day 7
UiPath Studio Web workshop series - Day 7
 
Meet the new FSP 3000 M-Flex800™
Meet the new FSP 3000 M-Flex800™Meet the new FSP 3000 M-Flex800™
Meet the new FSP 3000 M-Flex800™
 
20230104 - machine vision
20230104 - machine vision20230104 - machine vision
20230104 - machine vision
 
20150722 - AGV
20150722 - AGV20150722 - AGV
20150722 - AGV
 
Linked Data in Production: Moving Beyond Ontologies
Linked Data in Production: Moving Beyond OntologiesLinked Data in Production: Moving Beyond Ontologies
Linked Data in Production: Moving Beyond Ontologies
 
Building AI-Driven Apps Using Semantic Kernel.pptx
Building AI-Driven Apps Using Semantic Kernel.pptxBuilding AI-Driven Apps Using Semantic Kernel.pptx
Building AI-Driven Apps Using Semantic Kernel.pptx
 
activity_diagram_combine_v4_20190827.pdfactivity_diagram_combine_v4_20190827.pdf
activity_diagram_combine_v4_20190827.pdfactivity_diagram_combine_v4_20190827.pdfactivity_diagram_combine_v4_20190827.pdfactivity_diagram_combine_v4_20190827.pdf
activity_diagram_combine_v4_20190827.pdfactivity_diagram_combine_v4_20190827.pdf
 

Citrix XenDesktop on XenServer and IBM Hardware Sizing Baseline and Reference Configuration

  • 1. WHITE PAPER | Citrix XenServer Citrix XenDesktop on XenServer and IBM Hardware Sizing Baseline and Reference Configuration www.citrix.com
  • 2. Page 2 Contents Overview.............................................................................................................................................................3 Executive Summary...........................................................................................................................................3 Reference Architecture......................................................................................................................................4 Hardware Infrastructure Components ......................................................................................................................5 XenDesktop 5 Components .......................................................................................................................................7 XenServer Hosts.........................................................................................................................................................7 XenDesktop Desktop Delivery Controllers (DDCs).....................................................................................................8 Provisioning Services (PVS)........................................................................................................................................8 Citrix Web Interface...................................................................................................................................................9 Citrix Licensing...........................................................................................................................................................9 Virtual Desktops ......................................................................................................................................................10 Login Consultants VSI ..............................................................................................................................................11 Methodology.....................................................................................................................................................12 Pool Spin Up.............................................................................................................................................................12 User Workload.........................................................................................................................................................12 Findings.............................................................................................................................................................14 Success Criteria........................................................................................................................................................15 Test Results..............................................................................................................................................................16 References.........................................................................................................................................................26 Revision Description Managed By Date 1.0 Document created Roberto C. Moreno 3/14/2011 1.1 Document revised Ricardo Colon 3/17/2011 1.2 Document edited Roberto C. Moreno 3/17/2011 1.3 Document revised Zaid Haiman 3/18/2011 1.4 Document edited Roberto C. Moreno 3/18/2011 1.5 Document edited Roberto C. Moreno 3/19/2011 1.6 Document revised Zaid Haiman 3/21/2011 2.0 Final Draft Zaid Haiman 4/6/2011
  • 3. Page 3 Overview The overall goal of this White paper is to provide sizing baselines for the deployment of Citrix XenDesktop on IBM BladeCenter Servers running Citrix XenServer. A laboratory environment was created to test and validate the different components of the IBM hardware and the XenDesktop solution. The main focus of all tests conducted in the laboratory was to define the optimum number of virtual desktops that can be hosted on a single IBM BladeCenter HX5 server. Reference configuration of the environment, methodology for conducting tests, and the findings from tests performed are presented in this document. This document assumes that the reader has an architectural understanding of blade server hardware, storage area networks, and the Citrix XenDesktop Platinum components. Executive Summary The laboratory environment used for conducting all tests was based on the IBM BladeCenter HX5 Blade Server and the IBM System Storage N5200. A total of two BladeCenter HX5 servers were used. Each featured dual 1.86GHz Nehalem-EX processors with Intel Hyper-Threading enabled. Both servers were equipped with 50GB solid state drives (SSD), dual port Gigabit NICs, and Qlogic Fiber Channel adapters. Memory configurations included 64GB of DDR3 RAM for the blade hosting virtual infrastructure servers and 128GB for the blade hosting the virtual desktops. The IBM System Storage N5200 featured a cluster of two controllers and the IBM EXN4000 disk expansion unit with 14 FC 15K HDDs. The N5200 was attached to the HX5 Blade servers via 2Gbit Fiber Channel. The latest release of the Citrix XenDesktop 5 components available at the time of testing were used to build the laboratory virtual desktop infrastructure. The XenDesktop 5 Site consisted of:  Two (2) Desktop Delivery Controllers for brokering and managing the virtual desktops,  Two (2) Provisioning Servers for OS provisioning. Citrix Licensing and Web Interface servers were also deployed as part of the environment. All of the XenDesktop 5 infrastructure servers were virtualized on XenServer 5.6 FP1. Tests conducted within the laboratory environment validated the Citrix XenDesktop 5 hosted VDI solution running on the XenServer platform with IBM BladeCenter HX5 servers. Single server scalability tests show a single IBM BladeCenter HX5 server being capable of supporting an optimum number of 80 virtual desktops with a “medium user” workload using Microsoft Office 2010 on Microsoft Windows 7 Professional.
  • 4. Page 4 Memory utilization was determined to be the main bottleneck limiting the server from effectively hosting additional desktops with the XenServer virtual desktop host reporting 98% of RAM utilization with 80 active desktops. A combination of tools was used to simulate user load, monitor resource utilization, and gather data in order to validate the reference configuration. The latest version of the Login Virtual Session Indexer (VSI) 3 from Login Consultants was used for load testing and capturing system response time data. Other tools used for benchmarking and monitoring the environment included Citrix XenCenter, Citrix VM Performance Utility, Windows Performance Monitor, and NetApp’s ONTAP. Reference Architecture This section details the various components used in the laboratory to configure the environment. Configurations and specifications of the hardware infrastructure, XenDesktop 5, and LoginVSI components is presented in the following subsections. This configuration is only provided for reference purposes; specific configurations for production environments will vary based on each implementation’s needs. Figure 1 – Laboratory Components
  • 5. Page 5 Environment Summary: 2 XenServer Hosts Hosting of infrastructure VMs and virtual desktops 2 Desktop Deliver Controllers Brokering and managing virtual desktops 2 Provisioning Servers Virtual desktop OS provisioning 1 Citrix Web Interface / AD Domain Controller Provides access to XenDesktop resources and directory services 1 Citrix Licensing Server Licensing for all XenDesktop components 5 LoginVSI Launchers Initiates user sessions Table 1 Hardware Infrastructure Components All hardware used to build the laboratory environment was provided by IBM. The environment consisted of IBM BladeCenter HX5 servers and N Series System Storage. The blade chassis provided was the IBM BladeCenter S equipped with Ethernet and Fiber Channel switch modules. Installed on the chassis were two BladeCenter HX5 blade servers: one for hosting infrastructure servers and the other for the hosting of virtual desktops. Each server was configured with dual Intel Nehalem-EX processors, DDR3 VLP RAM, Solid State Drives, Qlogic Fiber Channel adapters, and onboard Broadcom network interface cards. Storage was provisioned by IBM System Storage N5200 controllers and the EXN4000 disk expansion unit connected to the servers via Fiber Channel. The aggregate used for the test environment consisted of six FC 15K 300 GB disks in a RAID 4 configuration. This provided the best performance and capacity combination based on the amount of disks available and the requirements of the environment. Additional details of the configuration of all hardware components used is presented tables 2 – 4 below. Figure 2 - IBM BladeCenter S
  • 6. Page 6 Component LAB Configuration IBM BladeCenter S Two BladeCenter HX5 Blades Four power supply modules Four cooling modules One management module One BNT Layer 2/3 Copper Gigabit Ethernet Switch Module One Brocade 8Gigabit Fiber Channel Switch Module Table 2 Figure 3 - IBM BladeCenter HX5 Component LAB Configuration IBM BladeCenter HX5 XenServer Virtual Desktop Host  Two Intel Xeon Processors L7555  Intel Hyper-Threading Technology enabled  128 GB RAM (16 x 8GB DDR3 SDRAM)  50GB SSD Drive (SATA 1.5GB)  Broadcom 5709S onboard NIC with dual Gigabit Ethernet ports  Qlogic 8Gb Fiber Channel Expansion Car IBM BladeCenter HX5 XenServer Management Server Host  Two Intel Xeon Processors L7555  Intel Hyper-Threading Technology enabled  64 GB RAM (16 x 4GB DDR3 SDRAM)  50GB SSD Drive (SATA 1.5GB)  Broadcom 5709S onboard NIC with dual Gigabit Ethernet ports  Qlogic 8Gb Fiber Channel Expansion Card Table 3
  • 7. Page 7 Figure 4 - IBM N5200 Figure – 5 IBM EXN4000 Component LAB Configuration IBM System Storage N Series N5200 (2864-A20) Modular Disk Storage System Dual Controller Active/Active Configuration DATA ONTAP 7.3.4 2 x Intel Xeon @ 2.8 GHz 4 GB RAM 1 GB Nonvolatile Memory 8 x 2Gb Fiber Channel Ports 8 x 1Gb Ethernet Ports Disk Expansion Unit EXN4000 4 x 2/4Gb Fiber Channel Ports 14 x 300 GB 15K RPM FC Disk Drives Table 4 XenDesktop 5 Components The XenDesktop 5 solution included the XenServer Hypervisor, XenDesktop Desktop Delivery Controllers (DDC), Provisioning Servers (PVS), Citrix Web Interface and License servers. All of the XenDesktop infrastructure servers were virtualized on XenServer. With a proper design and best practices applied none of these components require the use of physical servers. Both the XenDesktop DDCs and Provisioning Servers where deployed in pairs to provide redundancy and high availability. No performance gains from load balancing were expected as a single instance of either of these components could adequately handle the load from the single server scalability tests. XenServer Hosts Two XenServer hosts where deployed, each dedicated to a particular function; the management server host was used for all infrastructure servers, and the virtual desktop host was used for all virtual desktops. XenServer 5.6 FP1, the latest release at the time of testing, was used for both hosts. NOTE: It was necessary to enable “Remote Console” under Console Redirection Settings in the BladeCenter HX5 server BIOS in order to install XenServer 5.6 FP1. Not doing so resulted in the Installation process hanging.
  • 8. Page 8 For the installation of XenServer, the locally attached solid state drives (SSD) of the host were used. While XenServer 5.6 FP1 supports boot-from-SAN configurations, it was decided to use local storage to simplify configurations and dedicate the storage array to virtual machine disk operations. Storage repositories for virtual machines were created using provisioned LUNs over Hardware HBA. To handle network traffic a NIC bond was created using the two available NICs on each host. This bond was used for both XenServer management traffic and virtual machine network traffic. To better handle the number of virtual machines, the amount of RAM assigned to Dom0 on the virtual desktop host was increased to 2.94 GB from the default 752 MB, as recommended by Citrix. XenDesktop Desktop Delivery Controllers (DDCs) The laboratory XenDesktop Site consisted of two DDCs. The site’s database was configured using the locally installed copy of SQL 2008 Express on the first DDC deployed. In XenDesktop 5, collections of virtual machines (VMs) or physical computers are managed as a single entity called a catalog. A catalog is a collection of machines of the same type. The machine type defines the type of hosting infrastructure used for desktops and the level of control that users have over their desktop environment. For the tests conducted in this laboratory, the catalog was created using the streamed machine type. The streamed machine type enables you to deliver desktops to VMs and blade PCs that have been configured to load the operating system over the network leveraging Citrix Provisioning services. NOTE: For the purpose of this laboratory, Citrix Provisioning Services was used for provisioning all virtual desktops used in tests. Citrix Provisioning Services technology has been validated and proven effective in large scale deployments of XenDesktop. The XenDesktop 5 Machine creation services (MCS) was not used during testing. As with previous versions, XenDesktop 5 uses desktop groups to allocate virtual desktops to users and groups. A single desktop group was created from the streamed machine catalog for the laboratory tests. Provisioning Services (PVS) The Provisioning Services farm consisted of two Provisioning servers. The farm’s database was created on the same SQL 2008 Express instance used for the XenDesktop Site. The farm was configured with a single site and device collection for all target devices. Target devices were created using Provisioning Server’s Auto add feature. A single vDisk store was configured to create the master virtual desktop image (vDisk) that was used for all target devices. For the laboratory tests the store was configured using the local hard disk subsystem of the Provisioning Servers to store the vDisks. This provides the easiest way of implementing vDisk high availability, suitable for environments with few Provisioning servers such as this. The vDisks were manually synchronized between the Provisioning Servers.
  • 9. Page 9 Provisioning Services PXE and TFTP services were used to allow the target devices to boot from the network. The environment’s DHCP service scope options were configured to provide the target devices with the required information to connect with the Provisioning Servers and obtain their assigned vDisk. Citrix Web Interface The Citrix Web Interface 5.4 is included with XenDesktop 5 and it is installed by default on each of the DDCs during the XenDesktop setup wizard. Although both of the DDCs Web Interface sites were functional, they were not used in any of the tests performed. For testing purposes, version 5.2 of the Web Interface was used in order to match the LoginVSI reference configuration provided by Citrix and described in section 3.4. Citrix Licensing The functionality of all the components required that valid licenses specific to each product were available. For the purpose of this laboratory, a XenDesktop Platinum License file was used. This file included individual licensing for each of the components used. Additional details of the configuration of the XenDesktop components used are presented on table 5 below. Component LAB Configuration Management Server Host XenServer 5.6 FP1 2 x 8 Core Intel L7555 @ 1.86 GHz (32 Logical Cores Total) 64 GB RAM @ 1067 MHz 50 GB SSD Local Storage 550 GB Storage Repository (Hardware HBA) NIC Bond (2 x 1Gb NICs) Virtual Desktop Host XenServer 5.6 FP1 2 x 8 Core Intel L7555 @ 1.86 GHz (32 Logical Cores Total) 128 GB RAM @ 1067 MHz 50 GB SSD Local Storage 450 GB Storage Repository (Hardware HBA) NIC Bond (2 x 1Gb NICs) XenDesktop – Desktop Delivery Controllers (DDC) Windows Server 2008 R2 Standard Edition XenServer 5.6 VM 2 vCPU 4 GB RAM 24 GB VHD XenDesktop 5 Controller XenDesktop 5 Desktop Studio Citrix Web Interface 5.4 *DDC 1 – SQL 2008 Express (XD and PVS Data Stores) Provisioning Services / PXE / TFTP Windows Server 2008 R2 Standard Edition XenServer 5.6 VM 4 vCPU 8 GB RAM 24 GB VHD 30 GB VHD (vDisk Store) Provisioning Server 5.6 SP1
  • 10. Page 10 Component LAB Configuration Citrix Web Interface / Domain Controller / DNS / DHCP Windows Server 2008 R2 Standard Edition XenServer 5.6 VM 2 vCPU 4 GB RAM 40 GB VHD Citrix Web Interface 5.2 Citrix Licensing Server Windows Server 2008 R2 Standard Edition XenServer 5.6 VM 1 vCPU 1 GB RAM 24 GB VHD Citrix Licensing Server 11.6.1 XenDesktop Platinum License Table 5 Virtual Desktops All of the virtual desktops deployed for testing were based on the Windows 7 Professional operating system. Desktop virtual machines were configured for a normal user workload following Citrix guidelines in “XenDesktop Planning Guide: Hosted VM-Based Resource Allocation.” Each virtual desktop was configured 1 vCPU, 1.5GB of RAM, and a 4GB locally attached VHD for write cache. To prepare the master image, a new virtual machine was created on the XenServer virtual desktop host. Windows 7 was installed and updated with all available security patches and updates available from Microsoft. For the purposes of this test all applications required for testing were installed locally, not streamed or hosted. The following additional configurations were also implemented for increased performance:  All user profiles for the test user accounts were created locally on the master image according to the “Login VSI 3.0 Admin Guide.”  All Windows 7 optimizations for XenDesktop and Provisioning were configured following the “Windows 7 Optimization Guide” provided by Citrix.  All virtual desktops were configured with a locally attached VHD for PVS Write Cache  Following Login Consultants and Citrix recommendations, page file settings were modified to use a fixed size equal to the amount of memory allocated to the virtual machine, and the location of the page file was set to use the locally attached write cache VHD. The virtual desktop image was then converted to a Provisioning Services vDisk with the use of the PVS Imaging Wizard.
  • 11. Page 11 Table 6 below details the configuration used for all virtual desktops. Component LAB Configuration Virtual Desktop Windows 7 Professional Edition with SP1 XenServer 5.6 VM 1 vCPU 1.5 GB RAM 25 GB vDisk 4 GB VHD (Write Cache) MS Office 2010 Professional Plus Virtual Desktop Agent 5.0.0.148 Provisioning Services Target Device 5.6.1 Adobe Flash Player 10 Adobe Reader 9.1 Adobe Shockwave Player 10.2.0.023 Bullzip PDF Printer 6.0.0.865 FreeMind 0.8.1 Java 6 Update 21 GPL Ghostscript Lite 8.64 Kid-Key-Lock 1.2.1.0 Table 6 Login Consultants VSI Login Consultants VSI is a tool designed for benchmarking Server Based Computing (SBC) and Desktop Virtualization (VDI) solutions including Citrix XenDesktop. The free version, VSI 3.0 Express, was used for all tests performed. VSI consists of 4 components:  AD Domain controller for user accounts and standard policies  A file share for central configuration and logging  Launcher workstations to initiate the sessions  Target platform (VDI or SBC) where the user load scripts are installed and performed Domain user accounts and standard policies were automatically configured using the AD Deployment setup included with VSI 3.0. As part of the setup a new root OU is created, test users are created and GPOs are imported and linked. The file share and related permissions were configured manually following the “Login VSI 3.0 Admin Guide.” A total of five launchers were deployed with capacity configured for twenty sessions each. The launchers were prepared for integration with XenDesktop following the “Benchmarking Citrix XenDesktop 4 using Login Consultants VSI 2.1” white paper from Citrix. While some of the product versions described on the white paper are different from the versions used in this laboratory, the configurations are still valid.
  • 12. Page 12 Table 7 below details the configuration of the VSI Launchers. Component LAB Configuration VSI Launcher Windows Server 2003 R2 32-Bit Standard Edition XenServer 5.6 VM 2 vCPU 2 GB RAM 8 GB VHD Login VSI Launcher 3.0 Login VSI Analyzer 3.0 Citrix Online plug-in 12.1.0.30 Python 2.6.150 Python 2.6 pywin32-214 Table 7 Methodology All tests were performed using the laboratory environment and reference configuration described on section 3 of this document. All testing concentrated in determining the optimum amount of desktops that could be hosted on a single IBM BladeCenter HX5 server. Testing was done in two stages: pool spin up and user workload. During both stages, data was gathered from the XenServer hypervisor, XenDesktop infrastructure servers, virtual desktops, and storage to analyze for performance and resource utilization. Tests were performed several times to ensure results were consistent from one test to the next. Once it was determined that results were in fact consistent, a total of three official tests were executed. The highest and lowest scoring tests were then discarded and the remaining test result was selected to present the findings in this document. Pool Spin Up The pool spin up stage consisted of the process involved in getting all virtual desktops within the test desktop group available for use from a halted state. In preparation for the test, the desktop group is put on maintenance mode and all hosted virtual desktops are kept on a halted state within the XenServer host. To start the test, maintenance mode is disabled for the test desktop group and all virtual desktops are then started via XenDesktop 5 desktop group power management settings. As the virtual desktops boot using the provisioned Windows 7 image and register with the Desktop Delivery Controllers, they become available to users. The test is considered complete once all virtual desktops within the test desktop group show as available on the XenDesktop 5 desktop studio management console. User Workload The user workload was the most significant stage in validating the XenDesktop 5 solution with the reference configuration included in this document. The test consisted in generating simulated user load and measuring response time of the various tasks performed during the simulation. The process of simulating user load involved all phases of a user session including user logon, work time, and user logoff.
  • 13. Page 13 User workload was simulated using the Login Virtual Session Indexer (VSI 3.0) from Login Consultants. VSI is a benchmarking methodology based on the amount of simultaneous sessions that can be run on a single machine running either bare metal or virtualized operating systems. VSI is the standardized benchmark for Server Based Computing (SBC) and Virtual Desktop Infrastructure (VDI) environments and used by key vendors in the desktop virtualization space. The tool is 100% platform and protocol independent which allows for easy replication of results across most environments. The Express version of the VSI 3.0 tool was selected for all testing performed. VSI simulates realistic user workloads. The medium workload used for testing simulates an office worker using generic applications like Office, IE (including Flash applets) and Adobe Acrobat Reader. Like real users, the scripted session will leave multiple applications open at the same time. Every session will average about 20% minimal user activity, similar to real world usage. The medium workload is the only workload available in VSI Express and it is the default workload in VSI. The following list outlines the automated workflow performed by the medium workload:  This workload emulated a medium workload user using Office 2010, IE 8 and PDF.  Once a session has been started the medium workload will repeat every 12 minutes.  During each loop the response time is measured every 2 minutes.  The medium workload opens up to 5 applications simultaneously.  The type rate is 160ms for each character.  Approximately 2 minutes of idle time is included to simulate real-world users. Each loop will open and use:  Outlook 2010 - browse 10 messages.  Internet Explorer - one instance is left open (BBC.co.uk), one instance is browsed to Wired.com, Lonelyplanet.com and heavy flash application gettheglass.com.  Word 2010 - one instance to measure response time, one instance to review and edit document.  Bullzip PDF Printer & Acrobat Reader - the Word document is printed and reviewed to PDF.  Excel 2010 - a very large randomized sheet is opened.  PowerPoint 2010 - a presentation is reviewed and edited.  7-zip - using the command line version the output of the session is zipped. Sessions were launched at 30 second intervals until the total amount of sessions configured were started. The test was then allowed to run for about 1 hour to allow each session to go through a minimum of 2 loops before logoff. In addition to the user workload simulated by the Login VSI tool, a virtual desktop was given to a real person for use during testing. This was of great help in getting a feel for what system response time was like during tests.
  • 14. Page 14 Monitoring of Solution Components During all stages of testing data was gathered and analyzed to obtain performance and resource utilization results of the various components involved. The following set of tools was used to capture and analyze data:  Login Consultants VSI 3.0 Express – For capturing user experience data based on system response time.  Citrix XenCenter 5.6 FP1 – For real time performance monitoring of the XenServer host.  Windows Performance Monitor – For monitoring and capturing performance data of the Desktop Deliver Controllers and Provisioning Servers.  Netapp DATA ONTAP 7.3.4 – For monitoring and capturing performance data of the SAN.  Citrix Virtual Machine Performance Utility (StormVM) – For Disk I/O performance monitoring of the XenServer host storage repository used for virtual desktop write cache disks. Findings This section describes the findings of the tests performed to validate the XenDesktop 5 solution and defines the success criteria for a successful test run. Data from test results is presented in graphs detailing performance and resource utilization of the main components during each of the testing stages. Tests results show that with the architecture described in section 3 of this document it was possible to run up to eighty (80) Windows 7 virtual desktops with 1.5 GB of RAM on a single IBM BladeCenter HX5 server without scarifying performance and user experience. Pool spin up test results show that all 80 virtual desktops could be started and made available to users in 12 minutes. The Intel Nehalem-EX processors proved to be highly efficient under this type of load. Test results show average CPU utilization between 20% and 40% with 5 virtual desktops per core. Windows 7 virtual desktops 1 vCPU 1.5 GB RAM 25GB vDisk 4 GB PVS Write Cache XenServer 5.6 FP1 VMs/Host VMs/Core IBM BladeCenter HX5 Dual 8 Core (1.86GHz L7555) 128 GB RAM 80 5 Table 8 The factor limiting additional density on the server was available RAM. XenServer reported 98% of total memory utilization when the 80 virtual desktops where active. The BladeCenter HX5 Server available at the time of this writing had a maximum capacity of 128GB of RAM per single-wide blade. However, IBM is set to release a new version of the HX5 which will support up to 256 GB of RAM and Dual 10 Core processors. These additional resources should allow for greater virtual desktop density on a server with the same form factor.
  • 15. Page 15 The IBM System Storage N5200 effectively processed all I/O operations requested by the XenServer hosts. Citrix guidelines state that a typical virtual desktop on a working state would generate a load of about 14 IOPS1 and require 125kB/s of disk throughput. Based on these guidelines, the N5200 was expected to manage a load of approximately 640 (8 x 80) IOPS with 80% being random write operations and 10MB/s (125kB/s x 80) of disk throughput during the user workload stage. Test results show the N5200 handling Fiber Channel Protocol (FCP) OPS in the 500-600 range, and disk throughput near 10MB/s, very close to expectations. Results from additional stress tests conducted, show the N5200 supporting a high of 878 FCP OPS and 75MB/s FCP write throughput. As expected, none of the XenDesktop components showed significant resource utilization during any of the tests conducted. Critical components such as the Desktop Delivery Controllers and Provisioning Servers have been proven to work under greater loads. Success Criteria Success criteria to determine if a test cycle passed or failed was based on the VSI Max score. The Login VSI Max evaluates the user response time during increasing user load. VSI focuses on how many users can run on the system before it saturates. The operations executed by VSI to simulate user load are designed to measure system response time by completing activities that leverage resources such as CPU, memory, disk, the OS in general, printing and GDI, among others. These operations are specifically short by nature. When such operations are consistently long the system is saturated because of excessive queuing on any kind of resource. As a result, the average response times escalate. This effect is clearly visible to end-users. When such operations consistently consume multiple seconds the user will regard the system as slow and unresponsive. Response time for all tasks performed during the simulated user workload is measured in milliseconds. The Login VSI 3.0 Express medium workload used for testing has a pre-defined threshold of 4000ms to determine the VSI Max Pass or Fail rating. Tests that resulted in user response times above 4000ms were considered failed. While the VSI Max score was the single factor considered for the pass or fail of any given test, real user experience was also helpful in validating VSI results. During all tests conducted, a real user was given access to a virtual desktop in the same manner that the VSI Launchers established connections. User feedback regarding system response time was then evaluated and used to confirm the pass or fail rating given by the VSI Max score. While user feedback is a very subjective measure it is generally used in production environments to determine if the system is performing according to expectations. 1 During certain events, such as logon storms, IOPS will be higher than these estimated values.
  • 16. Page 16 Test Results The following graphs demonstrate results from all data gathered during tests. Results from both stages of testing, pool spin up and user workload, are presented for the main components involved. Login VSI Graph 1 below demonstrates a successful test run where user experience was not affected by the 80 concurrent sessions running on the same host. All response times are below the 4000ms threshold defined by VSI 3.0 Express. Graph 1 - Login VSImax / 80 Sessions Vertical axis: Response Time in milliseconds Horizontal axis: Total Active Sessions  Red line: Maximum Response (worst response time of an individual measurement within a single session)  Orange line: Average Response Time within for each level of active sessions  Blue line: The VSImax average (Typically VSImax is reached when CPU Utilization is around 100%)  Green line: Minimum Response (best response time of an individual measurement within a single session) XenServer Figure 6 below shows 98% of RAM utilization by the virtual desktop host with 80 active VMs. This was the limiting factor in the amount of virtual desktops that could be hosted on a single server with the specified configuration. Graphs 2 and 3 below demonstrate resource utilization by the virtual desktop host during the Pool Spin Up and User Workload stage. Graph 2 shows cores 0-3 of the host running at about 80%, these cores are used by XenServer and were expected to be highly utilized during the startup of all VMs.
  • 17. Page 17 Figure 6 - XenServer Virtual Desktop Host / Memory Utilization Graph 2 - XenServer Virtual Desktop Host / Pool Spin Up
  • 18. Page 18 Graph 3 - XenServer Virtual Desktop Host / User Workload
  • 19. Page 19 IBM System Storage N5200 The N5200 proved to be more than adequate for a workload of 80 virtual desktops. Test data confirms that write operations account for approximately 80% of the total IOPS generated by VDIs. Graphs 4 and 5 below present results from the XenServer Virtual Machine Performance Utility (StormVM) disk I/O benchmark. The StormVM Disk I/O utility was used to conduct a stress test consisting of random write operations on the storage repository used for virtual desktop write cache. The stress test was conducted with virtual desktops being idle and during the user workload stage. The two spikes seen in graphs 7, 9, 11, and 13 represent the impact of the stress test performed during the user workload stage. Graph 4 - StormVM 1MB Random Writes / 80 Desktops IDLE Graph 5 - StormVM 1MB Random Writes / 80 Desktops Active
  • 20. Page 20 Graph 6 - Pool Spin UP / N5200 FCP OPS Graph 7 – User Workload / N5200 FCP OPS Graph 8 - Pool Spin UP / N5200 CPU-Disk Util Graph 9 – User Workload / N5200 CPU-Disk Util 0 50 100 150 200 250 300 350 400 450 500 13:04:00 13:05:45 13:07:30 13:09:15 13:11:00 13:12:45 13:14:30 FCP OPS 0 100 200 300 400 500 600 700 800 900 1000 15:02:00 15:13:30 15:25:00 15:36:30 15:48:00 15:59:30 16:11:00 FCP OPS 0% 2% 4% 6% 8% 10% 12% 13:04:00 13:05:45 13:07:30 13:09:15 13:11:00 13:12:45 13:14:30 CPU Util Disk Util 0% 10% 20% 30% 40% 50% 60% 70% 80% 15:02:00 15:12:00 15:22:00 15:32:00 15:42:00 15:52:00 16:02:00 16:12:00 CPU Util Disk Util
  • 21. Page 21 Graph 10 - Pool Spin UP / N5200 FCP Throughput Graph 11 – User Workload / N5200 FCP Throughput Graph 12 - Pool Spin UP / N5200 Disk R/W Graph 13 – User Workload / N5200 Disk R/W 0 500 1000 1500 2000 2500 3000 3500 4000 4500 5000 13:04:00 13:06:15 13:08:30 13:10:45 13:13:00 13:15:15 FCP kB/s in FCP kB/s out 0 10000 20000 30000 40000 50000 60000 70000 80000 15:02:00 15:15:30 15:29:00 15:42:30 15:56:00 16:09:30 FCP kB/s in FCP kB/s out 0 1000 2000 3000 4000 5000 6000 7000 8000 13:04:00 13:05:45 13:07:30 13:09:15 13:11:00 13:12:45 13:14:30 Disk kB/s read Disk kB/s write 0 10000 20000 30000 40000 50000 60000 70000 80000 90000 100000 15:02:00 15:15:30 15:29:00 15:42:30 15:56:00 16:09:30 Disk kB/s read Disk kB/s write
  • 22. Page 22 Desktop Delivery Controller As both DDCs test results were nearly identical, only the data from DDC1 is presented in this document. Data shows DDC resource utilization was very similar during both tests. Graph 14 - Pool Spin UP / DDC1 CPU Graph 15 – User Workload / DDC1 CPU Graph 16 - Pool Spin UP / DDC1 RAM Graph 17 – User Workload / DDC 1 RAM 0 10 20 30 40 50 60 70 80 90 100 % Processor Time:_Tot al (2vCPU) 0 10 20 30 40 50 60 70 80 90 100 % Processor Time:_Tot al (2vCPU) 0 500 1000 1500 2000 2500 3000 3500 4000 MemoryC ommitted Megabytes 0 500 1000 1500 2000 2500 3000 3500 4000 MemoryC ommitted Megabytes
  • 23. Page 23 Graph 18 - Pool Spin UP / DDC1 Network Graph 19 - User Workload / DDC1 Network 0 20 40 60 80 100 120 140 Network Interface: kB/s Received Network Interface: kB/s Sent 0 10 20 30 40 50 60 70 Network Interface: kB/s Received Network Interface: kB/s Sent
  • 24. Page 24 Provisioning Server As both PVS test results were nearly identical, only the data from PVS1 is presented in this document. As expected, the Provisioning server CPU and network utilization (Graph 20, 24) was the highest during the Pool Spin UP test. Grpahs 22 and 23 show that RAM utilization remained the same throughout both tests indicating the vDisk being cached in system RAM. This is further confirmed on graphs 26 and 27 where disk utilization remained idle nearly 100% of the time during both tests. Graph 20 - Pool Spin UP / PVS1 CPU Graph 21 – User Workload / PVS1 CPU Graph 22 - Pool Spin UP / PVS1 RAM Graph 23 – User Workload / PVS1 RAM 0 10 20 30 40 50 60 70 80 90 100 13:04:05 13:06:05 13:08:05 13:10:05 13:12:05 13:14:05 13:16:05 % Processor Time:_Tot al (4vCPU) 0 10 20 30 40 50 60 70 80 90 100 % Processor Time:_Tot al (4vCPU) 0 1000 2000 3000 4000 5000 6000 7000 8000 13:04:05 13:06:20 13:08:35 13:10:50 13:13:05 13:15:20 MemoryC ommitted Megabytes 0 1000 2000 3000 4000 5000 6000 7000 8000 MemoryC ommitted Megabytes
  • 25. Page 25 Graph 24 - Pool Spin UP / PVS1 Network Graph 25 - User Workload / PVS1 Network Graph 26 - Pool Spin UP / PVS1 Disk Graph 27 – User Workload / PVS 1 Disk 0 5000 10000 15000 20000 25000 30000 35000 13:04:05 13:06:20 13:08:35 13:10:50 13:13:05 13:15:20 Network Interface: kB/s Received Network Interface: kB/s Sent 0 2000 4000 6000 8000 10000 12000 14000 16000 18000 20000 Network Interface: kB/s Received Network Interface: kB/s Sent 0 10 20 30 40 50 60 70 80 90 100 13:04:05 13:06:05 13:08:05 13:10:05 13:12:05 13:14:05 13:16:05 PhysicalDis k: % Idle Time 0 10 20 30 40 50 60 70 80 90 100 PhysicalDis k: % Idle Time
  • 26. Page 26 References Reference Title http://support.citrix.com/article/CTX125630 Citrix XenDesktop with Login Consultants VSI http://support.citrix.com/article/CTX127050 XenDesktop - Windows 7 Optimization Guide http://support.citrix.com/article/CTX127277 XenDesktop Planning Guide: Hosted VM-Based Resource Allocation http://support.citrix.com/article/CTX128645 Design Considerations for Virtualizing Provisioning Services Included with Login VSI 3.0 Express LoginVSI 3.0 Admin Guide http://support.citrix.com/article/ctx127065 XenServer Virtual Machine Performance Utility http://support.citrix.com/proddocs/index.jsp Citrix eDocs http://support.citrix.com/article/CTX126531 Configuring Dom0 Memory in XenServer 5.6 .
  • 27. Page 27 About Citrix Citrix Systems, Inc. (NASDAQ:CTXS) is the leading provider of virtualization, networking and software as a service technologies for more than 230,000 organizations worldwide. Its Citrix Delivery Center, Citrix Cloud Center (C3) and Citrix Online Services product families radically simplify computing for millions of users, delivering applications as an on-demand service to any user, in any location on any device. Citrix customers include the world’s largest Internet companies, 99 percent of Fortune Global 500 enterprises, and hundreds of thousands of small businesses and prosumers worldwide. Citrix partners with over 10,000 companies worldwide in more than 100 countries. Founded in 1989, annual revenue in 2008 was $1.6 billion. ©2009 Citrix Systems, Inc. All rights reserved. Citrix®, Access Gateway™, Branch Repeater™, Citrix Repeater™, HDX™, XenServer™, XenApp™, XenDesktop™ and Citrix Delivery Center™ are trademarks of Citrix Systems, Inc. and/or one or more of its subsidiaries, and may be registered in the United States Patent and Trademark Office and in other countries. All other trademarks and registered trademarks are property of their respective owners.