SESSION CODE: VIR-EXL308
Glen Gooda
Senior Consultant
Microsoft
VIRTUALIZING MICROSOFT
EXCHANGE
SERVER WITH HYPER-V
What We Will Cover
► Why Microsoft Virtualization for Microsoft Server
Applications?
► Base Hypervisor Performance
► ESG Performance Results
► Exchange Server Virtualization Best Practices
► Links to Reference Material
Trends – Changing Market Landscape
►
Virtualization is exploding resulting in VM proliferation and impacting OS share
Unix
6%
Other
1%
Number of physical servers shipments used for virtualization will grow to 1.7M+ in 2012 at a CAGR of 15%
IDC Server Virtualization Forecast
14,000,000
12,000,000
10,000,000
Linux 21%
Unpaid
Windows
11%
Licensed
Windows
61%
8,000,000
6,000,000
4,000,000
2,000,000
0
Y2005 Y2006 Y2007 Y2008 Y2009 Y2010 Y2011 Y2012
Physical Units
Logical Units
19% of physical server shipments will be used for virtualization, increasing from 11.7% in 2007
Why Microsoft Virtualization
for Microsoft Server
Applications
Microsoft Virtualization for Server Applications
Enterprise
Applications
Line of Business (LOB)
Custom Applications
Management
Platform
Microsoft Virtualization: The Best Choice for
Microsoft Server Applications
*Built-in Virtualization
with one-stop support
*Large Partner
Ecosystem
Increased Deployment
Options
*Deep Application
Knowledge
*Physical & Virtual
Management
*Cross-Platform &
Hypervisor Support
*A comparable solution can
cost almost 4 times more†
*Lower On-going
Costs
Virtualization-Friendly
Licensing
*Only available with Microsoft Virtualization
†Based on a comparison of Microsoft® System Center Server Management Suite Datacenter with VMware® vSphere Enterprise Plus with VMware vCenter Server.. Assumes a five host configuration, 2 processors on each host, 2
years support costs for both products, and no operating system costs included.. The Microsoft solution can use either the free Microsoft Hyper-V Server 2008 R2 hypervisor or an existing Windows Server 2008 R2
hypervisor. Based on Microsoft estimated retail prices and published VMware prices available at https://www.vmware.com/vmwarestore as of 08/04/2009 for purchases in the United States. Actual reseller prices may vary.
Windows Server 2008 R2 SP1
► Many of the scenarios discussed leverage features new to R2 and R2 SP1
► Live Migration & HA
► New Processor Support
– Improved Performance & Lower Costs
► Enhanced Scalability
– Greater VM density & Lower TCO
► Networking enhancements
– Improve performance & 10 Gb/E ready
► Dynamic VM capabilities
► Enhancements to Server Core
– Ease management & Lower TCO
► Dynamic Memory & Remote FX
in SP1
Hyper-V Configuration Guidelines
► Hyper-V Root Configuration
Plan for 1GB+ memory reserve for the management OS in the root partition
Plan for one dedicated NIC for management purposes
Plan (ideally) for one dedicated NIC for live migration
Separate LUNs/Arrays for management OS, guest OS VHDs and VM storage
Management OS and VHD LUNs should employ RAID to provide data
protection and performance
– Challenge for blades with 2 physical disks
–
–
–
–
–
Hyper-V Configuration Guidelines
► Hyper-V Guest Configuration
– Fixed-sized VHDs for Virtual OS
• Need to account for page file consumption in addition to OS requirements
OS VHD Size (minimum 15GB) + VM Memory Size = Minimum VHD size
– Account for space needed by additional files by VM
• Example for SQL: OS VHD Size + (VM Memory Size) + Data Files + Log Files
(c) 2011 Microsoft. All rights reserved.
Dynamic Memory in SP1
► Overview
– A memory management enhancement for Hyper-V
– Enables customers to dynamically grow and decrease the memory of a VM
– Available as a feature in Windows Server 2008 R2 SP1
► Benefits
–
–
–
–
Better consolidation ratios with predictable performance
Enables linear scalability for deployment environments
Designed for production use
Recommendations for use for server applications still being evaluated.
► With the SP1 release, we will meet all the Gartner/Burton group
requirements for an Enterprise ready Hypervisor
How does it work?
► VM memory configuration parameters:
– Initial (what VM will boot with)
– Maximum (what VM can grow to)
► Memory is pooled and dynamically distributed across VMs
► Memory is dynamically allocated/removed based VM usage
with no service interruption
► Guest enlightened: guests & Hyper-V work TOGETHER
► Memory is added and removed via synthetic memory driver
(memory VSC) support
How is it different?
► There is a difference between memory overcommit and overcommitting
memory
– We warned of the dangers of overcommitting memory
– We pressed against the use of overcommit from a marketing perspective
► Dynamic Memory allows for predictable, consistent performance even
after overcommitting memory
► Key is avoid the performance cliff which is possible with the VMware
solution
► The features VMware has does NOT provide significantly more scale than
Hyper-V with DM does (despite what VMware will tell our customers)
How does VMware do it?
► Memory Ballooning and Un-ballooning
– Set VM memory, reservation, and limit
– Will add and remove as necessary, between the reservation and limit
► Memory Compression
► Page Sharing
– Optimized common memory pages
– Affected by ASLR and SLAT
► Hypervisor Memory Paging
– Pages at the Hypervisor when physical memory is completely committed
► All these (italicized) are only used after all physical memory is committed,
which is not recommended by VMware or Microsoft!
Base Hypervisor
Performance
Project Virtual Reality Check
► Available at www.virtualrealitycheck.net
► Performed by Ruben Spruijt and Jeroen van de Kamp
► Not sponsored by any one company, although VMware and Citrix
have assisted the site
► Results are focused on running and replacing Terminal Server
workloads only, on vSphere, Hyper-V, and XenServer
► Results are not for redistribution or validation, although
they are public
► Phase II results from February 2010, with significant
increase in vSphere performance
(c) 2011 Microsoft. All rights reserved.
Project VRC Results
► Enable EPT/RVI results in a significant increase in capacity of VMs running TS
– vSphere – 90% increase
– XenServer – 95% increase
– Hyper-V – 154% increase
► When scaling x86 TS VMs w/o Hyper-threading, vSphere is 5% better than
both Xen and Hyper-V
Project VRC Results
► When scaling x86 TS VMs w/ Hyper-threading, Xen and Hyper-V are 15%
better than vSphere
– When scaling up to 100 TS sessions, response times for all three hypervisors are fairly
equal
– Beyond 100 sessions, vSphere response times increases with each new session
► When scaling x64 TS VMs, Xen and Hyper-V are within 13.6% of bare metal,
and are 27% better than vSphere
(c) 2011 Microsoft. All rights reserved.
2010 ESG Paper
► 3rd Party Performance Validation White Paper, sponsored by Microsoft
► Key Points
– Hyper-V is easy to install to get running for those administrators familiar with
Windows
– Hyper-V provides high availability with transparent and automatic failover of
Virtual Machines (VMs)
– Hyper-V Live Migrations provides for zero-downtime migrations of VMs
between Hyper-V servers
– Most importantly, the report shows Hyper-V performance versus physical is
excellent, with 95% to 99% of the performance of physical disks and 89% to
98% of performance of the tested workloads versus physical.
2010 ESG Paper
► All this strongly supports the position that Hyper-V performance is not a
bottleneck or blocker to implementation and that customers can run
almost all their workloads on Hyper-V
► Publically available and you can distribute to your customers/clients
► http://www.enterprisestrategygroup.com/2010/07/microsoft-hyper-v-r2scalable-native-server-virtualization-for-the-enterprise/
► http://www.infostor.com/index/articles/display/5976242552/articles/infos
tor/esg-lab-review/2010/july-2010/microsoft-hyper-v.html
Hyper-V Fixed Virtual Hard Disk
Performance Overhead Analysis
Native vs. Fixed VHD IO Performance Analysis
Disk IOs per second
(IOPS)
(Window 2008 R2 Physical Disk(Native) vs. Fixed VHD(VM), ESG Lab Iometer workloads)
SQL Log
OLTP Database
Physical Disk (Native)
File Server
Fixed VHD (VM)
Exchange DB
Hyper-V R2 Application Overhead Analysis
Virtual vs. Physical Performance Analysis
(Relative difference in elapsed time per operation; Windows 2008 R2 data center)
Application install
Directory level copy
Subsequent copies
SQL query
0%
10%
20%
30%
40%
Native
50%
60%
70%
Hyper-V R2 VM
80%
90%
100%
Microsoft/Intel iSCSI Test
► Used Windows Server 2008 R2,
Intel Xeon 5500 processors, and
Intel 10Gbps Ethernet Adapters
► Reached over One Million IOPS
over a single 10 Gbps Ethernet link
using a software iSCSI initiator on
Native HW
► Reached over 700,000 IOPS over a
single 10 Gbps Ethernet link using
a software iSCSI initiator on HyperV to the Guest OS
Microsoft/Intel iSCSI Test
► Native Performance
► In-Guest VM Performance
Microsoft/Intel iSCSI test
► iSCSI Performance can be as good or better than either Fibre Channel
or FCoE. iSCSI can offer a compelling offering for those who don’t
want to invest in FC.
► Windows Server and Hyper-V is not a limiting factor to IO
performance. There shouldn’t be any significant concern around IO
for virtualizing with Hyper-V.
► More details (along with a great Photoshop picture) at
http://gestaltit.com/all/tech/storage/stephen/microsoft-and-intelpush-one-million-iscsi-iops/
ESG Performance
Results
ESG Lab Summary 2011
► Available online:
http://www.microsoft.com/virtualization/e
n/us/solution-business-apps.aspx
► Useful links, resources, case studies, white
papers and webcasts
2011 ESG Test Lab - Physical
HP BL680C
up to 24 cores and
128 GB RAM per
blade
SAN
2x4 Gbps FC per server
RAID-10 Pools:
EMC CX4-960
155 15K RPM FC disk
drives
Data (88):
Logs(16):
OS(24):
Apps(16):
http://download.microsoft.com/download/F/0/9/F09FEDCA-1FEA-4DB4-B131-563145BF0B6F/ESG%20Preso%20Microsoft%20Hyper-V%20Performance%20Exchange%20Mar%2011_Wide.pdf
2011 ESG Test Lab - Virtual
Application: Exchange 2010
VM configuration: 4 vCPU, 4 GB
RAM
Microsoft Windows Server 2008
R2 SP1
Hypervisor: Microsoft Hyper-V R2
Physical OS: Windows Server 2008
R2 SP1
Hyper-V R2
SAN
Virtual machine images: Fixed VHD
Exchange data and logs: Fixed VHD
http://download.microsoft.com/download/F/0/9/F09FEDCA-1FEA-4DB4-B131-563145BF0B6F/ESG%20Preso%20Microsoft%20Hyper-V%20Performance%20Exchange%20Mar%2011_Wide.pdf
Microsoft Exchange Server Jetstress 2010
► Scale from 5,000 to 20,000 simulated Exchange mailboxes
► 5,000 mailboxes per VM
► 250 MB mailboxes
► 0.15 IOPs per mailbox
► 2 databases/2 copies per VM
http://download.microsoft.com/download/1/7/F/17FB551C-0905-4A04-AB46-2EBA616CFDF3/ESG%20Preso%20Microsoft%20Hyper-V%20Performance%20SharePoint%20Mar%2011_Wide.pdf
Hyper-V R2 Virtual Application Workload Scalability
25,000 (Up to 20,000 Exchange 2010 mailboxes, Windows 2008
25
R2 SP1)
20,000
20
15,000
15
10,000
10
5,000
5
0
0
1
2
3
Hyper-V R2 Virtual Machines
Exchange DB Read
Response Time (ms)
Exchange 2010 Mailboxes
2011 Exchange Workload Results
4
http://download.microsoft.com/download/F/0/9/F09FEDCA-1FEA-4DB4-B131-563145BF0B6F/ESG%20Preso%20Microsoft%20Hyper-V%20Performance%20Exchange%20Mar%2011_Wide.pdf
Exchange IO Operations per second
2011 Exchange Workload Results
Hyper-V R2 Virtual Application Workload Scalability
8,000 (Up to 20,000 Exchange 2010 mailboxes, Windows 2008
R2 SP1)
7,000
6,000
5,000
4,000
3,000
2,000
1,000
0
1
2
3
4
Hyper-V R2 Virtual Machines
http://download.microsoft.com/download/F/0/9/F09FEDCA-1FEA-4DB4-B131-563145BF0B6F/ESG%20Preso%20Microsoft%20Hyper-V%20Performance%20Exchange%20Mar%2011_Wide.pdf
Exchange Tested Solutions
►
►
►
►
►
Microsoft and participating server and storage vendors
Key design decision points for common scenarios
Cost effective resource planning (server, storage, network)
Device Access Group and multi-site designs for HA
Documented results of exhaustive performance tests
►
►
500 Mailboxes in a Single Site Running Hyper-V on Dell Servers
9000 Mailboxes in Two Sites Running Hyper-V on Dell M610 Servers, Dell EqualLogic Storage, and F5 Load Balancing
Solutions
15000 Mailboxes in Two Sites Running Hyper-V on Unisys ES7000 Servers and Hitachi Adaptable Modular Storage 2000
Family
16000 Mailboxes in a Single Site Deployed on IBM and Brocade Hardware
20000 Mailboxes in Two Sites Running Hyper-V on Dell R910 Servers, EMC CLARiiON Storage, and Brocade Network Solutions
32400 Mailboxes In Three Sites Running Hyper-V on Cisco Unified Compute System Blade Servers and EMC CLARiiON
Storage
►
►
►
►
http://download.microsoft.com/download/F/0/9/F09FEDCA-1FEA-4DB4-B131-563145BF0B6F/ESG%20Preso%20Microsoft%20Hyper-V%20Performance%20Exchange%20Mar%2011_Wide.pdf
Exchange Tested Solutions with Hyper-V
35,000
Exchange 2010 Mailboxes
30,000
25,000
20,000
15,000
10,000
5,000
0
1
2
3
Exchange Tested Solutions
4
http://download.microsoft.com/download/F/0/9/F09FEDCA-1FEA-4DB4-B131-563145BF0B6F/ESG%20Preso%20Microsoft%20Hyper-V%20Performance%20Exchange%20Mar%2011_Wide.pdf
Exchange Tested Solutions on Hyper-V
Hyper-V R2 VMs
1.
2.
3.
4.
Solution
Mboxes
Sites
Cores
RAM(GB)
Drives
Mbox
CAS/Hub
Total
1
9,000
2
36
432
144
9
9
18
2
15,000
2
96
512
136
12
8
20
3
20,000
2
128
768
160
16
16
32
4
32,000
3
96
1,152
432
24
12
36
http://technet.microsoft.com/en-us/library/gg513522.aspx
http://www.microsoft.com/downloads/en/details.aspx?FamilyID=4D83E5AC-7D9A-47C6-A843-A5520D118FC4
http://www.microsoft.com/downloads/en/details.aspx?FamilyID=6FD7E23F-30C8-4FB7-BCA3-A5DE9AA9A8E4
http://www.emc.com/collateral/hardware/white-papers/h7337-exchange-unified-cisco-hyper-v-wp.pdf
http://download.microsoft.com/download/F/0/9/F09FEDCA-1FEA-4DB4-B131-563145BF0B6F/ESG%20Preso%20Microsoft%20Hyper-V%20Performance%20Exchange%20Mar%2011_Wide.pdf
2011 Exchange Results Summary
► 20,000 simulated Exchange 2010 users on a single server in
four Hyper-V R2 VMs with Jetstress
► Exchange tested solutions from 9,000 through 32,000
mailboxes
► The 32,000 mailbox solution was deployed in 36 Hyper-V R2
VMs over three sites
► The performance, scalability and low overhead of Hyper-V R2
can be used to reduce costs and improve the
manageability, flexibility, and availability of consolidated
Exchange applications.
http://download.microsoft.com/download/F/0/9/F09FEDCA-1FEA-4DB4-B131-563145BF0B6F/ESG%20Preso%20Microsoft%20Hyper-V%20Performance%20Exchange%20Mar%2011_Wide.pdf
Exchange Server
Virtualization
Best Practices
Why Virtualize Exchange
Host in Datacenter
VM
1
Exchange 2010
MBX
DC 1
File &
Print Server
Exchange 2010
MBX
DC 2
Database
Server
2
3
NLB
DAG
Exchange 2010
CAS & HUB
Exchange 2010
CAS & HUB
Exchange 2010
UM
Management Server
Consolidate under-utilised servers into a single virtualized hosts
Lower costs by reducing space needs and power consumption
Rapid provisioning of a mobile infrastructure
Updated Support Guidance
–
–
–
–
Support for virtualized Exchange servers since Exchange Server 2007 SP1
Exchange 2010 release continued support for virtualization
Expanding support scenarios
Release of Exchange 2010 Virtualization Guidance whitepaper
Ex 2007
Ex 2010
RTM
Ex 2010
SP1 (Now)
Any hypervisor validated under Windows SVVP



All storage used by an Exchange guest must be block level storage



Virtual storage must be fixed size, SCSI pass-through, or iSCSI



Taking virtual snapshots of Exchange guest, not supported



Virtual processor-to-logical processor ration no greater than 2:1



Exchange HA in combination with hypervisor clustering or migration

Unified Messaging role supported

Support Guidance (continued)
► Storage
–
–
–
–
Fixed sized VHD, SCSI or iSCSI pass-through.
Dynamic or Thin Disks are not supported
Differencing or Delta Disks are not supported
Disk space should be 15GB + VM Memory Size
Support Guidance (continued)
► Other considerations
– Only the management software in the parent partition. No other roles
should be running.
– VM Snapshots are not supported
– 2:1 vCPU to physical CPU ratio is supported but 1:1 is best
– Current guidance to not use any memory management or
oversubscription
(c) 2011 Microsoft. All rights reserved.
Deployment Recommendations
► Exchange application is not ‘virtualization aware’
► Core Exchange Design Principles Still Apply
– Design for Performance, Reliability and Capacity
– Design for Usage Profiles (CAS/MBX)
– Design for Message Profiles (Hub/Edge)
Deployment Recommendations
► Virtualization Design Principles Now Apply
– Design for Performance, Reliability and Capacity
– Virtual machines should be sized specific to the Exchange role
(EDGE, HUB, CAS, MBX, multi-role)
– Hosts should be sized to accommodate the guests that they will
support
(c) 2011 Microsoft. All rights reserved.
Best Practices:
Capacity, Sizing and Performance
Sizing Process Overview
► Start with the physical server sizing process
– Calculator & TechNet guidance
► Account for virtualization overhead
► Determine VM placement
– Account for VM migration if planned
► Size root servers, storage, and network infrastructure
Guest Sizing Rules of Thumb
► Size Mailbox role first
– CPU ratios for other roles based on Mailbox role sizing
– Mailbox role performance is key to user experience
– High availability design significantly impacts sizing
► Don’t oversubscribe resources
– Size based on anticipated peak workload, don’t under provision
physical resources
► Don’t forget network needs
Guest Sizing for Unified Messaging
► Newly supported for virtualization
– Requires Exchange 2010 SP1 (or greater)
► Role is susceptible to poor voice quality and/or latency if
undersized
► Requires min. 4 virtual processors
► UM must be able to utilize physical processors on demand
► Consider network requirements (low latency, sufficient
bandwidth) to meet UM needs
► Tests show that 4VP/16GB VM can handle 40 concurrent calls
with VM Preview and 65 calls without
Root Server Sizing
► Root server storage sizing includes space for the OS & required
hypervisor components, plus connectivity to storage for guest VMs
– Don’t forget about high availability of storage if required (multipath HBAs or iSCSI NICs, redundant paths, etc.)
► Network sizing is critical: number of interfaces and bandwidth
– Consider app connectivity, storage networking, heartbeats, CSV,
VM migration
Root Server Sizing
► CPU sizing should include root needs plus per-guest overhead
– Follow hypervisor vendor recommendations
► Memory sizing should not assume oversubscription
– Follow hypervisor vendor recommendations
– Provide memory for root plus sum of running VM requirements
– Memory for Hyper-V root = the larger of 512MB or the per-VM value (summed
for running VMs) of 32MB for the first 1GB of virtual RAM + 8MB for each
additional GB of virtual RAM
• Example: 8 VMs running, each with 32GB RAM. Root requires 8 * (32MB +
8MB*31) = 2240MB
Virtual Processors
► Scale up CPU on VMs as much as possible
– Don’t deploy 4 x 1 vCPU machines vs. 1 x 4 vCPU machine: take
advantage of Exchange scalability
► Don’t oversubscribe CPUs unless consolidating with P2V, or similar
scenario
► Generally assume 1 logical CPU == 1 virtual CPU, don’t assume that
a hyperthreaded (SMT) CPU counts
Best Practices:
Server Deployment
Locating Virtual Machines
► VM placement is important for high availability
► Don’t co-locate DAG database copies on physical hosts
► Exchange unaware of VM location relative to other VMs
– No path correction in transport to avoid data loss
► Ensure peak workload can run in standard VM locations
– OK to move temporarily for maintenance assuming high
availability requirements are met and current workload can be
serviced
Storage Decisions
► Exchange performance and health highly dependent on
availability and performance of storage
► Many options for presentation of storage to VMs
–
–
–
–
VHD
FC
iSCSI, FCoE
DAS
► Optimize for performance and general design goals
– We recommend looking for options that provide large mailboxes
and low cost
Storage Decisions
► Exchange storage should be on spindles separate from guest OS
VHD physical storage
► Exchange storage must be fixed VHD, SCSI pass-through or iSCSI
– Preference is to use SCSI pass-through to host queues, DBs, and
logfile streams
– Hyper-V Live Migration suggests Cluster Shared Volumes with
fixed VHD (faster “black-out” period)
► FC/SCSI HBAs must be configured in Root OS with LUNs presented
to VMs as pass-through or VHD
Storage Decisions
► Internet SCSI (iSCSI)
– Standard best practices for iSCSI connected storage apply
(dedicated NIC, jumbo frames, offload, etc…)
– iSCSI initiator in the guest is supported but need to account for
reduced performance
► Exchange storage must be block-level
– Network attached storage (NAS) volumes not supported
(c) 2011 Microsoft. All rights reserved.
Exchange VM Deployment
► Exchange setup must be run when VM is provisioned
– Not “sysprep friendly”
► Possible to script Exchange setup to fully automate Exchange VM
provisioning
► Build “starter image” with desired OS, patches, pre-reqs, and
Exchange install binaries
Best Practices:
High Availability & VM Migration
High Availability And Disaster Recovery
► Exchange High Availability Definition
– Automatic switch over of application services which doesn’t compromise the
integrity of application data
– Selection of “active” data set occurs within the application automatically
► Exchange Disaster Recovery Definition
– Manual fail over of application services with high retention of data integrity
– Selection of “active” data set occurs manually outside the application,
Exchange application provides support to minimize data loss through
replication
Exchange 2010 High Availability
► Database Availability Group (DAG)
– A group of up to 16 Exchange Server 2010 Mailbox servers that
provide automatic database-level recovery
– Uses continuous log replication and a subset of Windows Failover
Clustering technologies
– Can extend across multiple datacenters/AD sites
Exchange 2010 High Availability
► Benefits of Exchange Native Data Protection
– Protection from database, server or network failure
– Automatic failover protection and manual switchover control is
provided at the mailbox database level instead of at the server
level.
– Support for up to 16 copies, support for lag copies
(c) 2011 Microsoft. All rights reserved.
Host Based Failover Clustering
► Host Based Failover Clustering HA
– Using Host Based Failover Clustering and automatically failing
VMs to an alternate cluster node in the event of a critical
hardware issue (virtualization platform independent)
► What you need to be aware of:
– Not an Exchange Aware Solution
– Only protects against server hardware/network failure
– No HA in the event of storage failure / data corruption
– Trend is larger mailboxes = larger database sizes = longer time to
recover from data loss = DAG
– Requires a shared storage deployment
VM Migration and Exchange 2010
► Physical Computer Maintenance
– Operating System/Application Updates
– Hardware Maintenance
► Rebalancing Workloads
– Dynamic redistribution of VM’s to optimize workload on physical
hardware
► Green IT
– ‘Off Peak’ Virtual Machine Consolidation
VM Cluster & Migration Considerations
► Minimize “outage” during migration operations
– Consider CSV rather than pass-through LUNs for all Mailbox VM
storage
► Disable migration technologies that save state and migrate: always
migrate live or completely shut down
► Consider relaxing cluster heartbeat timeouts
– Cluster nodes considered down after 5 seconds by default
► Be aware of additional network interface requirements for VM
migration technologies – size network appropriately
Best Practices:
Coexistence With Other Workloads
Private Cloud Considerations
► Given fixed resource requirements, isolate Exchange within private
cloud as much as possible
► Be prepared to apply different resource management polices to
Exchange VMs vs. other workloads which may be less mission critical
► Use private cloud as pre-built infrastructure, not necessarily dynamic
– Based on deployment sizing, understand overall resource
requirements and allocate accordingly from pool of cloud
resources
Resource Allocation & Balancing
► Disable hypervisor-based auto tuning features
– Dynamic memory
– Storage tuning/rebalancing
► Exchange Mailbox role IOPS heavily dependent on ESE cache,
dynamic memory can negatively impact
► Size for calculated resource requirements – no reliance on dynamic
tuning should be needed
Enhancing Exchange
on Hyper-V
System Center Integration
• Protect Exchange up to every 15 minutes
• ‘Lossless Restore’ – within 1 committed
transaction
• Understands & protects scenarios such as DAG
• Proactively monitor Exchange infrastructure
• Drive down the cost of ensuring service levels
• New Correlation Engine help to monitor more
efficiently
• Rapidly provision Exchange infrastructure
• Automate migrations for load balancing of VMs
• Seamlessly migrate workloads
Citrix NetScaler VPX
► Hyper-V Virtual Appliance
– Can be deployed through SCVMM
► Speed up deployment with application templates & configuration
wizards
► Once deployed, NetScaler will load balance Exchange 2010 Client
Access servers
► Ships with PRO-enabled OpsMgr MP
– Load balance, optimize, secure & report
– Dynamically scale infrastructure based on monitoring
Partner Evidence: Exchange Virtualization
► HP recommended configurations for Exchange 2010
virtualization using Hyper-V R2
– Sizing for 20,000 users, 512 MB mailboxes
• All roles virtualized and Live Migration for MBX servers
• Hardware: HP ProLiant BL460c G6 (BL460c G6) server blade and HP
StorageWorks 8400 Enterprise Virtual Array (EVA8400) storage systems
– http://h20195.www2.hp.com/V2/GetPDF.aspx/4AA1-2127ENW.pdf
– Sizing for 5,000 users, 256 MB mailboxes
• All roles virtualized and Live Migration for MBX servers
• Hardware: HP ProLiant BL460c G6 (BL460c G6) server blades and HP
LeftHand P4500 (P4500) storage systems
– http://h20195.www2.hp.com/V2/GetPDF.aspx/4AA1-9895ENW.pdf
Summary
► Why Microsoft Virtualization for Microsoft Server
Applications?
► Why Hyper-V?
► ESG Performance Results
► Exchange Server Virtualization Best Practices
► Surrounding Ecosystem
► Resources
Related Content
Breakout Sessions:
EXL303: Exchange Server 2010 : High Availability Concepts
EXL304 : Load Balancing with Microsoft Server 2010
EXL308 : Exchange Virtualisation : Is it right for you?
EXL306: Best Practices for Virtualization Of Microsoft Exchange 2010
SVR-VIR308 : iSCSI: Getting Best Performance, High-Availability and
Virtualisation
VIR307 : Understand How Microsoft Virtualisation Compares to VMWare
VIR/COS301 : Best Practises for Private Cloud Implementation
Track Resources
Cloud Power http://www.microsoft.com/cloud/
Private Cloud http://www.microsoft.com/privatecloud/
Windows Server http://www.microsoft.com/windowsserver/
Windows Azure - http://www.microsoft.com/windowsazure/
Microsoft System Center http://www.microsoft.com/systemcenter/
Enrol in Microsoft Virtual Academy Today
Why Enroll, other than it being free?
The MVA helps improve your IT skill set and advance your career with a free, easy to access training
portal that allows you to learn at your own pace, focusing on Microsoft technologies.
What Do I get for enrolment?
►
►
►
Free training to make you become the Cloud-Hero in my Organization
Help mastering your Training Path and get the recognition
Connect with other IT Pros and discuss The Cloud
Where do I Enrol?
www.microsoftvirtualacademy.com
Then tell us what you think. TellTheDean@microsoft.com
© 2010 Microsoft Corporation. All rights reserved. Microsoft, Windows, Windows Vista and other product names are or may be registered trademarks and/or trademarks
in the U.S. and/or other countries.
The information herein is for informational purposes only and represents the current view of Microsoft Corporation as of the date of this presentation. Because Microsoft must respond to changing
market conditions, it should not be interpreted to be a commitment on the part of Microsoft, and Microsoft cannot guarantee the accuracy of any information provided after the date of this
presentation. MICROSOFT MAKES NO WARRANTIES, EXPRESS, IMPLIED OR STATUTORY, AS TO THE INFORMATION IN THIS PRESENTATION.
(c) 2011 Microsoft. All rights reserved.
Resources
www.msteched.com/Australia
www.microsoft.com/australia/learning
Sessions On-Demand & Community
Microsoft Certification & Training Resources
http:// technet.microsoft.com/en-au
http://msdn.microsoft.com/en-au
Resources for IT Professionals
Resources for Developers
(c) 2011 Microsoft. All rights reserved.