Don’t Install WMF 3.0 On VMM Managed W2008 R2 Hyper-V Hosts

Microsoft has published a KB article (KB2795043) that explains the following scenarios:

On System Center Virtual Machine Manager, you may experience one of the following symptoms:

  • A Windows Server 2008 R2 SP1 Hyper-V host has a status of Needs Attention in the VMM Console.
  • or

  • Adding a Hyper-V host or cluster fails.

The fix is … to uninstall the WMF 3.0 update (KB2506143).  There’s a bit more to it than that. You also need to reboot the host and then run:

winrm qc

… and then do another reboot of the host.

I know; it’s far from an ideal situation.  But there’s the workaround for you.

KB2549448 – W2008 R2 Clustering Uses Default Time-Out After You Configure Regroup Timeout Setting

Microsoft issued another clustering update for Windows Server 2008 R2.  This one is where:

The Cluster service still uses the default regroup time-out value of 6 seconds after you configure the regroup time-out setting on a computer that is running Windows Server 2008 R2. This time-out value may not be long enough for some cluster nodes to rejoin a large scale cluster after you restart the Cluster service.
Note To obtain the time-out value (ClusSvcRegroupOpeningTimeout), type the following command at a command prompt, and then press Enter:

cluster /prop

Microsoft says you can fix this by Installing the available hotfix.

After you install the following hotfix package, the following regroup settings become effective:

  • ClusSvcRegroupOpeningTimeout
  • ClusSvcRegroupPruningTimeout
  • ClusSvcRegroupStageTimeout
  • ClusSvcRegroupTickInMilliseconds

KB2549472 – W2008 R2 Cluster Node Cannot Rejoin Cluster After Node is Restarted Or Removed From Cluster

Microsoft issued a hotfix over the holidays to deal with an issue where:

  • You configure a failover cluster that has some computers that are running Windows Server 2008 R2.
  • One cluster node is offline or is removed from the cluster.
  • You try to rejoin the cluster node to the cluster after a while.

In this scenario, the cluster node cannot rejoin the cluster.

According to Microsoft, the cause is that:

When a cluster node tries to rejoin a cluster after it is restarted or removed from the cluster, the virtual IP address of the cluster node may be changed. However, other cluster nodes still use old route records in their route database. Therefore, the cluster node cannot rejoin the cluster.

A supported hotfix is available from Microsoft.

KB2741477 – Can’t Add NICs To VM on W2008 R2 Clustered Host

Microsoft has released a hotfix for when you cannot add network adapters to a virtual machine on a Windows Server 2008 R2-based failover cluster.

Consider the following scenario:

  • You create a failover cluster on some computers that are running Windows Server 2008 R2 SP1. 
  • You enable the Hyper-V role on both cluster nodes.
  • You configure a virtual machine as a cluster resource.
  • You try to add a network adapter to the virtual machine by using the Refresh Virtual Machine Configuration wizard.

In this scenario, you cannot add the network adapter successfully until the virtual machine is removed from the cluster.

A supported hotfix is available from Microsoft.

KB2575625 – W2008 R2 Cluster Starts Failover After 80 Seconds When You Shutdown Active Node

Microsoft has released a hotfix for when the cluster service initiates a failover after a delay of about 80 seconds when you shutdown the active node in Windows Server 2008 R2.

Consider the following scenario:

  • You configure a failover cluster on some computers that are running Windows Server 2008 R2.
  • You add a crossover cable between two nodes of the cluster.
  • You shutdown the active node of the cluster.

In this scenario, the Cluster service initiates a failover after a delay of about 80 seconds.
Note The Cluster service usually initiates the failover within 5 seconds.

This issue occurs because the Cluster service incorrectly waits for Global Update Manager (GUM) updates.

A hotfix has been released to fix this issue.

Expanded Supported Guest Operating Systems on WS2012 Hyper-V

The support for Windows Server 2012 guest OSs has been expanded by Microsoft.  On the server OS side the changes are mostly on the Linux side; OpenSuse and Ubuntu are now listed as supported. Previously they were not supported but they worked.

Guest operating system (server)

Maximum number of virtual processors

Notes

Windows Server 2012

64

Integration services do not require a separate installation because they are built-in.

Windows Server 2008 R2 with Service Pack 1 (SP 1)

64

Datacenter, Enterprise, Standard and Web editions. Install the integration services after you set up the operating system in the virtual machine.

Windows Server 2008 R2

64

Datacenter, Enterprise, Standard and Web editions. Upgrade the integration services after you set up the operating system in the virtual machine.

Windows Server 2008 with Service Pack 2 (SP 2)

8

Datacenter, Enterprise, Standard and Web editions (32-bit and 64-bit). Install the integration services after you set up the operating system in the virtual machine.

Windows Home Server 2011

4

Edition information is not applicable. Install the integration services after you set up the operating system in the virtual machine.

Windows Small Business Server 2011

Essentials edition – 2

Standard edition – 4

Essentials and Standard editions. Install the integration services after you set up the operating system in the virtual machine.

Windows Server 2003 R2 with Service Pack 2 (SP2)

2

Standard, Web, Enterprise, and Datacenter editions (32-bit and 64-bit). Install the integration services after you set up the operating system in the virtual machine.

Windows Server 2003 with Service Pack 2

2

Standard, Web, Enterprise, and Datacenter editions (32-bit and 64-bit). Install the integration services after you set up the operating system in the virtual machine.

CentOS 5.7 and 5.8

64

Download and install Linux Integration Services Version 3.4 for Hyper-V.

CentOS 6.0 – 6.3

64

Download and install Linux Integration Services Version 3.4 for Hyper-V.

Red Hat Enterprise Linux 5.7 and 5.8

64

Download and install Linux Integration Services Version 3.4 for Hyper-V.

Red Hat Enterprise Linux 6.0 – 6.3

64

Download and install Linux Integration Services Version 3.4 for Hyper-V.

SUSE Linux Enterprise Server 11 SP2

64

Integration services do not require a separate installation because they are built-in.

Open SUSE 12.1

64

Integration services are built-in and do not require a separate download and installation.

Ubuntu 12.04

64

Integration services are built-in and do not require a separate download and installation.

Note that Linux (physical or virtual) requires numa=off in the grub file if you have more than 7 CPUs or 30 GB RAM in the (virtual) machine. That’s a Linux thing, not a Hyper-V or virtualisation thing.

The client OS list now is:

Guest operating system (client)

Maximum number of virtual processors

Notes

Windows 8

32

Integration services do not require a separate installation because they are built-in.

Windows 7 with Service Pack 1 (SP 1)

4

Ultimate, Enterprise, and Professional editions (32-bit and 64-bit). Upgrade the integration services after you set up the operating system in the virtual machine.

Windows 7

4

Ultimate, Enterprise, and Professional editions (32-bit and 64-bit). Upgrade the integration services after you set up the operating system in the virtual machine.

Windows Vista with Service Pack 2 (SP2)

2

Business, Enterprise, and Ultimate, including N and KN editions. Install the integration services after you set up the operating system in the virtual machine.

Windows XP with Service Pack 3 (SP3)

2

Professional. Install the integration services after you set up the operating system in the virtual machine.

Windows XP x64 Edition with Service Pack 2 (SP 2)

2

Professional. Install the integration services after you set up the operating system in the virtual machine.

You can see the supported list of guest OSs for the legacy versions of Hyper-V (2008, 2008 R2) here.

KB2712156–AVX Does Not Work In AM On W2008 R2Host With AMD CPU That Is AVX Capable

Microsoft released a hotfix to deal with an issue where AVX instructions do not work in a virtual machine on a Windows Server 2008 R2-based computer that has an AMD CPU which supports the AVX feature.

Consider the following scenario:

  • An AMD CPU that supports the Advanced Vector Extensions (AVX) feature is installed in a computer that is running Windows Server 2008 R2.
    Note AMD introduces support for the AVX feature in Bulldozer-based multicore processors.
  • You install the Hyper-V server role on the computer.
  • You create a virtual machine on the computer, and then you start the virtual machine.

In this scenario, AVX instructions do not work as expected on the virtual machine.  This issue occurs because the CPUID maximum function limit is not updated to reflect the support for the AVX feature on AMD processors.

A supported hotfix is available from Microsoft.

Choosing Hardware Or Software Functionality In Virtualisation

I say this a lot: “The great thing about Windows Server 2012 is that you have options”. It’s true, so very true.

Software can only ever do so much until it impacts the performance of the host.  Hardware will always do things more efficiently, enabling lower latency, more throughput, and allowing greater scalability.  Windows Server 2008 R2 recognised this by giving us some hardware offloads:

  • Virtual Machine Queuing (VMQ) improved inbound networking for VMs
  • Second Level Address Translation (SLAT – Intel EPT or AMD NPT/RVI) greatly improved the overall performance of memory paging intensive VMs by offloading VM to host memory mapping to the CPU rather than having the hypervisor do it.

This all continues in WS2012:

  • Receive Side Scaling (RSS): Uses queues on the NIC and scales out beyond core 0 on the CPU to enable greater scalability for network processing
  • Dynamic Virtual Machine Queuing (DVMQ): Using the same queues as RSS (and therefore being incompatible on the same NICs) it does something similar to RSS but for virtual machine traffic.
  • IPsec Task Offload: Policy driven network encryption that is offloaded to capable NICs
  • Datacenter Bridging (DCB): Protocol-based QoS can be offloaded to the NIC for non-virtual NIC traffic, giving better performance than by using the OS Packet Scheduler (and being able to do QoS and Priority Flow Control for RDMA)

Single Root I/O Virtualisation (SR-IOV) is a great example of the pros and cons of software versus hardware functionality.  Without SR-IOV we have traditional virtual switch networking:

image

Packets come in the NIC, through the drivers, into the virtual switch for filtering and routing, back down the stack into the VM Bus (ring –1 and DEP/No Execute) and up into the virtual NIC in the VM.  It’s software processing the entire way in and out again.

Now compare it with SR-IOV enabled on a host and VM:

image

You get something weird: the VM now has hardware connectivity to a Virtual Function (VF) on the NIC.  There is on VF used for every vNIC and a pNIC has a number of VFs. Now software has been removed from the process of copying packets to the vNIC in the VM.  Latency is improved, and the scalability of the host is improved too (more CPU available to apps in VMs).

The Pros of SR-IOV are easy to see.  But do you need it?  Maybe if you need the lowest possible latency.  Maybe if you’re going to have hugely dense hosts or network processing could spike your CPUs.  These are choices that we make for ESXi and Hyper-V.  But where ESXi cannot vMotion one of these VMs, we can with Hyper-V.  That’s nice. 

On the con side, you cannot team the pNIC when usign SR-IOV because that NIC team (in the management OS) is completely bypassed.  You’d have to create a NIC team in every VM … yuck!  Double-yuck if deploying self-service clouds.  So it is a trade-off.

You see a similar thing with DR replication.  I love the concept of setting a SAN to replicate LUNs to another location and just leaving it.  But it’s just not that simple.

On the pro side, SAN replication is set-and-forget (to some extent, assuming you monitor and manage by exception).  That’s great for a cloud.  VMs or services are put into one tier of storage and they auto-replicate to the DR site.  Put into another tier and they are not.  As a former hosting engineer, I love that.

On the con side, SAN replication is expensive (not so good in public cloud).  It’s not only vendor lock in, but it’s usually SAN model lock in.  I don’t love that.  And while multi-suite clusters are a great idea, I think the WAN, physical networking, storage, servers, and virtualisation solution is one of the most complicated things you can build in IT infrastructure.  While a consultant might be up to that, are the admins who are left holding the baby skilled enough to maintain it?

Hyper-V Replica is built-in and free to use.  It’s very granular, and it’s designed to be able to work on commercial broadband.  I love that.  There’s no vendor lock in, and no requirement for partners.  Another plus.  It’s simple to set up and maintain.  But there are scalability/performance considerations to any host-based replication.  Hyper-V Replica also has resource requirements.  That’s always the way with software.  SAN based replication will not have those same host resource requirements, but it has other complexity/budgetary requirements.

Choosing the right solution for any site is a classic consultant’s “it depends”.  There is never one right answer.  Know the requirements of the customer/site, know the possible solutions, and find the best fit.

KB2652137 – Communications Fail When You Use W2008 R2 Provider Package With WS2012 iSCSI Target

Another hotfix last night, this time for a scenario when communications fail when you try to use the Windows Server 2008 R2 provider package to communicate with a Windows Server 2012 iSCSI target.

You have a Windows Server 2008 or a Windows Server 2008 R2 server that runs applications such as Microsoft SQL Server. You have a Windows Server 2012 server that is configured for the iSCSI Software Target. When you try to use the Windows Server 2008 or the Windows Server 2008 R2 provider package to communicate with the iSCSI target, communications fail.

This problem occurs because the DCOM Remote Protocol is no longer used for the iSCSI Software Target in Windows Server 2012. The WMI interfaces are now used in the provider to communicate with the iSCSI target.

The resolution is to:

To resolve this problem, install a Windows Server 2012-aware provider package on the iSCSI initiator. The new provider package implements the iSCSI Software Target WMI Provider to communicate with the iSCSI target service.

The update, “iSCSI Target Storage Providers (VDS/VSS) for downlevel application servers”, supports installation on Windows Server 2008 Service Pack 2 (SP2) or Windows Server 2008 R2 Service Pack 1 (SP1).

EDIT#1:

If you are installing the WS2012-aware provider package on down level operating systems then you really should read this blog post by Jane Yan, paying particular attention to the credential configuration step.  Credit: Andreas Erson.

KB2727972–W2008 R2 Cluster Node Crashes When You Restart A Node

Microsoft has posted a hotfix for a scenario where a cluster node crashes when you restart a computer in a Windows Server 2008 R2 environment.

  • You deploy a failover cluster in a Windows Server 2008 R2 environment.
  • You enable the Volume Shadow Copy Service on a cluster disk.
  • You replace a host bus adapter (HBA) on a cluster node.
  • You restart the cluster node.

In this scenario, the cluster node crashes.  This issue is triggered by a deadlock that occurs when the computer restarts.

A supported hotfix is available from Microsoft.