Archive

Posts Tagged ‘Management’

I am Speaking at Teched Australia 2011

June 29, 2011 Leave a comment

I am absolutely thrilled to announce I will be presenting the following two sessions at Tech.Ed Australia 2011 :

SCVMM 2012: Deployment, Planning, Upgrade

This session provides a scenario rich detailed walk through of VMM 2012 deployment, planning, and upgrade scenarios. Come and learn how to best plan your next VMM rollout

SCVMM 2012 Fabric Lifecycle: Networking and Storage

This session provides a scenario rich detailed walk through of  new and more robust networking and storage features in VMM 2012. In this session you will learn how to discover, configure, and provision networking
and storage fabric for use with the private cloud

Came along! It will be an excellent session.

Tech.Ed Australia 2011 is on the Gold Coast between the 30th August and the 2nd September, registrations are now open. Find out more at http://australia.msteched.com/

Advertisements

Validate SCSI Device Vital Product Data (VPD) test fails after you install W2008 R2 SP1

June 22, 2011 Leave a comment

If you found this error :” Failed to get SCSI page 83h VPD descriptors for cluster disk
<number> from <node name> status 2″ after applying SP1 to your W2008R2 cluster, Microsoft has released a fix for it.

The List Potential Cluster Disks storage validation test may display a warning message that resembles the following: “Disk with identifier <value> has a Persistent Reservation on it. The disk might be part of some other cluster. Removing the disk from validation set”

The  hotfix resolves an issue in which the storage test incorrectly runs on disks that are online and not in the Available Storage group.

More details:

You configure a failover cluster that has three or more nodes that are running Windows Server 2008 R2 Service Pack 1 (SP1).

You have cluster disks that are configured in groups other than the Available Storage group or that are used for Cluster Shared Volumes (CSV).

These disks are online when you run the Validate SCSI Device Vital Product Data (VPD) test or the List Potential Cluster Disks storage validation test

More info : http://support.microsoft.com/kb/2531907

 

Plan your organization’s migration to a private cloud with the Hyper-V Cloud Fast Track Assessment!

June 17, 2011 Leave a comment

Use the MAP Toolkit to plan your organization’s migration to a private cloud with the Hyper-V Cloud Fast Track Assessment!

New MAP features  to:

  • Build portfolios of web applications and databases to migrate to Windows Azure and SQL Azure.
  • Assess your environment’s readiness for Office 365 or Internet Explorer 9
  • Identify and migrate databases from competing platforms like Oracle and MySQL to Microsoft SQL Server.
  • Consolidate your servers on to Hyper-V Cloud Fast Track Infrastructures

The beta of the MAP Toolkit v6.0 is now available. To get involved in the beta program

https://connect.microsoft.com/

 

 

SCVMM 2008 Ports and Protocols

June 7, 2011 Leave a comment

SCVMM 2008, SCVMM 2008 R2 and SCVMM 2008 R2 SP1 default ports :

Connection type Protocol Default port Where to change the port setting
VMM server to VMM agent on Windows Server–based host (control) WS-Management 80 at VMM setup, registry
VMM server to VMM agent on Windows Server–based host (file
transfers)
HTTPS (using BITS) 443 (Maximum value: 32768) Registry
VMM server to remote Microsoft SQL Server database TDS 1433 Registry
VMM server to P2V source agent DCOM 135 Registry
VMM Administrator Console to VMM server WCF 8100 at VMM setup, registry
VMM Self-Service Portal Web server to VMM server WCF 8100 at VMM setup
VMM Self-Service Portal to VMM self-service Web server HTTPS 443 at VMM setup
VMM library server to hosts BITS 443 (Maximum value: 32768) at VMM setup, registry
VMM host-to-host file transfer BITS 443* (Maximum value: 32768)

* VMM 2008 R2 : port 30443 (http://support.microsoft.com/kb/971816)

Registry
VMRC connection to Virtual Server host VMRC 5900 VMM Administrator Console, registry
VMConnect (RDP) to Hyper-V hosts RDP 2179 VMM Administrator Console, registry
Remote Desktop to virtual machines RDP 3389 Registry
VMware Web Services communication HTTPS 443 VMM Administrator Console, registry
SFTP file transfer from VMWare ESX Server 3.0 and VMware ESX Server 3.5
hosts
SFTP 22 Registry
SFTP file transfer from VMM server to VMWare ESX Server 3i hosts HTTPS 443 Registry

More info  : http://technet.microsoft.com/en-us/library/cc764268.aspx

 

SCVMM 2012 Management ports and protocols. Detailed

June 7, 2011 Leave a comment

Here are the list of ports/protocols for the new SCVMM 2012.

From To Protocol Default
port
Where to change port setting
VMM management server P2V
source agent (control channel)
DCOM 135
Load Balancer HTTP/HTTPS 80/443 Load balancer configuration provider
WSUS server (data channel) HTTP/HTTPS 80/8530
(non-SSL), 443/8531 (with SSL)
These ports are the IIS port binding with WSUS. They cannot be changed from VMM.
WSUS server (control channel) HTTP/HTTPS 80/8530
(non-SSL), 443/8531 (with SSL)
These ports are the IIS port binding with WSUS. They cannot be changed from VMM.
VMM agent on Windows Server–based host (data
channel for file transfers)
HTTPS
(using BITS)
443
(Maximum value: 32768)
Citrix XenServer host (customization data
channel)
iSCSI 3260 On XenServer in transfer VM
XenServer host (control channel) HTTPS 5989 On XenServer host in: /opt/cimserver/cimserver_planned.conf
remote Microsoft SQL Server database TDS 1433
VMM agent on Windows Server–based host (control
channel)
WS-Management 5985 VMM setup
VMM agent on Windows Server–based host (control
channel – SSL)
WS-Management 5986
in-guest agent (VMM to virtual machine control
channel)
WS-Management 5985
Storage Management Service WMI Local
call
Cluster PowerShell interface PowerShell n/a
P2V source agent (data channel) BITS User-Defined P2V cmdlet option
VMM library server hosts
file transfer
BITS 443
(Maximum value: 32768)
VMM setup
VMM host-to-host file transfer BITS 443
(Maximum value: 32768)
VMM Self-Service Portal VMM
Self-Service Portal web server
HTTPS 443 VMM setup
VMM Self-Service Portal web server VMM
management server
WCF 8100 VMM setup
Console connections (RDP) virtual
machines through Hyper-V hosts (VMConnect)
RDP 2179 VMM console
Remote Desktop virtual
machines
RDP 3389 On the virtual machine
VMM console VMM
management server
WCF 8100 VMM setup
VMM management server (HTTPS) WCF 8101 VMM setup
VMM management server (NET.TCP) WCF 8102 VMM setup
VMM management server (HTTP) WCF 8103 VMM setup
Windows PE agent VMM
management server (control channel)
WCF 8101 VMM setup
VMM management server (time sync) WCF 8103 VMM setup
WDS provider VMM
management server
WCF 8102 VMM setup
Storage Management Service  SMI-S Provider CIM-XML Provider-specific
port
VMM management server VMware
ESX Server 3i hosts
HTTPS 443

Others

Connection Type Protocol Default port Where to change port setting
OOB Connection – SMASH over WS-Man HTTPS 443 On BMC
OOB Connection IPMI IPMI 623 On BMC
BITS port for VMM transfers (data channel)
BITS 443 VMM setup
VMware ESX Server 3.0 and VMware ESX Server 3.5 hosts SFTP 22
VMware Web Services
communication
HTTPS 443 VMM console

Note: When you install the VMM management server you can assign some of the ports that it will use for communications and file transfers between the VMM components.

Hyper-V : Best Practices and Supported scenarios regarding Exchange Server 2010

May 19, 2011 Leave a comment

 The following are the supported scenarios for Exchange 2010 SP1 :

  • The Unified Messaging server role is supported in a virtualized environment.
  • Combined Exchange 2010 high availability solutions (database availability groups (DAGs)) with hypervisor-based clustering, high availability, or migration solutions that will move or automatically failover mailbox servers that are members of a DAG between clustered root servers

HyperV Guest Configuration

Keep in mind that because there are no routines within Exchange Server that test for a virtualized platform, Exchange Server behaves no differently programmatically on a virtualized platform than it does on a physical platform.

Determining Exchange Server Role Virtual Machine Locations

When determining Exchange Server Role virtual machine locations, consider the following general best practices:

  • Deploy the same Exchange roles across multiple physical server roots (to allow for load balancing and high availability).
  • Never deploy Mailbox servers that are members of the same Database Availability Groups (DAGs) on the same root.
  • Never deploy all the Client Access Servers on the same root.
  • Never deploy all the Hub Transport servers on the same root.
  • Determine the workload requirements for each server and balance the workload across the HyperV guest virtual machines.

Guest Storage

Each Exchange guest virtual machine must be allocated sufficient storage space on the root virtual machine for the fixed disk that contains the guest’s operating system, any temporary memory storage files in use, and related virtual machine files that are hosted on the root machine.Consider the following best practices when configuring Hyper-V guests:

  • Fixed VHDs are recommended for the virtual operating system.
  • Allow for a minimum of a 15-GB disk for the operating system, allow additional space for the paging file, management software, and crash recovery (dump) files. Then add Exchange server role space requirements.
  • Storage used by Exchange should be hosted in disk spindles that are separate from the storage that hosts the guest virtual machine’s operating system.
  • For Hub Transport servers, correctly provision the necessary disk space needed for the message queue database, and logging operations.
  • For Mailbox servers, correctly provision the necessary disk space for databases, transaction logs, the content index, and other logging operations. .

Guest Memory : Dynamic Memory should be disabled
Memory must be sized for guest virtual machines using the same methods as physical computer deployments. Exchange—like many server applications that have optimizations for performance that involve caching of data in memory—is susceptible to poor system performance and an unacceptable client experience if it doesn’t have full control over the memory allocated to the physical computer or virtual machine on which it is running.
Many of the performance gains in recent versions of Exchange, especially those related to reduction in input/output (I/O) are based on highly efficient usage of large amounts of memory. When that memory is no longer available, the expected performance of the system can’t be achieved. For this reason, memory oversubscription or dynamic adjustment of virtual machine memory must be disabled for production Exchange servers.

Deployment Recommendations

When designing an Exchange Server 2010 virtualized environment, the core Exchange design principles apply. The environment must be designed for the correct performance, reliability, and capacity requirements. Design considerations such as examining usage profiles, message profiles, and so on must still be taken into account.

See this article (Mailbox Storage Design Process) as a starting point when considering a high availability solution that uses DAGs.

Because virtualization provides the flexibility to make changes to the design of the environment later, some organizations might be tempted to spend less time on their design at the outset. As a best practice, spend adequate time designing the environment to avoid pitfalls later.

Group the Exchange Server roles in such a way that balances workloads on the root servers. Mixing both roles on the same HyperV root server can balance the workloads and prevent one physical resource from being unduly stressed, rather than if the same roles were put on the same hosts

The updated support guidance applies to any hardware virtualization vendor participating in the Windows Server Virtualization Validation Program (SVVP).

Best Practices for Virtualizing Exchange Server 2010 with Windows Server 2008 R2 Hyper-Vwhitepaper. This whitepaper is designed to provide technical guidance on Exchange server roles, capacity planning, sizing and performance, as well as high availability best practices.

Complete system requirements for Exchange Server 2010 running under hardware virtualization software can be found in Exchange 2010 System Requirements. Also, the support policy for Microsoft software running in non-Microsoft hardware virtualization software can be found here.

CentOS now have official support as guest VM in Hyper-V

May 18, 2011 2 comments

Effective immediately, Microsoft will support Windows Server2008 R2 Hyper-V to run CentOS.  

CentOS is a popular Linux distribution for Hosters, and this was the number one requirement for interoperability that we heard from that community.

This development will enable MS Hosting partners to consolidate their mixed Windows + Linux infrastructure on Windows Server Hyper-V; reducing cost and complexity, while betting on an enterprise class virtualization platform. .

How will support work?
Call Microsoft CSS. Support will cover installation issues as well as configuration issues.

What version of the Linux Integration Services support CentOS?

 The existing Hyper-V Linux Integration Services for Linux Version 2.1 support CentOS. The following features are included in the Hyper-V Linux Integration Services 2.1 release:

· Symmetric Multi-Processing (SMP) Support: Supported Linux distributions can use up to 4 virtual processors (VP) per virtual machine.

· Driver support for synthetic devices: Linux Integration Services supports the synthetic network controller and the synthetic storage controller that were developed specifically for Hyper-V.

· Fastpath Boot Support for Hyper-V: Boot devices take advantage of the block Virtualization Service Client (VSC) to provide enhanced performance.

· Timesync: The clock inside the virtual machine will remain synchronized with the clock on the host.

· Integrated Shutdown: Virtual machines running Linux can be gracefully shut down from either Hyper-V Manager or System Center Virtual Machine Manager.

· Heartbeat: Allows the host to detect whether the guest is running and responsive.

· Pluggable Time Source: A pluggable clock source module is included to provide a more accurate time source to the guest.

The Linux Integration Services are available via the Microsoft Download Center here: http://www.microsoft.com/downloads/details.aspx?displaylang=en&FamilyID=eee39325-898b-4522-9b4c-f4b5b9b64551

 From Wikipedia:

CentOS is a community-supported, mainly free software operating system based on Red Hat Enterprise Linux (RHEL). It exists to provide a free enterprise class computing platform and strives to maintain 100% binary compatibility with its upstream distribution. CentOS stands for Community ENTerprise Operating System.

Red Hat Enterprise Linux is available only through a paid subscription service that provides access to software updates and varying levels of technical support. The product is largely composed of software packages distributed under either an open source or a free software license and the source code for these packages is made public by Red Hat.

CentOS developers use Red Hat’s source code to create a final product very similar to Red Hat Enterprise Linux. Red Hat’s branding and logos are changed because Red Hat does not allow them to be redistributed.

CentOS is available free of charge. Technical support is primarily provided by the community via official mailing lists, web forums, and chat rooms. The project is not affiliated with Red Hat and thus receives no financial or logistical support from the company; instead, the CentOS Project relies on donations from users and organizational sponsors.