BP Virtualization PDF
BP Virtualization PDF
Trademarks
The trademarks, service marks, logos and graphics (collectively “Trademarks”) appearing on Mitel's
Internet sites or in its publications are registered and unregistered trademarks of Mitel Networks
Corporation (MNC) or its subsidiaries (collectively “Mitel”) or others. Use of the Trademarks is prohibited
without the express consent from Mitel. Please contact our legal department at [email protected] for
additional information. For a list of the worldwide Mitel Networks Corporation registered trademarks, please
refer to the website: http://www.mitel.com/trademarks.
Chapter 1 :
Introduction
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
Benefits of virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
Software requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
VMware software requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
Hyper-V software requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Mitel Product version support for VMware and Hyper-V . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Hardware requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
Chapter 2 :
Deploying on a VMware infrastructure
Deployment considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
Deployment types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
Preparing the VMware infrastructure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
Host servers and clustering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
Networking . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
VMware support by Mitel product release . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
Host server sizing and resource management considerations . . . . . . . . . . . . . . . . . . . . . . . . . 21
Increasing resources for larger dimensions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
CPU and memory resource reservation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
Resource dimensioning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
Storage provisioning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
Resource sharing guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
Other virtual appliance management considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
VMware features. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
iii
Mitel Virtual Appliance Deployment
vSphere Replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
VMware Fault Tolerance (FT) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
VMware Distributed Resource Scheduler (DRS) and vMotion . . . . . . . . . . . . . . . . . . . . . . . . . 45
Chapter 3 :
Deploying on Microsoft Hyper-V Infrastructure
Chapter 4 :
Installing and Upgrading
iv
Table of Contents
Chapter 5 :
Maintenance and Troubleshooting
Maintenance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
Backups . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
Backup Methods in Hyper-V . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
Backup methods in VMware . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
VMware Data Protection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
Backup routine recommendations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
Restores . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
MiCollab Virtual system recovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
Monitoring resource usage and performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89
Troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
Troubleshooting virtual appliances in Hyper-V environment . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
Troubleshooting virtual appliances in VMware environment . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
Mitel Virtualization Diagnostic Tool (VDT) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
Collecting VMware logs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91
v
Mitel Virtual Appliance Deployment
Mitel virtual machines are blocked during power up due to checking file systems unexpected
inconsistency error . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106
Mitel virtual machines experience issues on power up after powering off . . . . . . . . . . . . . 107
Mitel resiliency solutions not working (such as MiVoice Business resiliency or MBG clustering)
107
Mitel application message record/playback or voice prompt issues . . . . . . . . . . . . . . . . . . 107
Mitel application repeatedly resets after startup, or voice quality issues . . . . . . . . . . . . . . 108
APPENDIX A:
GLOSSARY
APPENDIX B:
VMWARE VSPHERE STORAGE SETUP GUIDELINES
VMware vSphere storage setup guidelines for Mitel virtual appliances . . . . . . . . . . . . . . . . . . . 121
Storage virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
Hard disk drive types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
Mechanical hard disk drive . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122
Solid state disk (SSD) drive . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122
Types of storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122
Local storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122
Networked storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122
RAID levels . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123
RAID 0 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124
RAID 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124
RAID 10 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124
RAID 5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124
RAID 6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124
Virtual disk modes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124
Independent persistent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .124
Independent non-persistent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
Dependent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
Virtual disk types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
Thick-provisioned . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
Thin-provisioned . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
Partition alignment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
SAN multipathing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
File systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
VMFS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
NFS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
Raw Disk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
vi
Table of Contents
Recommendations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
Storage network . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
RAID level . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127
Virtual disk mode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127
Virtual disk types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127
Partition alignment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127
SAN multipathing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127
File system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127
Other considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
CPU utilization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
Separate physical NIC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
Jumbo frames . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
vSphere flash read cache . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
Storage hardware acceleration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
VMware-specific optimizations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
vSphere pluggable storage architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
Reference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
APPENDIX C:
VMWARE VCLOUD DIRECTOR
vii
Mitel Virtual Appliance Deployment
APPENDIX D:
SECURITY IN MITEL/VMWARE VIRTUAL NETWORKS
viii
List of Tables
ix
List of Tables
x
Chapter 1
INTRODUCTION
Mitel Virtual Appliance Deployment
2
Introduction
Introduction
This guide describes the equipment, network, and configuration considerations that apply when
setting up Mitel virtual appliances on servers enabled with VMware® vSphere™ virtualization
or Microsoft Hyper-V.
Note: This guide provides information about VMware and Hyper-V applications and
requirements that were valid for previous releases, and which may not be up-to-date.
Always refer to your current VMware or Hyper-V documentation for the latest information.
Supported in VMware:
• MiVoice Business Virtual
• MiCollab Virtual
• MiVoice Border Gateway Virtual
• MiContact Center Office
• MiContact Center Business
• MiContact Center Enterprise
• MiVoice Call Recording
• Virtual Open Integration Gateway (vOIG)
• Virtual Contact Center (vCC), Virtual MiVoice Call Accounting, Virtual IVR, Virtual Contact
Center Bundle
• MiVoice Business Express
• MiCloud Management Portal (formerly Oria)
• MiCloud Management Gateway
• MiVoice Office 250
• MiVoice MX-ONE
• MiVoice 5000
• MiVoice Office 400
• Mitel Mass Notification
Note:
1. MiCollab Client Virtual has been discontinued as a standalone product. The server
function for all of the clients is now included in MiCollab.
2. MiCollab Outlook Plugin and MiCollab Audio, Web and Video Conference sharing
are not supported in the virtual environment.
Supported in Hyper-V:
• MiVoice Business Virtual
• MiCollab Virtual
3
Mitel Virtual Appliance Deployment
Note: MiCollab Client Virtual has been discontinued as a standalone product. The server
function for all of the clients is now included in MiCollab.
The following Mitel applications are available as VMware-Ready® applications that you can
install in a virtual machine:
• MiContact Center (also available as a virtual appliance, as shown in the list above)
• MiVoice Business Reporting (also available as a virtual appliance, as shown in the list
above)
• MiVoice Enterprise Manager
Benefits of virtualization
Virtualization can offer significant benefits to a data center; they are described on the VMware
and Hyper-V Web sites, and summarized here:
• Reduced capital expenditure on physical servers
Dedicated physical servers or appliances are not required for each application, resulting in
server consolidation and hardware savings when using a virtual infrastructure. In addition,
multiple operating systems can run on a single server.
• Reduced operations and maintenance costs
Having fewer servers reduces time, effort, and cost for server management. Placing Mitel
virtual appliances in the virtual infrastructure with other virtualized business applications
provides further cost savings through integrated IT processes.
• Reduced power consumption
Virtual data centers have inherent power savings due to the reduced number of servers.
In VMware deployments, VMware Distributed Power Management (DPM) and the vMotion
feature reduce the number of servers running at any given time.
• Improved application availability
VMware offers several high availability features. One example is the VMware High
Availability feature, which can monitor server and virtual appliance health. In the case of a
host server failure, the virtual appliances can be restarted on a different server, minimizing
down time. When server maintenance is necessary, Mitel virtual appliances can be moved
to another server using vMotion, avoiding a service outage.
Hyper-V availability features are not available for Mitel products at this time.
4
Introduction
By consolidating Mitel virtual appliances within the customer's virtual infrastructure, the
disaster recovery features provided by VMware will be applied consistently to all of the
critical services.
Hyper-V availability features are not available for Mitel products at this time.
• Seamless telephony integration when using Mitel virtual appliances
Mitel’s adoption of virtualization has allowed continued compatibility and the ability to
integrate with existing customer network and telephony infrastructure, just as Mitel physical
server-based applications such as the 3300 ICP do.
• Faster system deployment
Mitel virtual appliances are inherently simpler to deploy into an existing data center.
Increasing communication capacity or adjusting overall workload is easier and faster than
the equivalent operations in traditional hardware and dedicated servers.
• Update roll-back
VMware features provide methods to quickly and easily restore a virtual appliance to its
exact state before an upgrade, should a problem arise during an application or virtual
appliance upgrade. On a physical system, a failed upgrade usually requires rebuilding the
system and performing a data restore.
Hyper-V features are not available for Mitel products at this time.
This Solutions Guide discusses configuration and performance at the general solution level.
The planning and configuration guidelines discussed here apply to the deployment of all Mitel
virtual products. Product installation and configuration procedures and engineering guidelines
are described in the product documentation for each individual product. See the Documentation
page on Mitel OnLine.
For more information about designing multi-node Mitel networks, both physical MiVoice
Business-based and MiVoice Business Virtual-based, refer to the Multi-Node Networking
Solutions Guide.
5
Mitel Virtual Appliance Deployment
Software requirements
Microsoft Hyper-V and VMware have specific software requirements, as detailed in the following
sections:
• “VMware software requirements” on page 6
• “Hyper-V software requirements” on page 7
6
Software requirements
Notes:
1. For a breakdown of the VMware products available for use with of the Mitel virtual appliances, see
Table 7.
2. VMware also provides a bundled product collection known as vCloud Suite which contains several
of the specific products listed in the table (vSphere, vCloud Networking and Security, vCloud
Director, Site Recovery Manager). vCloud Suite also contains several products not listed in the
table (vCenter Operations Manager, vCloud Automation Center).
See http://www.vmware.com/products/vCloud-suite. Availability of features and products in the
vCloud Suite lineup for use with Mitel products are as listed in the table.
Table 3 shows the Mitel product versions supported for use in a virtualized data center.
7
Mitel Virtual Appliance Deployment
PREVIOUS RELEASES
MiVoice Business Virtual / vMCD all releases since 6.0 SP2 all releases since 7.02
MiVoice Border Gateway Virtual all releases since 7.1 all releases since 8.0 SP1
MiCollab Virtual / vMAS all releases since 5.0 all releases since 6.0
MiCollab Client Virtual / vUCA all releases since 6.0 SP1 all releases since 6.0 SP4
NuPoint Unified Messaging Virtual all releases since 6.0 all releases since 7.0
MiContact Center Virtual (MiCC Virtual) all releases since 7.1 SP1 all releases since 7.1 SP1
MiContact Center Enterprise 9.2 not supported
MiContact Center Business Virtual 8.1 and Service Packs 8.1
Virtual Customer Service Manager (vCSM) all releases since 6.0 not supported
Virtual Open Integration Gateway (vOIG) all releases since 2.0 not supported
MiContact Center Office Virtual all releases since 6.1 SP1 not supported
Oria all releases since 5.0 not supported
MiVoice Call Recording all releases since 8.1 SP1 not supported
MiVoice Business Express all releases since 6.0 all releases since 7.1
Note: Refer to the Engineering Guidelines for each product; see Table 9 on page 35.
8
Hardware requirements
Hardware requirements
VMware hardware requirements:
At minimum, Mitel virtual appliances running on both VMware and Hyper-V virtual machines
require:
• Currently shipping Intel Xeon E3v2 / E5 / E7 Server Class series processors with a minimum
of 4 cores, 2 GHz, Hyper-threading enabled, and Extended Page Table support
(http://ark.intel.com/)
• Legacy server technology may also be used but must meet the minimum requirements
shown in Table 4:
Notes:
1. Intel VT consists of Intel VT-x in the processor, Intel VT-d in the I/O chip set, and Intel VT-c in the
network adapter chip set. The Intel Xeon 5500 Series incorporates Intel VT as do many newer Intel
processors. Refer to http://ark.intel.com/VTList.aspx.
2. Refer to the Release Notes for the Mitel product version you are using for an up-to-date list of
supported processors.
3. Some processors allow selection of power management modes in BIOS settings (Low Power vs.
High Performance, for example). It is recommended that hosts used to support real-time sensitive
applications such as MiVoice Business and MiVoice Border Gateway, have their BIOS setting
configured to maximize performance. This reduces the risk of voice quality issues and missed
interrupts, for example.
9
Mitel Virtual Appliance Deployment
MiVoice Business Virtual known issues Mitel Virtual Appliance Quick Reference Guide
MiVoice Business Express MiVoice Business Express Deployment Guide
MiCollab MiCollab Installation and Maintenance Guide
MiCollab Client Administrator’s Guide
MiCollab Audio, Web, and Video Configuration and Maintenance
Manual
MiCollab Engineering Guidelines
MiCollab Client Engineering Guidelines
MiVoice Border Gateway Virtual installation MiVoice Border Gateway Installation and Maintenance Guide
and maintenance
MiContact Center Office MiContact Center Office Technician’s Guide
MiContact Center Business MiContact Center Business Deployment Guide
MX-ONE MiVoice MX-ONE Virtualization, Description
MiCloud Management Portal (Oria) MiCloud Management Portal Installation and Maintenance Guide
deployment MiCloud Management Portal Engineering Guidelines
The main VMware documentation Web page http://www.vmware.com/support/pubs/
The main VMware vSphere 5.5 Web page, http://www.vmware.com/support/pubs/vsphere-esxi-vcenter-server-p
with links to documentation, videos, and ubs.html
release notes
The VMware Hardware Compatibility guide http://www.vmware.com/resources/compatibility/search.php
vSphere resource management VMware vSphere Resource Management Guide
vSphere performance best practices Performance Best Practices for VMware vSphere
Microsoft Hyper-V documentation Microsoft Technet Library
10
Chapter 2
DEPLOYING ON A VMWARE
INFRASTRUCTURE
Mitel Virtual Appliance Deployment
12
Deploying Mitel virtual applications on a VMware virtual infrastructure
13
Mitel Virtual Appliance Deployment
With virtualized voice, rather than having to handle voice communications with a separate
budget and separate sets of hardware, processes, tools, and often staff, voice can be treated
like any other business application in the data center. Rather than managing boxes, IT managers
can manage the overall services that IT provides to the business, and in the process, reap the
benefits and cost savings of a simplified test, development, and production cycle, streamlined
administration, and a single disaster recovery and business continuity plan that applies to the
whole data center.
Mitel virtual appliances are bundled with Mitel Standard Linux (MSL) and packaged in Open
Virtualization Format (OVF) for deployment in a VMware environment. Mitel virtual applications
are delivered as OVA (.ova) files, ready to import into a VMware ESXi Hypervisor.
Given the potential complexity of your solution, and the time-sensitivity of the functions, it is
important that the systems be configured to:
• Ensure that the Mitel products work optimally with each other and with your existing
infrastructure.
• Ensure that the overall solution is optimized for use with the virtual (VMware) infrastructure.
• Ensure that the VMware servers have the capacity to deliver the standard of service you
need.
• Increase the availability of your services and applications.
14
Architecture and topology of Mitel virtual solutions
Figure 2: MiVoice Business Virtual application layer resiliency with vSphere High Availability
This scenario consists of a Mitel Voice Over IP (VOIP) and Unified Communications (UC) virtual
solution in a single data center, similar to the previous scenario. This scenario adds MiVoice
Business Virtual controllers for application layer resiliency that ensures immediate voice
recovery, and VMware High Availability (HA) to provide rapid recovery and return to the primary
MiVoice Business Virtual, in the event the host or virtual machine fails.
MiVoice Border Gateway Virtual application layer resiliency (clustering) ensures immediate
availability of voice connectivity to the service provider. A failed MiVoice Border Gateway Virtual
recovers using HA, restoring full service quickly.
The resumption of voice connectivity occurs in preference to data network connectivity: SIP
and PRI connectivity. In this example, PRI is the front-end to the virtual appliances with a
physical Mitel 3300 ICP acting as the media gateway and connection to each of the data centers.
15
Mitel Virtual Appliance Deployment
Deployment considerations
The following considerations will affect the design of your deployment:
• “Deployment types” on page 16
• “Preparing the VMware infrastructure” on page 17
• “VMware support by Mitel product release” on page 20
• “Host server sizing and resource management considerations” on page 21
Deployment types
There are three primary deployment types used for the installation of Mitel virtual applications:
• Small and Medium-size Business (SMB)
SMB deployments of virtual applications are usually used in smaller businesses, for server
consolidation. Management is generally not the primary issue, but running Mitel Unified
Communication and Collaboration (UCC) applications on VMware virtual servers offers
cost savings because fewer servers are required.
An SMB deployment is likely to have a small number of host servers, and possibly just a
single stand-alone host. vCenter Server-related functions such as vMotion and HA
virtualized environment availability features may not be present. Mitel applications generally
coexist with non-Mitel applications in SMB environments.
• Enterprise
Enterprise-level deployments are used either to fit the Mitel system into the existing virtual
IT infrastructure, or to create a new virtual IT infrastructure specifically dedicated to
Enterprise UCC functions.
Rather than setting up a Mitel network on your premises with Mitel hardware and software,
you can have your UCC and other applications hosted in a Service Provider data center,
in a Public Cloud, for a monthly hosting charge or similar billing arrangement, using one of
the deployment models discussed below.
This environment is likely to be a large-scale cloud cluster in which you lease capacity to
run your virtual applications. Sophisticated management, including vCenter Server and
related operations, are usually present, but not directly accessible by the customer.
16
Deployment considerations
The Mitel virtual UCC solutions are well-suited to the following deployment models:
• Software as a Service (SaaS) - Service Providers or Enterprise IT host the Mitel UCC
solution as a set of software applications within a VMware vSphere shared server infra-
structure, and offer it to customers as a service. There may also be other non-Mitel
applications offered as part of the overall service. Customers, in this context, may be indi-
vidual end-customers, Mitel resellers managing the UCC and other applications on behalf
of the end-customer, or Enterprise IT (Private Cloud) providing service to organizations
within the enterprise.
• Unified Communications as a Service (UCaaS) - Similar to SaaS, Service Providers or the
Enterprise IT host the Mitel UCC solution as a set of applications within a VMware vSphere
shared server infrastructure, and offer UCC service to customers. Customers may be indi-
vidual end-customers, Mitel resellers managing the UCC applications on behalf of the
end-customers, or Enterprise IT (Private Cloud) providing service to organizations within
the enterprise.
• Infrastructure as a Service (IaaS) - Infrastructure providers lease out computing and network
resources (for example: vCPU, GHz, RAM, HDD, and network access) required to host the
Mitel UCC solution on their VMware vSphere shared infrastructure. The end customer, or
a Mitel reseller acting on their behalf, must handle all aspects of installation, administration,
management and maintenance of the UCC applications.
• Customer Premise Equipment - Mitel certified dealers or customers install and configure
Mitel UCC virtual applications in the VMware environment on the customer's premise. The
virtual infrastructure is directly managed by the customer. This is primarily suited to Enter-
prise and SMB type deployments.
When deciding on the host server computing environment, and planning how to set up one or
more clusters of virtual machines, consider the following:
• Plan for adequate overall computing capacity across the cluster of host servers to accom-
modate immediate and foreseeable future needs for overall CPU and memory. This should
include availability plans, if one or more hosts fail. See “Host server sizing and resource
management considerations” on page 21.
• Plan to deploy on multiple hosts of similar capacity and compatible processor type within
each cluster so that you can enable Distributed Resource Scheduler (DRS), vMotion, High
Availability (HA), Distributed Power Management (DPM), and other VMware availability and
optimization features.
17
Mitel Virtual Appliance Deployment
Note: Mitel expects vMotion and Storage vMotion to function properly with Mitel virtual
appliances in accordance with supported configuration from VMware. Similarly, SAN,
VMware Virtual SAN (vSAN), and other storage connectivity is expected to work in
VMware-supported configurations, subject to the I/O requirements listed in Table 8.
Refer to Mitel application-specific Engineering Guidelines documentation (Table 9) for
vMotion host and storage performance requirements.
If you want to use vMotion in a configuration that may be outside these specifications,
contact Mitel Professional Services for assistance.
• Configure hosts into clusters, with VMware Distributed Resource Scheduler (DRS) enabled.
DRS continuously monitors utilization across a resource pool, and allocates available re-
sources among virtual machines according to business needs. See “Resiliency, high
availability, and disaster recovery” on page 39 for more information about DRS.
• Ensure that all hosts that will be used to run real-time sensitive applications, such as MiVoice
Business Virtual, NuPoint UM Virtual, and MBG Virtual, are configured for maximum per-
formance. This includes processor and BIOS settings requirements as listed in “Minimum
hardware requirements” on page 9.
Networking
Follow VMware best practices for configuration of networking capacity and fault tolerance. Mitel
recommends the following minimum guidelines:
• Supply plenty of bandwidth: 1 Gbit/s across each of at least four physical ports on each
host server is considered the minimum for most deployments.
• Separate the Virtual Local Area Network (VLAN) and IP interfaces for storage traffic (as-
suming Storage Area Network (SAN), VMware Virtual SAN (vSAN), Network File System
(NFS), or similar IP storage is used) on a separate VLAN or subnet. This provides adequate
throughput for storage operations, while isolating traffic from other flows. Preferably, as per
VMware best practices, you should configure storage networking on two or more physical
ports on the host servers, directed at different network switches to provide resiliency. You
should also consider configuring these ports for NIC bonding (using Link Aggregation Con-
trol Protocol (LACP) or similar) to improve overall throughput and/or resiliency.
• If VMware vMotion features such as Distributed Resource Scheduler (DRS) and High Avail-
ability (HA) will be used, separate VLAN, IP interface, and multiple physical ports should
be used for vMotion traffic.
• Voice traffic to and from the Mitel virtual appliances should be separate from data traffic.
Implement a dedicated voice VLAN, and support it through multiple physical ports. The
same recommendation applies for physical implementations.
Storage
In planning connections to, and stability of, the network storage, consider the following
guidelines:
• For networked storage, ensure a robust Storage Network to support storage network traffic
with minimal network delays.
18
Deployment considerations
• Ensure adequate throughput and IOPS to support all hosted applications in the storage
device. Refer to the following documentation for detailed requirements.
• “VMware vSphere storage setup guidelines for Mitel virtual appliances” on page 121
• Product-specific documentation, for detailed requirements
• VMware documentation for storage set-up. VMware Publication: vSphere Storage (for
version number)
• Deploy with support for replication, including remote replication if you plan to implement
Site Recovery Manager (SRM), or a similar disaster recovery scheme.
• Deploy storage with multi-pathing, with a minimum of two physical ports on each host server
and on each of the storage units, directed at different network switches. This provides
resiliency.
• RAID protection of all storage is strongly recommended.
Mitel virtual appliances support for various storage architectures matches that of VMware
vSphere, unless specifically stated otherwise. This includes iSCSI, NFS, Fibre Channel, VSAN,
VSA (VMware and other vendors), and host-local storage. Host-local storage has a major
limitation in that the virtual application cannot be moved to a different host server using vMotion
for load balancing or maintenance, or protected by HA. For details, refer to the Mitel
documentation for each virtual appliance, available at Mitel OnLine.
Note: Regardless of the storage technology used, ensure that the storage used by Mitel
applications meets or exceeds the specific requirements of the applications involved, in
terms of IOPS. See Table 8, “VMware required resource reservation and allocation,” on
page 21, and application-specific Engineering Guidelines.
In addition to Network Storage, Mitel supports on-board server storage that can be enabled
with VMware vSphere Storage Appliance (VSA), and is ideal for smaller businesses that have
a small data center and do not want to invest in expensive off-board network storage solutions.
VSA allows for support of HA, DRS, and DPM with on-board server storage for up to three
servers. (The three-host limit is imposed by the VMware implementation of VSA, as of vSphere
5.1.) vMCD Release 6.0 and MiVoice Business Virtual Release 7.0 support VSA.
Note: VMware VSA may be lower performance than hardware-based high performance
shared storage. If VMware VSA is to be used, ensure that VSA can meet the performance
requirements of the applications using this as storage.
The VMware vSAN storage solution is fully integrated with vSphere. It automatically aggregates
server disks in a cluster to create shared storage that can be rapidly provisioned from VMware
vCenter during VM creation. It is an object-based storage system and a platform for VM Storage
Policies designed to simplify virtual machine storage placement decisions for vSphere
administrators. It is fully integrated with core vSphere features such as VMware vSphere High
Availability (vSphere HA), VMware vSphere Distributed Resource Scheduler™ (vSphere DRS)
and VMware vSphere vMotion®.
Its goal is to provide both high availability and scale-out storage functionality. It can also be
considered in the context of quality of service (QoS) because VM Storage Policies can be
19
Mitel Virtual Appliance Deployment
created that define the level of performance and availability required on a per-virtual machine
basis. See “VMware vSphere storage setup guidelines for Mitel virtual appliances” on page 121.
20
Deployment considerations
Notes:
1. MCDs and vMCDs using MCD clustering and IP Trunks are not compatible in SRM for versions
earlier than MiVoice Business Release 7.0. Only SIP trunks and standalone vMCDs inter-operate
with SRM. A deployment configuration has been tested with MiVoice Business Virtual that uses
resiliency along with SRM for MiVoice Business Virtual Release 7.x.
2. After SRM recovery—that is, after the system becomes active on the recovery site—you must
configure and test SRM at the recovery site to ensure the ability to move back to the original site
or other alternate site. Then you can use SRM to move the system back to the original site.
3. Horizon View is no longer supported for MiCollab Client 8.0. Earlier versions of MiCollab Client
supported Horizon View.
Table 8 shows the resource allocation and resource reservation requirements for each of the
Mitel virtual appliances.
Refer to the Engineering Guidelines for each product for more detailed information about
deployment of the virtual version.
CURRENT
RELEASES
21
Mitel Virtual Appliance Deployment
22
Deployment considerations
MiCollab 8.1 5000 users 8 120 GB 8 GHz 16.0 GB 1 MB/s 200 IOPS
Client (in
MiCollab)
NuPoint 9.1 120 ports 8 120 GB 5 GHz 16.0 GB 1.6 MB/s 80 IOPS
UM (in
MiCollab)
MiCollab 8.1 5000 users, 250 8 120 GB 12.5 16.0 GB 2.8 MB/s 200 IOPS
Multi Tenant tenants GHz
MiCollab 8.1 250 users 2 40 GB 1 GHz 5.0 GB
Client
1500 4 40 GB 1.5 GHz 7.0 GB
Standalone
2500 6 120 GB 3 GHz 8.0 GB
5000 users 6 120 GB 3.5 GHz 10.0 GB 1.5 MB/s 200 IOPS
MiVoice 10.1 125 devices 1 20 GB 1 GHz 1 GB 1.7 MB/s 4 IOPS 60
Border @6ccs/device13
Gateway
250 devices 2 20 GB 2 GHz 2 GB 1.7 MB/s 4 IOPS 125
Virtual3
@6ccs/device13
1250 devices 3 40 GB 5 GHz 2 GB 5 MB/s 27 IOPS 250
@6ccs/device13
2500 devices 4 40 GB 6 GHz 3 GB 5 MB/s 27 IOPS 500
@6ccs/device13
5000 devices 4 40 GB 6 GHz 4 GB 5 MB/s 27 IOPS 500
@3ccs/device13
5000 devices 6 40 GB 9 GHz 4 GB 5 MB/s 27 IOPS 830
@6ccs/device13
8000 devices 6 40 GB 9 GHz 4 GB 5 MB/s 27 IOPS 830
@3ccs/device13
SIP trunk only 2 20 GB 2 GHz 2 GB 1.7 MB/s 4 IOPS 200
200 channels
SIP trunk only 4 40 GB 6 GHz 3 GB 5 MB/s 27 IOPS 500
500 channels
MiContact 6.2 Entry level8 2 40 GB 2 GHz9 2.0 GB9 2.7 MB/s 17 IOPS
Center Office SP1
Mid-range8 2 100 GB 2 GHz9 4.0 GB9 3 MB/s 46 IOPS
Virtual (ESXi
8 9 9
6.0) Enterprise 2 200 GB 2 GHz 4.0 GB 3 MB/s 46 IOPS
23
Mitel Virtual Appliance Deployment
24
Deployment considerations
PREVIOUS
RELEASES
MiVoice 8.0 250 devices4 2 20 GB 2 GHz 1.5 GB 6 Mb/s 28 IOPS
Business SP3,
1500 devices4 3 20 GB 3 GHz 2.0 GB 9 Mb/s 28 IOPS
Virtual3 8.0
SP2, 2500 devices4 4 20 GB 5 GHz 2.0 GB 11 Mb/s 28 IOPS
8.0 5000 devices4 6 20 GB 8 GHz 2.0 GB 16 Mb/s 56 IOPS
SP1,
Embedded include 3.2 Mb/s 24 IOPS
8.0, Voice Mail2 d
7.3+,
7.2,
7.1, 7.0
MiVoice 6.3 500 users 2 60 GB 1.0 GHz 4.0 GB
MX-ONE12 SP1
1000 users 2 60 GB 1.5 GHz 4.0 GB
2000 users 2 60 GB 2.0 GHz 4.0 GB
3000 users 2 60 GB 2.5 GHz 4.0 GB
4000 users 3 60 GB 3.0 GHz 4.0 GB
5000 users 4 60 GB 4.0 GHz 4.0 GB
25
Mitel Virtual Appliance Deployment
26
Deployment considerations
27
Mitel Virtual Appliance Deployment
28
Deployment considerations
29
Mitel Virtual Appliance Deployment
MiContact 6.1. 6.0 Entry level8 2 40 GB 2 GHz9 2.0 GB9 2.7 MB/s 17 IOPS
Center Office
Mid-range8 2 100 GB 2 GHz9 4.0 GB9 3 MB/s 46 IOPS
Virtual
Enterprise8 2 200 GB 2 GHz9 4.0 GB9 3 MB/s 46 IOPS
Enterprise 7.0 4 45-165 1.0-2.0
Manager6 GB11 GB
Notes:
1. The VMware Hypervisor consumes resources from the platform it is running on: 1 vCPU, 2.0 GBytes RAM, 2
GHz CPU.
2. Voice Mail is based on twelve active channels. Resource usage is additional to that needed by the MiVoice
Business Virtual. (For lower numbers of users in vMCD, voice mail was based on eight active channels.)
3. The CPU, Memory, Storage Capacity, and CPU/memory reservations are set to their defaults at installation
time. The network and storage capacity are estimates of what is required to achieve the best voice quality.
Changing the Virtual Machine settings from vSphere may result in performance degradation or voice quality
issues. This applies to all of the appliances that handle voice traffic, including MiVoice Business Virtual, NuPoint
Virtual, MBG Virtual, MiCollab Virtual, and MiVoice Business Express.
4. The CPU reservation can be altered to support different system configurations. Note that the MiVoice Business
5000 user configuration is not available in the OVA, and must be configured manually. See “Increasing
resources for larger dimensions” on page 31.
5. Each virtual machine has an additional memory cost—overhead needed by VMware Hypervisor to manage
it. The overhead amount is based on the memory allocated and the number of vCPUs. In ESXi 5.0, the memory
overhead for a VM of the size of Mitel virtual appliances ranges from 100 MB to as high as 500 MB each. Refer
to vSphere Resource Management documentation.
6. Though not packaged as an OVA, these virtual appliances consume the resources specified here. The Memory
number is not a reservation, but a recommendation of what should be made available.
7. Hard disk space required depends on Operating System: Windows Server 2008 R2 (32/64 bit) requires 32
GB; Windows Server 2003 (32 bit) requires 16 GB; Windows 7 Enterprise, Professional, or Ultimate (32/64
bit) requires 25 GB; Windows XP Pro SP3 (32 bit) requires 10 GB.
8. MiContact Center Office ships with entry-level settings.
9. MiContact Center Office uses the Normal setting for CPU Shares and Memory Shares.
10. Although no resource reservations are required for Oria 4.0 SP2, it is recommended that the reservations
noted above be applied manually for Oria 5.0+.
11. MiVoice Enterprise Manager recommends 40 GB to manage two nodes, 120 GB to manage 50 nodes, and
160 GB to manage 250-1000 nodes.
12. For MiVoice MX-ONE, MiVoice 5000, and MiVoice Office 400, see the product documentation for the latest
information.
13. CPU requirements have increased due to the addition of Spectre & Meltdown protections. The CPU reservation
can be altered to support different system configurations. The default configuration for MBG Virtual Small
Business supports 250 devices and 125 simultaneous calls. The default configuration for MBG Virtual
Enterprise supports up to 5000 devices with a maximum of 500 simultaneous calls. Other configurations must
be configured manually. See 'Increasing resources for larger dimensions' below.
14. The Network requirements stated here are for both LAN and WAN combined.
For detailed information about deploying MiVoice Business Express, refer to the MiVoice Business Express
Deployment Guide. This guide also includes Engineering Guidelines information.
30
Deployment considerations
15. In all the Enterprise configurations for MiCollab, the MiVoice Border Gateway must be used only for
management. All media traffic is handled by a separate standalone MBG. The single-app MiCollab Client for
5000 users must be in co-located mode and all NuPoint UM traffic is handled by a separate standalone NuPoint
UM (no longer available for purchase).
16. Mitel publishes the application IOPS based on internal testing. To calculate the actual disk speed required to
support these applications, the RAID type, disks in the array, and read and write workload must be taken into
account.
17. Mitel has increased the OVA sizes from 150 to 250 for the SMB profile. If the customer has previously deployed
a 150 user system, the OVA can be reduced to meet the 150 user specifications, as previously published.
18. The IOPS requirement stated for MiVoice Business Express does not include the additional requirement for
MiCollab AWV video session recording.
19. Equal to #vCPUs at 1GHz each, provided 1 vCPU is one thread of a Hyper threaded CPU at 2GHz, or more.
Each Mitel virtual appliance is built with the resource reservation required for the performance
Service Level Agreements already set. Do not change the resource reservation. If you choose
to reduce the resource reservation, you do so at your own risk, and you must monitor
performance carefully to ensure that you always retain some headroom.
31
Mitel Virtual Appliance Deployment
Resource dimensioning
In keeping with VMware best practices and Mitel requirements, you must have a minimum of
2 GB of memory allocated for the VMware Hypervisor (ESXi).
Refer to the VMware vSphere Resource Management Guide and the Performance Best
Practices for VMware vSphere on the VMware Web site.
CAUTION: You must maintain CPU Headroom of 30% over the peak load to
guarantee service level agreement (SLA) performance.
Storage provisioning
During deployment of the OVA file, you must choose the provisioning format. The choices are:
• Thin provisioned format - The VM storage is initially allocated based on the storage required
at the time of deployment, and can expand as storage needs increase.
This is more efficient in terms of total storage requirement, but can also be considerably
slower as the storage grows and becomes more fragmented.
• Thick provisioned format - Causes immediate allocation of all storage required. This is less
efficient in terms of total storage requirement (does not allow over-commit), but it guarantees
optimum performance. Mitel virtual appliances that are sensitive to voice quality, including
MiVoice Business Virtual, MiVoice Border Gateway, NuPoint Virtual, and MiVoice Business
Express require thick provisioning to ensure good voice quality from the first call. For more
information, refer to the product-specific documentation.
32
Deployment considerations
Thick provisioned format includes two additional options: Lazy-zeroed and Eager-zeroed.
Mitel recommends using the Lazy-zeroed option. For more information about storage
provisioning and these settings, see “Disk provisioning” on page 63.
Note: If you are using a storage technology with de-duplication capability, then the
difference between thin and thick-provisioned formats is reduced, and you may find that
Thin provisioning is sufficient.
The Mitel recommendation is to use Thick provisioning for all Mitel virtual appliances.
Contact your storage provider for information about the over-commit capabilities of the storage
option you are using.
One of the major advantages of virtualization is that you can co-locate multiple VMs on one
host server.
Because most of Mitel's virtual applications provide real time voice, the customer will experience
voice quality issues if insufficient CPU resources are allocated. Mitel builds in VMware resource
reservation settings that will prevent resource limitation issues for typical customer installations.
VMware uses resource reservation to guarantee that the specified amount of the resource
(CPU cycles and memory) is always available to the VM, regardless of the needs of other VMs
sharing the resource.
The host CPU utilization must be maintained below 70% to prevent voice quality issues, as
discussed in “Resource dimensioning” on page 31.
Should performance or quality issues arise after reducing the resource reservation below the
settings provided by Mitel, the first action Mitel Product Support will suggest is to restore the
reservations to Mitel-mandated values.
Mitel virtual appliances may be installed into the Virtual Infrastructure, where they will share
resources with third-party virtual appliances and applications. A primary consideration in this
scenario is ensuring that resource allocations for Mitel voice operations (MiVoice Business
Virtual and other real-time voice applications) are adequate for good voice quality. Giving them
too little resource reservation and priority may result in voice quality issues. These issues may
be very transitory, based on what all the virtual appliances are doing at any given point in time.
This can make troubleshooting very difficult.
There are two ways to ensure good performance of Mitel real-time applications; by setting a
guaranteed resource reservation for the time-sensitive applications, or by giving each
application a share-based percentage of the total CPU.
Setting a guaranteed resource reservation is the best way to ensure that enough CPU is
allocated to the voice applications; a percentage-based allocation does not guarantee the
required performance level in cases where the CPU is overloaded.
When co-locating Mitel and third-party applications, you must ensure that restart priorities are
set correctly. MiVoice Business Virtual should be set to the highest priority. This sets the MiVoice
33
Mitel Virtual Appliance Deployment
Business Virtual VM ahead of other non-critical appliances in the event that a host fails. For
example, you would set instances of MiVoice Business Virtual at the top priority, with MiCollab
Virtual at a slightly lower priority, followed by your other business applications. Refer to
“Resiliency, high availability, and disaster recovery” on page 39 for more information.
Should performance or quality issues arise after you have reduced the resource reservation
below the settings provided by Mitel, you will be asked to restore the resource reservations
before support is offered for SLA issues. Maximum loading, including all Mitel and third-party
virtual appliances, is 70%.
In some cases where the MiVB instances are running less traffic and possibly less users than
the standard deployments, it may be possible to tune and reduce the GHz reservations. This
can improve instance density per server. For further details on settings, refer the MiVoice
Business Engineering Guidelines (MiVB Release 8.0 onwards) and Engineering Guidelines for
MiVoice Business for ISS and MiVoice Business Virtual for earlier releases.
If the installer is unfamiliar with determining and applying the resource settings, it is
recommended to contact Mitel Professional Services.
Deploying multiple Mitel applications on one virtual server has advantages in that the various
Mitel applications have similar requirements, and they are intended to work together. This
generally simplifies resource considerations, and can result in lower maintenance costs.
When co-locating multiple instances of different Mitel virtual appliances, consider setting up a
resource pool. In a resource pool, you can set priorities for resource allocation within the pool.
For more information about resource pools, see “Using Resource Pools” on page 32.
Minimum storage requirements are product specific, and different for every Mitel virtual
appliance. For product-specific storage requirements, refer to the applicable Mitel Engineering
Guide. For the I/O characterization specifics, see Table 8 on page 21.
This section describes some additional items to consider when you deploy Mitel virtual
appliances and applications.
Anti-affinity rules
Some applications, especially when deployed in resiliency configurations, require that VMs for
the applications do not reside on the same physical host server. Rules that ensure that certain
VMs remain on different servers are called anti-affinity rules.
For example, a MiVoice Business Virtual primary controller VM should never be allowed to run
on the same hardware as the corresponding MiVoice Business Virtual secondary controller
VM. VMware vCenter anti-affinity rules should be set to ensure that the primary and secondary
are never moved to the same host server, including during automated actions such as vMotion
or HA. See “Resiliency, high availability, and disaster recovery” on page 39 and the VMware
documentation for more information.
34
Deployment considerations
Intel's Hyper-Threading Technology makes one core appear as two logical processors, but the
two logical processors share most of the core's resources. If one logical processor is running
at high utilization, the other logical processor will have very little capacity that can be used
before the performance of both logical processors suffers.
AMD Opteron processors use Rapid Virtualization Indexing, which helps accelerate the
performance of many virtualized applications by enabling hardware-based virtual machine
memory management.
ESXi assigns adjacent vCPU numbers to the logical processors on the same core. VMware
attempts to keep the vCPUs of different VMs on separate cores. Use caution if you decide to
use VMware advanced features to override VMware default behavior—assigning vCPUs to
VMs. If two high-priority VMs have vCPUs on the same core, performance problems are likely.
CPU utilization shown in VMware management tools (Performance tab in vSphere Client, for
example) can be misleading. If half the vCPUs are running at near 100% (on different cores)
and the other half of the vCPUs are running near 0%, the system is actually fully loaded, although
the average CPU may show 50% (the average of all the vCPUs).
Similarly, the GHz value for the system is really closer to the GHz sum of all the cores, not the
sum of all the vCPUs.
With servers hosting Mitel virtual appliances that are less demanding—appliances that do not
have to handle voice traffic—the assigned vCPU capacity may exceed the number of logical
processors (twice the number of physical cores) by up to 50%. For example, a server with 8
physical cores provides 16 logical cores when hyperthreading is enabled. In this case, with
non-voice traffic, you can provision 50% more vCPUs, for a total of 24 vCPUs.
Also refer the MiVoice Business Engineering Guidelines (MiVB Release 8.0) for MiVoice
Business on ISS and MiVoice Business Virtual.
Using 3D acceleration
By default, 3D acceleration is not enabled in ESXi 5.1 and up. You can enable 3D support on
virtual machines that have the Windows 8+ guest operating system.
1. For the virtual machine you are enabling 3D acceleration on:
• Verify that the virtual machine compatibility is ESXi 5.5+.
• Verify that the virtual machine is powered off.
2. Right-click the virtual machine in the inventory.
3. Click Edit Settings.
4. In the left column of the Virtual Hardware tab, expand Video Card.
5. Select Enable 3D Support.
35
Mitel Virtual Appliance Deployment
Product-specific guidelines
For additional product-specific restrictions and requirements, see the Mitel Engineering
Guidelines for each product.
36
VMware features
VMware features
VMware offers many features and capabilities, many of which are compatible or applicable for
Mitel virtual applications. Table 10 lists the basic VMware features, and specifies which are
expected to work with Mitel virtual appliances. See also Table 12, “Mitel compatibility with
VMware resiliency features,” on page 40.
Note: Mitel maintains support for current VMware vSphere/vCenter releases as outlined
in this document. VMware also provides a number of infrastructure services that require
the underlying VMware (vSphere / vCenter) such as SRM, vCNS, and vCloud Director
that do not have any product integrations. As such, these VMware capabilities are
supported by the customer and do not require direct product support.
37
Mitel Virtual Appliance Deployment
38
Resiliency, high availability, and disaster recovery
For those Mitel products that do not offer application layer availably solutions, the VMware high
availability (HA) features are highly recommended.
Whether or not you use VMware features for high availability, critical host servers must be
arranged to ensure that one physical or electrical disruption cannot disable all servers in the
cluster.
MiVoice Border Gateway (MBG) also includes resiliency capabilities; supported MiNet sets can
support a resiliency list of up to four IP addresses. If a set loses its connection and cannot
re-establish it, it will try the next IP address on its list until it has exhausted all IP addresses on
its resiliency list. For MiNet sets supporting persistent resiliency lists, this resiliency list is
retained through power-cycling. The resiliency list can be manually configured with arbitrary IP
addresses. Ideally, they will be configured to correspond to multiple MBG servers.
39
Mitel Virtual Appliance Deployment
Note: Mitel maintains support for current VMware vSphere/vCenter releases as outlined
in this document. VMware also provides a number of infrastructure services that require
the underlying VMware (vSphere / vCenter) such as SRM, vCNS, and vCloud Director
that do not have any product integrations. As such, these VMware capabilities are
supported by the customer and do not require direct product support.
40
Resiliency, high availability, and disaster recovery
Notes:
1. For MiContact Center Office, these features must comply with the Windows licensing restrictions.
2. Mitel expects vMotion and Storage vMotion to function properly with Mitel virtual appliances in accordance
with supported configuration from VMware. Similarly, SAN or other storage connectivity is expected to work
in VMware-supported configurations, subject to the I/O requirements listed in Table 8.
If you want to use vMotion in a configuration that may be outside these specifications, contact Mitel Professional
Services for assistance.
Note: VMware features such as vMotion, Storage vMotion, HA, SRM, cloning, templates,
and so on, do not require re-licensing of the application, because Mitel virtual application
licensing Application Record IDs (ARIDs) are based on globally unique IDs (GUIDs)
used by the ARID, and are not dependent on the specific hardware the application is
running on.
HA can be a very effective supplement to Mitel built-in resiliency features. HA can protect
against individual VMware host failures by restarting VMs that were running on the failed host,
on a different host. While some down-time is involved for the VMs during the HA operation (as
they are started on the new host), the inherent Mitel resiliency architecture provides zero
downtime for the overall solution. Use of VMware HA does not take the place of Mitel resiliency
operations, but supplements them by greatly reducing downtime of the nodes involved.
41
Mitel Virtual Appliance Deployment
In the example in Figure 3, if MiVB Virtual A is a primary node deployed on host X, with HA
protection to restart on some other host, and MiVB Virtual B is a secondary node deployed on
host Y, also with HA protection, then the following failure scenario can be configured:
1. Host X fails. MiVB Virtual A is now out of service.
2. All IP Phones homed to MiVB Virtual A immediately re-home (fail over) to MiVB Virtual B
on Host Y (their secondary) with near-zero loss of service (seconds to fail over, and no
loss of active calls).
3. VMware HA then ensures that MiVB Virtual A is restarted on Host Z (anti-affinity rules apply
so that MiVB Virtual A will not be restarted on the same host server as MiVB Virtual B),
and during this time service is maintained by the secondary controller: MiVB Virtual B on
Host Y.
4. When MiVB Virtual A is recovered a few minutes later, the IP Phones again re-home to
MiVB Virtual A, now running on Host Z. In this case, HA protection of the primary node
effectively reduces down time of the primary node to minutes.
In the physical world, it could take hours or even days to replace physical hardware running
the primary node, and the devices would continue to run on the secondary for the duration. Not
only is this non-optimal operation, it also adds the risk of additional down time, should node B
also fail during the Node A repair time.
42
Resiliency, high availability, and disaster recovery
In the reverse scenario where failure happens on host Y, the secondary MiVB Virtual is restarted
elsewhere, but with no loss of service since the IP Phones remain homed to their primary MiVB
Virtual A.
HA protection can also be applied to Mitel virtual appliances that do not support inherent
resiliency. While these applications are generally not considered mission critical, HA protection
will allow the application to be monitored and automatically restarted elsewhere in the case of
host failure. This will have the effect of reducing down time from hours or even days in the
physical world to minutes in the virtual world.
When planning and configuring VMware HA protection, it is very important to consider the
following:
• Anti-affinity rules must be put in place, such that primary and secondary nodes within a
cluster cannot be run on the same physical host. This also applies in the event of HA
operation restarting a node on a different host; the target host must also have the anti-affinity
rules applied. See “Anti-affinity rules” on page 34.
• Carefully consider HA restart priority settings for the protected virtual appliances. For ex-
ample, it may be desirable to give primary nodes higher restart priority than secondary
ones, or some application categories higher priority than others. For example, mission
critical MiVoice Business Virtual controllers and MBG Virtual instances would be set for a
higher restart priority than non-mission-critical applications.
• Plan the HA admission control policies to ensure adequate resources are reserved, in
advance, to accommodate all high priority HA-protected applications, especially mis-
sion-critical Mitel virtual appliances. The admission control policies also affect the restart
priority discussion above. See “Admission control for High Availability (HA) clusters” on
page 43 for recommendations for setting HA admission control policies.
• In circumstances where insufficient host resources exist after a failure, HA may override
anti-affinity rules set at the Distributed Resource Scheduler (DRS) level. This may result,
for example, in MiVoice Business Virtual primary and secondary nodes running on the same
host until the failed host can be repaired. For a single pair of MiVoice Business Virtual
controllers, a minimum of three host servers would be required to avoid this, each with
sufficient guaranteed spare capacity to accept and run the failed-over MiVoice Business
node.
When configuring your high-availability VM cluster, you choose one of the following three
admission polices:
• Host Failures Cluster Tolerates
• Percentage of Cluster Resources Reserved
• Specify a Failover Host
Alternatively, you can disable admission control if that makes sense for your implementation.
If availability is more important than Service Level Agreement (SLA) performance upon failure,
then, in VMware vSphere, setting Disable: Power on VMs that violate availability
constraints may be worth considering. Keep in mind that there is a risk of voice quality
degradation.
43
Mitel Virtual Appliance Deployment
If you disable admission control, you must manually manage the number of VMs allowed to
run on the hosts and/or reduce the VM reservations to ensure that the VMs will be able to restart
on another host in the event of a host failure; this is the source of the voice quality risk. Enabling
admission control is generally recommended.
The three available VMware admission control policies are described in the following sections.
The Host Failures Cluster Tolerates admission control policy uses slot size calculations to
determine whether enough capacity exists in the High Availability cluster to power on a virtual
machine. This mode ensures HA for all the VMs in the cluster, and it is easy to configure, since
VMware does all the resource calculations.
On the other hand, if a VM with a large resource reservation is introduced into the HA cluster,
it can limit the number of VMs that can be powered on at one time. The large resource
reservations set in the Mitel OVA files can change the slot size calculations for all the VMs in
the cluster. Since most applications do not have an explicitly-set resource reservation, they
work within the default 64 KB slot size, unless the slot size is changed by the introduction of
an extra-large VM.
To avoid the potential problems associated with the cluster-wide slot size being changed to a
large value by the introduction of Mitel appliances, you should set the maximum slot size
manually, before adding the Mitel appliances. If you set a maximum, virtual machines larger
than the set slot size are assigned multiple slots.
To reduce resource overhead due to the inefficiency in the slot size definition, you might consider
creating separate clusters for different MiVoice Business Virtual and MBG Virtual deployments.
For example, you could have a cluster with MiVoice Business instances dedicated and
optimized for 150 users and another cluster optimized for 2500 users. The second cluster could
use larger slot sizes.
Using this percentage-based admission control policy is potentially more cost effective. To allow
a trade-off between cost and availability SLA, for example, you can set the percentage to ensure
availability only for critical VMs (defined using VM Restart Priority in Virtual Machine Options
under the VMware HA configuration panel).
The flexibility in this option comes with some complexity, however, since you have to determine
the appropriate percentage for the required SLA. In addition, the percentage value may require
on-going adjustment depending on:
• The SLA level (how many VMs must be protected)
• Changes in available resources in the cluster
• How many simultaneous host failures must be accommodated
The complexity of determining, monitoring, and re-adjusting the percentage setting affects the
ongoing cost of maintaining the system, and the risk of undershooting customer SLA
requirements during failures (lowered voice quality or inability to support full MiVoice Business
resiliency, for example).
44
Resiliency, high availability, and disaster recovery
Another risk in using the Percentage of Cluster Resources Reserved admission control policy
is that, in a cluster running close to capacity, there might be enough aggregate resources for
a virtual machine to be failed over, but the resources might be fragmented across different
hosts. When that happens, they are not usable. A virtual machine can run on only one ESXi
host at a time.
Using the Specify a Failover Host admission control policy, you specify one host to serve as
a backup. It is kept idle so that it is always available to take over if another host fails. A couple
of disadvantages to this admission strategy are:
• One dedicated host per cluster (the maximum allowed) is probably not sufficient in a large
cluster.
• This resource is not used at all unless there is a failure, so under normal circumstances,
the fail-over host resource is wasted.
For more information and detailed procedures, see the VMware vSphere Administration Guide
and the VMware vSphere Availability Guide.
vSphere Replication
VMware vSphere Replication is a feature of the vSphere platform. It copies a virtual machine
to another location, within or between clusters, and makes the new copy available for restoration
through the vCenter Server Web-based user interface. It provides hypervisor-based virtual
machine level replication and recovery from a primary site to a secondary site.
vSphere Replication continues to protect the virtual machine by replicating changes made to
the virtual machine, to the copy. This ensures that the virtual machine remains protected and
is available for recovery without requiring restore from backup. vSphere Replication is provided
as a no-charge component of all eligible vSphere licenses.
Unified management of vSphere Replication is provided through the new vSphere Web Client.
This provides a common and unified screen for all aspects of virtual data center management,
including many aspects of protection, such as replication, backup and restore.
There are two levels of replication, the storage level and the hypervisor level. vSphere
(hypervisor level) replication is a part of SRM, and as a separately installed extension to vCenter
Server.
45
Mitel Virtual Appliance Deployment
Mitel virtual appliances are not compatible with the Fault Tolerance feature (see Table 12 on
page 40). FT is very resource heavy, especially in the level of CPU resources consumed on
multiple hosts, and the network load necessary to maintain synchronized state across the
instances on those hosts negatively affects the performance of the Mitel real-time applications.
A combination of Mitel application level resiliency and VMware HA, as described previously, is
recommended.
While not directly related to resiliency and availability, DRS can play an important role in ensuring
optimal performance during other operations. DRS ensures that VM resource reservations are
honored, and can re-distribute loads to ensure overall optimal performance of all applicable
VMs in the cluster. This can, in turn, help ensure optimal responsiveness and real-time behavior
(voice quality) of Mitel virtual appliances.
vMotion can also be used to manually migrate running VMs to different hosts in the overall
system, for example, in order to do maintenance of specific host servers (host software updates,
trouble-shooting of host issues, hardware configuration, and so on). Migration to different hosts
avoids downtime of the Mitel solution components during the host server maintenance
operation, increasing overall availability. This is not feasible in the physical world.
Mitel strongly recommends enabling DRS at the cluster level where Mitel virtual appliances are
deployed. To avoid the administrator having to monitor distribution behavior, the DRS
automation level can be set to Fully Automated.
Note: Since DRS uses vMotion to dynamically distribute loads at run time, and vMotion
operation can cause momentary pauses in voice streaming, it is an administrator
decision whether to enable full automation, and if enabled, what Migration threshold
to set (a more aggressive migration threshold will result in higher chance of vMotion
operating during active voice streaming).
When DRS is enabled for Mitel virtual appliances, such as MiVoice Business Virtual and MBG
Virtual that implement their own resiliency mechanisms, it is extremely important to set
anti-affinity rules within DRS configuration to ensure that primary and secondary elements (for
example, a primary MiVoice Business Virtual node and a secondary MiVoice Business Virtual
node) are never allowed to run on the same physical host, since a single host failure could then
defeat the inherent Mitel resiliency mechanisms. See also “Anti-affinity rules” on page 34.
46
Chapter 3
DEPLOYING ON MICROSOFT HYPER-V
INFRASTRUCTURE
Mitel Virtual Appliance Deployment
48
Deploying Mitel virtual applications in a Hyper-V virtual infrastructure
The latest releases of the following Mitel virtual appliances are supported on Hyper-V when
installed in the standalone product—Hyper-V Server 2012 R2—and when installed as a role in
Windows Server 2012 R2. Older versions of Hyper-V and older versions of the Mitel products
are not supported.
• MiVoice Business
• MiVoice Business Express
• MiCollab
• MiCollab Client
• NuPoint Unified Messaging (no longer available for purchase in stand-alone mode)
• MiVoice Border Gateway
• MiContact Center (also tested in Microsoft Hyper-V Server 2008)
Mitel currently supports Power On and Shutdown Guest functions in the Hyper-V environment.
Advanced features and clustering have not been fully tested and are not supported
For best practices for physical servers hosting Hyper-V roles, refer to the Microsoft
documentation, and the following recommendations:
• Ensure that all hosts that will be used to support real-time sensitive applications, such as
NuPoint UM Virtual, MiVoice Border Gateway Virtual, and so on, are configured for maxi-
mum performance. This includes processor and BIOS settings requirements as listed in
“Minimum hardware requirements” on page 9.
• Supply plenty of bandwidth: 1 GB across each of at least four physical ports on each host
server is considered the minimum for most deployments.
• Separate the Virtual Local Area Network (VLAN) and IP interfaces for storage traffic (as-
suming Storage Area Network (SAN), Network File System (NFS), or similar IP storage is
used). This provides adequate throughput for storage operations, while isolating traffic from
other flows.
49
Mitel Virtual Appliance Deployment
• Preferably, support storage networking on two or more physical ports on the host servers,
directed at different network switches to provide resiliency. You should also consider con-
figuring these ports for NIC bonding (using Link Aggregation Control Protocol (LACP) or
similar) to improve overall throughput.
• Voice traffic to and from the Mitel applications should be separate from data traffic. Imple-
ment a dedicated voice VLAN, and support it through multiple physical ports. The same
recommendation applies for physical implementations.
Adding NICs
Table 13 shows how many NICs you need for various configurations of the Mitel virtual
applications in the Hyper-V environment. Refer to the Microsoft documentation for instructions.
Storage
In planning connections to, and stability of, the network storage, consider the following
guidelines:
• Ensure adequate throughput to support the intended applications. Refer to the product-spe-
cific documentation for detailed requirements.
• Support storage using multi-path networking, with a minimum of two physical ports on each
host server and on each of the storage units, directed at different network switches. This
provides resiliency.
• RAID protection of all storage is strongly recommended.
Mitel virtual applications support for various storage architectures matches that of Hyper-V ,
unless specifically stated otherwise. This includes iSCSI, NFS, Fibre Channel, and host-local
storage. For details, refer to the Mitel documentation for each virtual application, available at
Mitel OnLine.
Note: Regardless of the storage technology used, ensure that the storage used by Mitel
applications meets or exceeds the specific requirements of the applications involved, in
terms of IOPS and latency. See Table 14, “Hyper-V resource requirements,” on page 51,
and application-specific Engineering Guidelines.
50
Deploying Mitel virtual applications in a Hyper-V virtual infrastructure
Table 14 shows the resource allocation and resource reservation requirements for each of the
supported Mitel virtual applications.
51
Mitel Virtual Appliance Deployment
PREVIOUS RELEASES
MiVoice 8.0, 7.3, 7.2, 250 devices 2 20 GB 50% 1.5 GB
Business 7.1, 2 GHz
Virtual 7.0 SP1
1500 devices 6 20 GB 50% 2.0 GB
6 GHz
2500 devices 8 20 GB 60% 2.0 GB
10 GHz
Embedded Voice included
Mail
MiCollab 8.0, 7.3, 7.2, 250 users 4 50 GB 50% 8.0 GB
Virtual 7.1, 7.0 4 GHz /
Multi-App 6.0 SP1 5.0 GB
1500 users 8 90 GB 75% 8.0 GB
12 GHz /
7.0 GB
MiCollab 8.0, 7.3, 7.2, 5000 users
Virtual 7.1
Single App 6.0 SP1
MiCollab 8.0, 6.3, 500 ports 16 120 GB 35% 16.0 GB
AWV 6.2+, 6.1, 12 GHz
6.0
MiCollab 8.0, 7.3, 7.2, 5000 ports 16 120 GB 50% 16.0 GB
Client 7.1, 7.0, 6.0 16 GHz
SP4
NuPoint UM 8.3, 8.2, 8.1, 120 ports 16 120 GB 30% 16.0 GB
8.0, 7.0 SP1 10 GHz
NuPoint UM 8.3, 8.2, 8.1, 60 ports 4 130 GB 50 % 4.0 GB
Standalone 8.0 4 GHz
7.0 SP1
120 ports 8 260 GB 50 % 6.0 GB
8 GHz
MiContact 8.1 Enterprise 8 120 GB 35% 16 GB
Center 6 GHz
Remote IVR 8.1 Enterprise 4 60 GB 25% 4 GB
2 GHz
52
Deploying Mitel virtual applications in a Hyper-V virtual infrastructure
Notes:
1. It is required that the memory allocation for Hyper-V is configured by setting the startup RAM to a
static value rather than configuring it as Dynamic Memory. It is also recommended that the Memory
weight be set to High to ensure that the virtual machine is given priority by Hyper-V.
2. The number of Virtual Processors is set using the Number of Virtual Processors setting.
Corresponding to this, the Virtual Machine Reserve Percentage should be set in accordance with
the values in the table. Note that this represents Mitel's recommended Virtual Processor and
Reserve Percentage based on a 2.0 GHz processor. The percentage allocation can be varied
depending on processor speed, but the GHz reservations must remain consistent and the number
of vCPUs should also remain consistent with what Mitel recommends.
3. While there is a small difference in the measurements made for Network I/O and IOPS on Hyper-V,
the numbers measured for ESXi generally apply and can be used as a reference when deploying
Mitel applications on Hyper-V. See Table 8, “VMware required resource reservation and allocation,”
on page 21.
4. Hyperthreading is recommended for optimal performance.
Resource dimensioning
You must have a minimum of 2 GB of memory allocated for the Hyper-V Hypervisor. Refer to
the Hyper-V documentation on the Microsoft web site. Refer to
http://technet.microsoft.com/en-us/library/hh831531.aspx.
CAUTION: You must maintain CPU Headroom of 30% over the peak load to
guarantee service level agreement (SLA) performance.
One of the major advantages of virtualization is that you can co-locate multiple VMs on one
host server.
Because most of Mitel's virtual applications provide real-time voice, the customer will
experience voice quality issues if insufficient CPU resources are allocated. Hyper-V uses
53
Mitel Virtual Appliance Deployment
priorities to guarantee that the specified amount of the resource (CPU cycles and memory) is
always available to the VM, regardless of the needs of other VMs sharing the resource.
Set the CPU priority to High for the voice-sensitive Mitel applications. Lower settings can be
used for the other applications.
Dynamic memory is not supported for Mitel Virtual Appliances. Mitel Virtual Appliances are
32-bit machines and Dynamic memory works only on 64-bit machines.
Assign the amount of Start up Memory from Table 14, “Hyper-V resource requirements,” on
page 51 and make sure Dynamic memory is not selected.
Clock Sync
The VM guest must not do clock sync with both NTP and sync-to-host. Turn off one of them.
The host NIC driver should be up to date. NIC Teaming should be used and Virtual Machine
Queue (VMQ) should be enabled.
Product-specific guidelines
For additional product-specific restrictions and requirements, see the Mitel Engineering
Guidelines for each product. These guides are listed in Table 9 on page 35:
The Engineering Guidelines also list limits, requirements, and recommendations related to:
• I/O considerations, including effects on performance
• Power conservation
• Security
54
Installing and upgrading on Hyper-V
Note: When choosing the license type to purchase, select the virtual license, but
installation is manual. As with installation on the Industry Standard Server (ISS), you
install MSL, and then the virtual appliance.
After creating the Hyper-V VM, follow the physical installation procedures for the particular Mitel
application. Refer to the documentation for the Mitel product you are installing for detailed
instructions.
Note: Hyper-V VMs running the MSL OS do NOT support the connection of USB drives
directly to the OS. This means that in Mitel virtual application installations such as
MiCollab Virtual, you cannot scan for attached ISOs on a USB drive.
CAUTION: Make sure that, while installing Windows Server 2012-R2 on the host
system, no USB drives are plugged in, either externally or internally. If there are USB
drives present, the installation fails with a "partition error" as it tries to install the
operating system on the USB device.
55
Mitel Virtual Appliance Deployment
Note: When choosing the license type to purchase, select the virtual license, but
installation is manual. As with installation on the Industry Standard Server (ISS), you
install MSL, and then the virtual appliance.
When configuring the virtual environment, the operating system should be set as CentOS Linux
6 (32 bit). The Guest Hardware should be configured with the number of processors, amount
of memory and disk size as specified in Table 14 on page 51 and the Engineering Guidelines
for the application being deployed. The disk should be added to the IDE Device.
The CPU priority choices are High, Medium, Low, and Custom. Set the CPU priority to High
for the voice-sensitive Mitel applications. Lower settings can be used for the other applications.
56
Installing and upgrading on Hyper-V
Before starting the machine (Action > Start) and before MSL is installed, you must modify the
Virtual Machine (in Settings) to match Table 14 for the specific product you are installing. This
is a mandatory step.
Refer to Mitel Standard Linux Installation and Administration Guide, available on Mitel OnLine.
Installing Mitel virtual applications in Hyper-V is very similar to installing on a physical server.
For Hyper-V installations, you use the same MSL and application ISO images as for installing
during the physical procedures, but you must still purchase the virtual version of the license for
the Mitel products.
Refer to the installation documentation for the specific Mitel product on Mitel OnLine for detailed
instructions for installing the appliance on a physical server.
After installation is complete, the Mitel virtual solutions for Hyper-V must maintain online
connectivity to the AMC and are subject to the same Sync Expiry rules in place for
vSphere-based deployments.
Because there is no pre-installed disk image option as there is for VMware installations, any
changes to the configuration of the virtual applications such as revised resource requirements
must be made manually.
As is done for physical server deployments, you will typically need to upgrade MSL and the
application together. Refer to the Installation and Maintenance Guides and Release Notes for
product-specific guidance.
There are a couple of upgrade variants that do not exist for physical servers, to ensure that you
can revert to your starting place if problems occur in the upgrade process.
• Before beginning the upgrade, you clone the virtual machine (VM). Then you begin the
upgrade on a copy of the VM. The old VM can be kept in a shut-down state until the new,
upgraded virtual application is verified to be working properly. In the event of a problem
with the new virtual application, the old virtual application can be powered up in seconds
and be back in service, exactly as it was before the upgrade was started.
• Alternatively, you can take a snapshot before beginning the upgrade process. Use of the
Hyper-V snapshot is recommended only for use during a system upgrade, and only when
the VM is powered down. You can take a snapshot before starting the upgrade, when the
system is out of service. If there is a problem with the upgrade, you can revert to the snapshot
57
Mitel Virtual Appliance Deployment
to restore the appliance to service. After the upgrade is complete and verified, you must
remove the snapshot. Removing the snapshot ensures adequate performance for voice-in-
tensive Mitel virtual applications.
Use this upgrade method for minor release and Service Pack releases.
• Fresh Install/Restore method
Use this upgrade method for major release , or if indicated in the product release notes.
58
Installing and upgrading on Hyper-V
8. After the first AMC synchronization has occurred, verify that the upgraded licenses have
been correctly applied.
Linux Integration Services allows Hyper-V to interact with the Mitel guest VM. This allows
operations such as shutting down the VM from Hyper-V. Microsoft has provided the software
for the Linux Integration Services to the open source communities.
The drivers necessary for Linux Integration Services are included in MSL (introduced in MSL
10.0). It is not necessary to add another Mitel Blade to get these components. Do not attempt
to upgrade these components yourself. Please refer your questions to Mitel Support.
59
Mitel Virtual Appliance Deployment
60
Chapter 4
INSTALLING AND UPGRADING
Mitel Virtual Appliance Deployment
62
Installing Mitel virtual appliances and applications
Note: Manually installing MSL in a VM, and then manually installing the corresponding
Mitel virtual appliance is not supported.
Note: MiContact Center Office: On first configuration after OVA file deployment,
configure the network settings. After new network settings have been applied, the
Network Configuration screen will no longer be available.
In the case of any network issues (IP address conflict, for example) close the installer
window (click Cancel) and troubleshoot the network settings from the Windows
Network and Sharing Center. Then the MiContact Center Office setup wizard can be
run and the configuration can be resumed.
Disk provisioning
Mitel virtual appliances that require low latency for voice streaming (such as MiVoice Business
Virtual, NuPoint Virtual, and MiCollab Virtual) require the use of thick provisioning of the disks.
Testing has found that the additional run-time overhead from using thin-provisioning noticeably
affects voice quality.
Thick-provisioned disks may be either eager-zeroed or lazy-zeroed. Mitel recommends the use
of lazy-zeroed disks.
Lazy-zeroed disks are zeroed on the first write to each block. The first write to each block is
much slower than every successive write. Conversely, eager-zeroed disks have all their blocks
zeroed during deployment when the disk is created for the virtual appliance.
Eager-zeroed disks take longer to deploy (on the order of a few minutes depending on the size
of the disk), but do not risk slow disk performance at random times during run-time, as
not-yet-used disk blocks are zeroed.
63
Mitel Virtual Appliance Deployment
Refer to VMware Knowledge Base article 1011170, for instructions for determining whether the
disk of an already-installed VM is lazy-zeroed or eager-zeroed.
Network driver
Mitel virtual appliances require low latency for voice streaming. Therefore, they must use the
VMXNET3 network driver.
Note: If you are installing multiple Mitel vApps, it is strongly recommended that you
deployed them one at a time. This will avoid having the guest properties for all of the
vApps displayed at the same time, which can cause confusion in completing the
deployment.
There may be product-specific settings that apply to the OVA you are deploying. There may
also be additional configuration steps to be performed after the OVA is deployed. Refer to the
product-specific documentation for the details.
Note:
• See Table 8, “VMware required resource reservation and allocation,” on page 21 for
detailed resource and capacity information.
• Refer to the product-specific installation guide for details and limitations particular to
installing the product OVA, and for post-install and configuration instructions.
The following procedure is an example of how to deploy vMiCollab using the vSphere 5.5 web
interface to an ESXi host via vCenter Manager. Note that the screens that are displayed during
the deployment wizard are dependent on your network configuration and VMware environment.
Details of the deployment for MiVoice Business, MiVoice Border Gateway, and MiVoice
Business Express are added in.
Note: It is strongly recommended that the deployment be done using the VMware Web
Client, as all the new features used for Mitel OVA deployment may not be available in
the older client.
1. Obtain the Mitel OVA for the product you are deploying.
CAUTION: Do not install Mitel Blades from any CD ROM drive that may be
attached to the virtual machine.
2. Log into the vSphere Web Client with your user name and password.
3. In the vSphere Client, In the upper right corner of the tool bar, click . Ensure that you
have installed the latest VMware Client Integration Plugin.
4. In the left-hand menu, click vCenter Inventory Lists.
64
Installing Mitel virtual appliances and applications
Note: When deploying in a vCenter environment, the wizard may prompt for a Data
store and Network Mapping if several options are available. Contact your Data Center
administrator for details about which Datastore or Network Mapping to use.
65
Mitel Virtual Appliance Deployment
15. Click Next. The Select a resource screen appears. When installing in a vCenter Server,
select the Resource Pool for the virtual instance of the OVA. This page is displayed only if
the cluster contains a resource pool.
16. If you are deploying the Mitel vApp in a vCenter Server, select the Host/Cluster for the
vMiCollab instance. The Host/Cluster screen only appears if you are deploying on a vCenter
Server.
17. Click Next. The Select storage screen appears.
18. Set the virtual disk format to Thick Provisioned Lazy Zeroed.
19. Select a destination Datastore.
20. Click Next. The Setup networks screen appears. This screen is different for each of the
Mitel products that use the custom Setup networks screen.
If the network defined in the OVF template does not match the name of the template on
the host to which you are deploying, you are prompted to configure the Network Mapping.
The required settings may be different, depending on your deployment configuration.
• Network Edge (Server-Gateway) Mode: In this configuration mode, the server func-
tions as a firewall/Internet gateway with two Ethernet interfaces. One interface is
connected to the internal network (LAN) while the other is connected to the external
network (Internet). Select the destination LAN and WAN networks for the OVF template.
These are the “Associated Networks” that are assigned in the LAN and WAN IP Pools.
You must assign the LAN and WAN destinations to different networks.
• LAN Only (Server-only) Mode: In this configuration mode, the server is only connect-
ed to the internal network (LAN). For this mode, select only a destination LAN network
for the OVF template. The WAN destination is ignored.
66
Installing Mitel virtual appliances and applications
If you are installing MiCollab, MiVoice Business, MiVoice Business Express, MiVoice Border
Gateway, or MiCloud Management Gateway, continue through the Properties screen
instructions.
If you are installing any other Mitel OVA, skip to Step 23.
67
Mitel Virtual Appliance Deployment
22. For certain Mitel appliances, the Properties screen appears. The Properties screen only
appears for deployments that use vCenter. These are the Mitel virtual appliances that use
the additional Properties screen.
• MiCollab
• MiVoice Business
• MiVoice Business Express
• MiVoice Border Gateway
• MiCloud Management Gateway
You use this screen to configure the MSL operating system parameters. Complete the fields
in this screen using the information that you recorded in preparing for installation. Mandatory
fields are highlighted with a red border. “Properties page - Networking settings by OVA type”
on page 69 describes how to fill in the Properties screen for each product.
CAUTION: Make sure to enter the correct values for each of the properties.
1. If any mandatory property values are missing, the wizard finishes, but the virtual
appliance will not be able to boot. A corresponding error message appears, and
you will have to re-start the OVF deployment from the beginning.
2. If all mandatory property values are set, but one or more of the values are invalid,
the Mitel Virtual Framework (MVF) blade attempts to retrieve the settings it can,
and set them as defaults for the MSL console. You will have to set parameters
using the MSL console configuration procedures. See the product-specific guide
for details.
3. You can use this screen only when setting up the LAN IP and WAN IP addresses
for the initial deployment of the appliance. After initial boot-up, you must use the
MSL server console for the product to modify the LAN IP or WAN IP addresses.
LOCALIZATION
Timezone Select the MSL time zone setting.
Default: North America.
Keyboard Select the keyboard used by the MSL console.
Restore from backup Select to open the MSL console and step through a
database restore. See the MSL Installation and
Administration Guide.
Default: Disabled
APPLICATION
Password This is the initial administrator password.
Hostname The host name to be used by the virtual machine. The
default is MiVB.
Note: This field can be left blank for template creation.
68
Installing Mitel virtual appliances and applications
Note: To create a blank template for cloning, leave the following fields empty:
Administrator Password, Hostname, Domain Name, LAN and WAN IP addresses.
After you create the clone, you must complete these fields before you can proceed with
deployment. You cannot clone an active (deployed) virtual machine.
LAN IP Address The IP address of the LAN side of the MiVoice Border Gateway.
69
Mitel Virtual Appliance Deployment
70
Installing Mitel virtual appliances and applications
Notes:
1. IP addresses must be specified for the LAN, WAN, and Optional LAN; otherwise, the virtual
appliance will not power on. If you are deploying the virtual machine in LAN only (server-only) mode,
set the WAN IP address to 0.0.0.0.
2. For Network Edge (Server-Gateway) deployments, ensure that the LAN IP and WAN IP addresses
are on different subnets and the Gateway IP address is on the subnet of the WAN IP address.
3. You can use this screen only when setting up the LAN IP and WAN IP addresses for the initial
deployment of the appliance. After initial boot-up, you must use the MiCollab server console
interface to modify the LAN IP or WAN IP addresses.
71
Mitel Virtual Appliance Deployment
25. After the OVA template file has been deployed, the new virtual machine appears in the
inventory list in the left side navigation pane.
26. Do one of the following:
• If you deployed the vApp on vSphere vCenter and used the Properties screen to
configure the MSL Operating System parameters, continue with configuring the vApp.
See the specific vApp deployment or installation guide for application configuration
details.
• If you did not use vSphere vCenter, you must configure the MSL Operating System
parameters. See the specific vApp deployment or installation guide for MSL configu-
ration details.
72
Upgrading Mitel virtual appliances and applications
Upgrading on Hyper-V
Hyper-V installation and upgrades follow the physical model. Refer to the installation guides
for the specific product or application to perform upgrades.
Upgrading on VMware
Mitel virtual appliances include three components, all of which must be considered when making
upgrade decisions:
• the virtual application
• the Mitel Virtual Framework (MVF) blade
• the Mitel Standard Linux (MSL) OS
Note: All products use MSL except for MiContact Center Office, which runs on
Microsoft® Windows®.
Note: The Blade upgrade is not available if the OVA dimensions have changed. If the
OVA dimensions have changed in the release, upgrade by installing the new OVA.
Refer to the Installation and Maintenance Guides and Release Notes for product-specific
guidance.
There may occasionally be releases where something in the OVA packaging itself is modified
and an OVA upgrade is required. In such cases, this requirement will be highlighted in the
Release Notes.
73
Mitel Virtual Appliance Deployment
Software upgrades
You upgrade each Mitel virtual appliance individually, using Mitel upgrade procedures. VMware
vCenter Update Manager is not supported for Mitel virtual applications. Log and core files must
be backed up separately if you want to keep them.
Do not update the virtual appliance manually. Manual updates miss changes that affect the
virtualization-specific aspects of the product, such as resource reservation changes, VMware
Tools updates, and hooks to the VMware environment. Following the procedures in this section
will ensure that any virtualization-specific changes are applied.
OVA Upgrade method - for major release upgrades, or as indicated in the product release notes:
1. Back up the existing version, using the MSL backup method.
2. Power down the virtual machine.
3. Apply the appropriate upgrade license, at the AMC, to the existing ARID for the Virtual Appliance.
4. Deploy a new virtual appliance using the new OVA file for the upgrade version of the virtual appliance.
5. After OVA deployment, ensure that your virtual machine has the resources required for the new release,
as appropriate for your environment. The OVA deployment will set default values that may need
adjustment for your specific needs.
6. Restore the backup data. The deployment process prompts you for the backup file to restore.
7. After the restore is complete, and the first AMC synch has occurred, verify that the upgraded licenses
have been correctly applied.
Note: MCD Release 6.0+ and MiVoice Business Virtual: Each MiVoice Business Virtual
requires two IP addresses, one for MSL and one for MCD.
MCD Release 5.x and previous: Each vMCD requires five IP addresses, one for MSL
and 4 contiguous IP addresses for the MCD space.
74
Upgrading Mitel virtual appliances and applications
VMware snapshot is recommended only for use during a system upgrade, and only when the
VM is powered down. You can take a snapshot before starting the upgrade, when the system
is out of service. If there is a problem with the upgrade, you can revert to the snapshot to try
again. After the upgrade is complete and verified, you must remove the snapshot. This ensures
adequate performance for voice-intensive Mitel virtual appliances. It is not necessary to clear
the hardware ID on the AMC.
Note: Do not restore MSL backups to your virtual machine. This is not supported.
75
Mitel Virtual Appliance Deployment
previous versions of ESXi. In a mixed environment, this will affect which servers vMotion or HA
can relocate each VM to, complicating operation of the virtual infrastructure.
Virtual hardware version 10 is not supported due to issues with some versions of vSphere client.
All Mitel virtual appliances can be deployed, and will run, on virtual hardware versions 7, 8, or
9. Current release Mitel virtual appliances are packaged to automatically deploy using the
highest available virtual machine version available on the ESXi host they are initially deployed
on.
If you are deploying on a mixed environment—a cluster containing multiple different ESXi
versions—start by deploying on a host server running the lowest ESXi version available within
the cluster. This will allow vMotion and HA operations on all hosts in the cluster.
Note: Under some circumstances, newer versions of ESXi may prevent normal virtual
machine startup due to interactive prompting and other issues that appear if they are
not running the latest virtual hardware version available on the host. This can affect
Power on, vMotion, HA, and other availability features. Virtual machine version upgrade
to the latest available version fixes these issues. These are not Mitel-specific problems,
but can affect any applications running on ESXi.
It is strongly recommended that you upgrade the virtual hardware of deployed Mitel
virtual machines to the latest available on the host ESXi. After the upgrade is done, you
cannot downgrade to an earlier Virtual Machine Version that may be required for running
on older ESXi versions.
Refer to VMware documentation for performing the upgrades, and for more information
about ESXi/Virtual Machine version compatibility.
76
Upgrading Mitel virtual appliances and applications
77
Mitel Virtual Appliance Deployment
Notes:
1. The vSphere Client can be used to manage VMs with VMware virtual hardware versions up to
version 9, but you will not be able to create new VMs. If you need to both create or manage VMs,
you must stay at virtual hardware version 7 or 8, and you must manage your VMs with the vSphere
Client.
78
Upgrading Mitel virtual appliances and applications
Table 21: vCenter and vSphere compatibility with Mitel virtual appliances
VSPHERE WEB CLIENT1 VCENTER SERVER AND VSPHERE CLIENT1
6.5 6.0 5.5 6.5 6.0 5.5 5.1 5.0
CURRENT RELEASES
Mitel Mass 6.5 Y
Notification
MiVoice Business 9.0 Y Y Y Y Y Y
Virtual
MiCollab Virtual 8.1 Y Y Y Y Y Y
MiCollab Client 8.1 Y Y Y Y Y Y Y
MiVoice Border 10.0 Y Y Y Y Y Y
Gateway Virtual
MiContact Center 6.2 SP1 * Y Y Y Y Y Y
Office
MiCloud 6.0 Y Y Y Y Y Y
Management Portal
(Oria)
MiCloud 5.0 Y Y Y Y Y Y
Management
Gateway
MiContact Center 8.1 Y Y
Business
MiVoice Business 8.0 Y Y Y Y Y
Express
MiVoice Enterprise 9.0
Manager
PREVIOUS RELEASES
MiVoice Business 8.0 Y Y Y Y Y Y
Virtual
7.3, 7.2, Y Y Y Y Y Y
7.1, 7.0
MiVoice Business 7.3, 7.2, Y Y Y Y Y Y
Express 7.1, 7.0
79
Mitel Virtual Appliance Deployment
Table 21: vCenter and vSphere compatibility with Mitel virtual appliances
VSPHERE WEB CLIENT1 VCENTER SERVER AND VSPHERE CLIENT1
6.5 6.0 5.5 6.5 6.0 5.5 5.1 5.0
vNuPoint UM 6.0 Y Y Y Y
MiCollab Client 8.0 Y Y Y Y Y Y Y
7.3, 7.2, Y Y Y Y Y Y
7.1, 7.0,
6.0
MiVoice Border 9.x, 8.x, Y Y Y Y Y Y
Gateway Virtual 7.x
Oria Virtual 5.3, 5.2, Y Y Y Y
5.1
Oria Virtual 5.0 SP1, Y Y Y Y Y
5.0
MiContact Center 8.0 Y Y Y
Business
vCSM 6.0 Y
vUIC 3.0 Y Y Y Y Y Y
Note: The vSphere Client can be used to manage VMs with VMware virtual hardware versions up to version 9,
but you will not be able to create new VMs. If you need to both create or manage VMs, you must stay at virtual
hardware version 7 or 8, and you must manage your VMs with the vSphere Client.
VMware Tools
VMware Tools is a software deliverable provided by VMware. It is installed and run inside the
Guest O/S, to integrate the virtual and guest environments, and to improve VM performance.
VMware Tools packages are designed to integrate with particular, well known commercial
operating systems like Red Hat Linux, Windows, and so on. In the case of Mitel Standard Linux
(MSL), which is not considered well known by VMware, VMware Tools distributions are
managed within the Mitel Virtualization Framework (MVF). Mitel virtual applications based on
MSL include the MVF Blade as part of their OVA packaging, with the MVF containing the latest
VMware Tools at time of product release. Do not attempt to install or upgrade the VMware
Tools using VMware vSphere functions.
In the case of Mitel applications containing MVF in their OVA packaging, in the vSphere Client
Summary Tab, the VMware Tools field displays “(3rd-party/Independent)”. This means that
VMware cannot manage or upgrade the VMware Tools in that Guest O/S. The status will be
displayed as Unmanaged in the vSphere 4.1 Client.
80
Upgrading Mitel virtual appliances and applications
While it is desirable to always upgrade VMware Tools to the latest available for the version of
vSphere being used, there are constraints on how this can be performed.
• For Mitel virtual appliances containing MSL 10.0 and above combined with MVF 2.0 and
above, an MVF Blade upgrade can be applied in order to upgrade VMware Tools. From the
MSL blades panel, it may be possible to upgrade MVF (and VMware tools) if a newer MVF
version is shown as an upgrade option through the Mitel AMC. OVA upgrade to a newer
version of the appliance containing the desired VMware Tools level is also applicable. See
“OVA upgrade advantages” on page 75.
• For Mitel virtual appliances that contain either MSL 9.x series or MVF 1.x series, MVF Blade
upgrade is not applicable. OVA upgrade to a newer version of the appliance, containing
the desired VMware Tools, is the only method available. See “OVA upgrade advantages”
on page 75.
• Those Mitel Applications that are not packaged as an OVA, but are installed into a pre-ex-
isting VMware Virtual Machine and Guest O/S, can have their VMware Tools software
installed and upgraded in the Guest O/S using the normal VMware-supplied methods. This
also applies to OVA-packaged virtual appliances based on Windows OS (MiContact Center
Office Virtual, for example). Refer to the VMware documentation for instructions.
Table 22 summarizes VMware Tools upgrade methods applicable to each Mitel virtual appliance
or application.
Table 22: VMware Tools packaging and upgrade by Mitel virtual appliance
VMWARE TOOLS
INCLUDED VMWARE TOOLS UPGRADE
CURRENT RELEASES
Mitel Mass Notification 6.5 Y Firmware upgrade
MiVoice Business Virtual 9.0 Y OVA upgrade
MiCollab Virtual 8.1 Y OVA to 6.0 and above
MiVoice Border Gateway 10.0 Y MVF blade upgrade or OVA to 8.1 and above
Virtual
NuPoint UM Virtual 8.3 Y MVF blade upgrade or OVA to 7.0 and above
(This is the final release of
NuPoint VM.)
MiVoice Business Express 8.0 Y MVF blade upgrade or OVA to 6.0 and above
Open Integration Gateway 4.0 Y MVF blade upgrade or OVA
MiContact Center Business 8.1 Y Use VMware methods
MiContact Center Office 6.2 SP1 Y
MiCloud Management 6.0 Y
Portal (Oria)
MiCloud Management 5.0 Y Use VMware methods
Gateway
MiVoice Enterprise 7.0 N Use VMware methods
Manager
81
Mitel Virtual Appliance Deployment
82
Chapter 5
MAINTENANCE AND
TROUBLESHOOTING
Mitel Virtual Appliance Deployment
84
Maintenance
Maintenance
After your system is up and running, there are regular tasks to perform to keep the system
running well:
• “Backups” on page 85
• “Restores” on page 88
• “Software upgrades” on page 74
• “Monitoring resource usage and performance” on page 89
Backups
Virtual machine backups can be done for the entire VM, as a whole, or at the product level,
and there are different ways to do the various types of backup. In general, you should start with
the Release Notes for each product for detailed information.
The following sections give background information on each backup and restore approach.
See the product-specific backup and restore methods, as found in the corresponding Mitel
product guides listed in Table 24.
To perform backups of Mitel virtual applications running on Hyper-V virtual machines, follow
the backup procedures in the Mitel product-specific documentation.
The sections that follow describe several ways of backing up the virtual appliance or the whole
VM, and the best ways to perform backups, depending on circumstances.
Snapshots are typically used to create a restoration point before installing a different version
of an application, or before performing potentially destabilizing or error-prone re-configuration
of the applications. After the new application is up and running, or re-configuration verified, the
restoration point must be removed to avoid any performance degradation. Snapshots should
not be used as a method of backing up the Mitel virtual appliance. Snapshots must be deleted
after restore is complete.
To perform a complete backup of your system, you can use VMware management functions
to back up the full virtual appliance VM, which includes all of the virtual applications running on
it. When creating backups of the whole VM, you must ensure that you have plenty of storage
space because the backups can be very large for a large system.
In the general case, you can take a snapshots of the whole VM while it is powered on (running),
suspended, or shut down.
Note: Live snapshots are not supported for Mitel virtual appliances.
Snapshots of a suspended Virtual Machine are also not supported.
Mitel virtual appliances must be powered down before a snapshot is saved.
85
Mitel Virtual Appliance Deployment
Snapshots preserve the state of a VM such that it can be returned to the same state repeatedly.
A snapshot captures the entire state of a virtual machine at a given moment in time. This
includes:
• Memory state: the contents of the virtual machine memory
• Settings state: the virtual machine settings
• Disk state: the state of all virtual disks on the virtual machine
• Power state: whether the virtual machine is running, suspended, or shut down
When reverting to a snapshot, these items are returned to their state at the time the snapshot
was taken. With snapshots, a user can create backup and restore positions in a linear process.
Snapshots are typically used as restoration points during a linear or iterative process, such as
installing update packages, or during a branching process, such as installing different versions
of an application.
Because Mitel virtual appliances are communicating with other devices on the network and
because the risk of a database corruption increases when live snapshots are taken, it is strongly
recommended that the Virtual Machine hosting a Mitel virtual appliances be powered down
prior to taking a snapshot.
Note: Keeping a snapshot consumes run-time resources, because the virtual disk
subsystem must keep track of changed blocks from the snapshot. This presents a load
on the system that can cause performance degradation; snapshots should be kept only
temporarily, and then deleted when they are no longer needed.
A VMware OVF export creates an image of the VM which can then be re-deployed in the same
way the Mitel virtual appliance OVA was deployed originally.
An OVF Export operation is supported only when the virtual appliance is shut down.
Mitel provides a backup strategy and capability with each individual Mitel application. Mitel
backups are done at the product level, so they use much less of the system’s resources, as
compared to VM snapshots or Data Recovery methods. Mitel application-specific backups are
an important part of upgrade strategy.
Note: Do not restore MSL backups to your virtual machine. This restore is not supported.
For information about Mitel product-specific backup and restore capabilities, see the
documentation for each product. Backup and restore procedures are described in the guides
listed in Table 24. These guides are available on Mitel Online.
86
Maintenance
vSphere Data Protection also provides a centralized management interface to enable backup
and recovery of your VMs directly through VMware vCenter Server, and automatically monitors
VMs that are moved by HA, vMotion, and DRS, so that scheduled backups can continue
uninterrupted.
During a backup, Data Recovery creates a quiesced snapshot of the virtual machine.
Deduplication is automatically performed with every backup operation.
Data Recovery can concurrently back up a maximum of eight virtual machines. To start multiple
backups, CPU utilization must be less than 90 percent. Due to memory constraints, Data
Recover does not support using more than two backup destinations simultaneously. If more
than two backup destinations must be used, configure them to be used at different times.
You can create backup jobs that include which virtual machines to backup, where to store the
backups, and for how long. For details on installing and running Data Recovery, refer to the
VMware documentation.
Note: To avoid adverse effects on NuPoint Virtual voice quality, use vSphere Data
Protection backups only at times when the call rate is less than 500 Calls Per Hour (CPH).
The VADP allows third-party backup software providers to access VMFS volumes directly,
off-loading the work from the main processors. A VADP-based backup can be expected to have
no adverse effects on voice quality.
Note: Third party backup applications have not been specifically qualified by Mitel, and
are not fully supported.
87
Mitel Virtual Appliance Deployment
You should rely primarily on the Mitel product-specific backup recommendations; see Table 24
for the guides containing this information for the various products.
Notes:
1. See Table 24 for the product-specific guides.
2. MiVoice Business Virtual requires both an MiVoice Business backup and an MSL backup.
3. vUIC requires only an application backup.
Restores
If you perform backups using the product-specific backup procedures, you should also use the
product-specific restore routines described in the product documentation. See Table 24 for the
product-specific guides.
Note: Do not restore MSL backups to your virtual machine. This is not supported.
You can recover a MiCollab Virtual system database on the same virtual machine by deploying
the latest MiCollab Virtual OVF file, restoring your MiCollab Virtual database backup, and then
installing the latest MiCollab application software.
For detailed instructions, refer to the MiCollab Installation and Maintenance Guide.
88
Maintenance
Tracking performance and resource use in an automatic way is especially useful when
monitoring critical resources. VMware vCenter Operations integrates directly with third-party
monitoring tools for a complete view of your infrastructure. You can set the upper and lower
bounds of the “normal” range, and configure alerts that appear when performance criteria stray
outside of the normal range. For more information, see the VMware White Paper, Sophisticated
Dynamic Thresholds with VMware vCenter Operations.
Note: NuPoint has specific limits for minimum I/O storage that may require monitoring.
See the Mitel NuPoint Unified Messaging Engineering Guidelines.
89
Mitel Virtual Appliance Deployment
Troubleshooting
Each Mitel virtual appliance is installed with its own built-in diagnostics for use in troubleshooting
problems local to the specific Mitel virtual application.
Troubleshooting is especially difficult in the service provider environment in which the customer
is renting computing capacity from the service provider. In this deployment model, the Mitel
customer has no access to the service provider's virtual infrastructure. Isolation of issues within
the virtual infrastructure from problems within the Mitel virtual solution is almost impossible
without some level of cooperation from the service provider.
For the full VDT reference, see “Virtualization Diagnostic Tool (VDT)” on page 94. This reference
is also maintained in the Mitel Standard Linux (MSL) on-line help.
90
Troubleshooting
Although you can collect logs using either the vSphere Windows client or the new Web client,
VMware plans to drop support for the Windows client in the future, so it may be more appropriate
to use the Web client.
VMware has documented the process for collecting diagnostics using the Web client here:
http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&
externalId=2032892.
This is the definitive reference for this process but for your convenience, the process is
presented below.
To collect ESX/ESXi and vCenter Server diagnostic data using the Web client:
1. Start the vSphere Web Client and log in to the vCenter Server system.
2. Under Inventory Lists, select vCenter Servers.
3. Click the vCenter Server that contains the ESX/ESXi hosts from which you want to export
logs.
4. Click the Monitor tab and click System Logs.
5. Click Export System Logs.
6. Select the ESX/ESXi hosts from which you want to export logs.
7. Optional: Select the Include vCenter Server and vSphere Web Client logs option.
8. Click Next.
9. Select the system logs that are to be exported.
Make sure that the vmware*.log for your VM is acquired by selecting the check box
VirtualMachines/logs and the host.d log is acquired by selecting Logs/System. If you are
not concerned about the size of the log archive it would be best to select more options.
10. Select Gather performance data to include performance data information in the log files.
Note: You can update the duration and interval time for the data collection.
91
Mitel Virtual Appliance Deployment
Note: The host or vCenter Server generates ZIP bundles containing the log files. The
Recent Tasks panel shows the Generate diagnostic bundles task in progress.
14. After the download is complete, click Finish or generate another log bundle.
Note: You can use the events log to troubleshoot issues, particularly when trying to
determine a time stamp.
VMware has documented the process for collecting diagnostics using the Windows client here:
http://www.google.com/url?q=https%3A%2F%2Fround-lake.dustinice.workers.dev%3A443%2Fhttp%2Fkb.vmware.com%2Fselfservice%2Fsearch.d
o%3Fcmd%3DdisplayKC%26docType%3Dkc%26docTypeID%3DDT_KB_1_1%26externalId
%3D653&sa=D&sntz=1&usg=AFQjCNE9KLhCRK-Sz9cyzRq-F_B9Zv3Ytw
This is the definitive reference for this process but as a convenience the process is presented
below.
To collect ESX/ESXi and vCenter Server diagnostic data using the Windows client:
1. Open the vSphere Client and connect to vCenter Server or directly to an ESXi 5.x host.
2. Log in using an account with administrative privileges or with the Global.Diagnostics
permission.
3. Select an ESXi host, cluster, or data center in the inventory.
4. Click File > Export > Export System Logs.
5. If a group of ESXi hosts are available in the selected context, select the host or group of
hosts from the Source list.
6. Click Next.
7. In the System Logs pane, select the components for which the diagnostic information
must be obtained. To collect diagnostic information for all the components, click Select All.
Make sure the vmware*.log for your VM is acquired by selecting the check box
VirtualMachines/logs and the host.d log is acquired by selecting Logs/System. If you are
not concerned about the size of the log archive, it would be best to select more options.
8. If required, select the Gather performance data option and specify a duration and interval.
9. Click Next.
10. In the Download Location pane, click Browse and select a location on the client's disk
where you want to save the support bundle.
92
Troubleshooting
The Downloading System Logs Bundles dialog box appears and provides progress
status for the creation and downloading of the support bundle from each source. A Generate
system logs bundles task is created.
93
Mitel Virtual Appliance Deployment
VDT is delivered to all Mitel virtual appliances through the MVF blade with the exception of the
Windows-based MiContact Center Office, and MiVoice Business Virtual. As VDT collects virtual
infrastructure data on a periodic basis, it may impose additional CPU and network loading on
the ESXi host, virtual network and vCenter.
By default, Mitel virtual appliances are delivered with the Virtualization Diagnostic Tool activated.
Note: Full VDT functionality requires Administrator level access to either the ESXi host
or the vCenter that manages that host.
Note: The VDT is packaged with the Mitel Virtualization Framework, and is supported
on, and run from, Mitel Standard Linux (MSL).
This introductory release of the VDT is not supported on Microsoft Windows, but is
available as a separate deliverable, if required.
Required credentials
Table 27 shows the VDT features, and the credential level required for each one.
94
Virtualization Diagnostic Tool (VDT)
The Virtual Machine Properties table displays information about the virtual machine and the
Mitel Virtual Framework (MVF). The information is presented in two columns:
• Current Dimensions: Lists the configuration at the time that the current Mitel Virtualization
page was loaded. Refreshing the page resets the settings.
• First Boot Dimensions: Lists the configuration after the Mitel OVA package has been
installed and the settings configured, but before the virtual machine has been powered on
for the first time.
SETTING DESCRIPTION
MVF Version The version number of the Mitel Virtualization Framework (MVF), a software package
that enable Mitel applications to run in a virtual infrastructure. MVF has the capacity to
support multiple operating systems and hypervisor products.
Virtualization Agent The version number of VMware Tools, a suite of utilities that enhances the performance
Version (VMware of the virtual machine's guest operating system and improves management of the virtual
Tools) machine.
Hypervisor Version The version number of the VMware ESXi hypervisor that hosts one or more virtual
machines and their guest operating systems.
vCPU count The number of virtual Central Processing Units (vCPUs) configured on this virtual
machine.
Memory (MB) The amount of virtual physical memory available for use by the operating system on this
virtual machine.
Disk size (GB) The virtual disk size available for use by the operating system on this virtual machine.
NIC count The number of virtual network interface cards configured on this virtual machine.
95
Mitel Virtual Appliance Deployment
SETTING DESCRIPTION
CPU Reservation The guaranteed minimum allocation of CPU resources for this virtual machine.
(MHz)
Memory Reservation The guaranteed minimum allocation of memory resources for this virtual machine.
(MB)
CPU Limit (MHz) The upper limit of CPU resources that can be allocated to this virtual machine. This limit
is expressed in concrete units (Megahertz) and cannot be exceeded.
Memory Limit (MB) The upper limit of for memory resources that can be allocated to this virtual machine. This
limit is expressed in concrete units (Megabytes) and cannot be exceeded.
vCPU Speed (MHz) The speed of the virtual CPU, which is dependant on the speed of your underlying
processor. So if you have a 12 cores and a processor speed of 3.36GHz, That means a
virtual machine with a single vCPU running a single threaded application can consume
3.36GHz.
This setting defines what a single vCPU will consume, not the aggregated amount
among multiple vCPUs on a single virtual machine. Accordingly, if you have two vCPUs
this figure should be doubled.
Storage Monitoring
Use the following settings to specify whether the tool performs file system monitoring and
whether it reboots on read-only state.
SETTING DESCRIPTION
File System Use this setting to specify whether the system monitoring is enabled or disabled. If the
Monitoring main feature is enabled (the default), the system will check the file system for disk errors
every five seconds. If any disk errors are discovered, the system will generate a warning
notification and send it to the administrator e-mail address configured on the Email
Settings screen.
Reboot on read-only Select Yes or No to specify whether the system reboots when it enters a read-only state.
state A read-only state occurs when there are I/O errors on the virtual machine disk drives.
The default setting is Yes.
Note: File System Monitoring must be enabled for this feature to be effective.
Diagnostic Overview
The VDT constantly monitors the system in order to report on three alarm conditions and the
state of the last nightly analysis.
SETTING DESCRIPTION
Hypervisor Version Indicates whether or not the version of VMware ESXi Hypervisor is supported. The
Hypervisor is also known as the Virtual Machine Monitor (VMM).
If your ESXi version is not supported and you must switch to a supported version in order
to restore monitoring functionality. For example, if you are running ESXi 4.0 or earlier,
you must upgrade to version 4.1 or later.
96
Virtualization Diagnostic Tool (VDT)
SETTING DESCRIPTION
Current Dimensions Indicates whether the currently configured application resource dimensions are
supported.
If your configuration is not supported due to a setting (vCPU count, Memory, Disk size,
or NIC count) being out of boundaries, you can resolve any performance issues, by doing
the following:
Revert to the default configuration for your deployment.
Contact Mitel Product Support for assistance.
AMC Connectivity Indicates whether the Virtual Machine can connect to the Mitel Application Management
Center (AMC) for licensing purposes.
The two states are Connected and Error. In the Error state, the VM cannot connect to the
AMC. Check the networking configuration and Application Resource ID (ARID). See the
Mitel Standard Linux Installation and Maintenance Guide for more information.
Last Nightly Analysis Indicates the date and time that the last nightly analysis was completed, and whether any
problems occurred while it was being run. Upon successful completion, the nightly
analysis generates the following log file:
NIGHTLY-REPORT-YYYY-MM-DD.txt
To enable the VDT to collect statistics for the virtual machine and the host, and then use the
statistics to generate log files, you must enter credentials for the vCenter server or ESXi
hypervisor.
Note: For optimum results, enter credentials for the vCenter. Entering credentials for
the ESXi may result in connectivity problems if settings are changed on the hypervisor.
SETTING DESCRIPTION
FQDN or IP address Enter the Fully Qualified Domain Name or IP address of the vCenter or ESXi hypervisor.
Username Enter the user name required to access the vCenter or ESXi hypervisor.
Password Enter the password required to access vCenter or ESXi hypervisor.
97
Mitel Virtual Appliance Deployment
SETTING DESCRIPTION
Nightly Analysis Specify the one-hour period during which the nightly analysis will be run each day. Select
Time hours between 0-1 and 23-24.
Upon successful completion, the nightly analysis generates the following log file:
NIGHTLY-REPORT-YYYY-MM-DD.txt.
After a connection is established, the system obtains performance statistics for the virtual
machine and the host. You can click Run Diagnostics to manually generate log files and an
on-line report.
Note: For a newly installed system, wait for at least 15 minutes before clicking Run
Diagnostics.
Note: Without credentials, the system will not collect statistics or generate log files for
virtualization diagnostics.
Log files
The system generates log files containing performance and configuration data, and statistical
events.
The following log files are generated automatically by the system on a periodic basis.
98
Virtualization Diagnostic Tool (VDT)
Although the log files are primarily intended for use by Mitel Product Support, you may use
them to troubleshoot basic issues with the following issues:
• Performance problems: The system analyzes performance data and if it detects five con-
secutive “out of bounds” events, an problem will be reported. For example, if the virtual
99
Mitel Virtual Appliance Deployment
machine waits longer than two seconds to be serviced by the host, five times in a row, the
system will report a “CPU Ready” error. System events are registered every twenty seconds.
• Configuration problems: The system checks configuration data and statistical events on an
ongoing basis. If a problem is found, and error is logged immediately.
See the MSL online help for detailed information concerning the system settings which control
the generation of log file problems.
PERFORMANCE
PROBLEMS DESCRIPTION
CPU Ready (seconds) The virtual machine has exceeded the maximum amount of time that it can wait to
be run on the physical CPUs. The default is 2 seconds
CPU Usage (percent) The virtual machine has exceeded its CPU capacity limit, which is expressed as a
percentage of the total amount available. For example, with a limit of 50%, if the
virtual machine has four CPUs with 2 GHz processors, and you are running an
application that requires 6 GHz (75% of capacity), the limit has been exceeded by
25%. The default is 50%.
Disk Latency (seconds) The virtual machine has exceeded the maximum amount of time permitted for a
SCSI command to be issued by the guest operating system to the virtual machine
hard disk. The default is 0.02.
Network Usage (MB) The virtual machine has exceeded the maximum network utilization (combined
transmit and receive rates, in Megabytes per second). The default is 50.0 MB.
Memory Swapped (MB) The virtual machine has exceeded the maximum amount of memory, in
Megabytes, that can be swapped into memory from disk. The default is 0 MB.
Memory Use (MB) The virtual machine has exceeded the maximum amount of memory capacity that
it can use, expressed as a percentage of the total amount available. For example,
if the virtual machine has 4 GHz of memory, and you are running an application that
requires 3 GHz (75% of capacity), an event will be registered. The default value is
50%.
Number of Packets The virtual machine has exceeded the maximum number of received packets that
Dropped (average) can be dropped at the network interface. The default value is 0.
Disk Usage (MB) The virtual machine has exceeded the maximum amount of data, in Megabytes per
second, that can be read from the virtual machine hard disk. The default value is
30 MB.
Configuration Detection Description (Yes/No)
High VM-to-host CPU ratio If Yes is displayed, the ESXi host has exceeded the virtual CPU to host CPU ratio,
which is 0.79 by default. For example, if five virtual machines with 4 GHz vCPUs
are powered on, and the host has 8 physical/16 logical cores, then the ratio is 4 +
4 + 4 + 4 + 4 ÷ 16 = 1.25. Since 1.25 exceeds 0.79, a potential configuration issue
is detected.
High VM-to-host Memory If Yes is displayed, the ESXi host has exceeded the virtual memory to host memory
ratio ratio, which is 1.20 by default. For example, if five VMs are powered on, each using
2 GHz of memory, and the host has 8 GHz of physical memory, then the ratio is 2
+ 2 + 2 + 2 + 2 ÷ 8 = 1.25, which will cause an event to be registered.
Snapshots Present If Yes is displayed, the system checks to determine if snapshots are supported on
the virtual machine. Because snapshots create considerable disk I/O load, use of
this feature may degrade the voice quality of calls.
100
Virtualization Diagnostic Tool (VDT)
PERFORMANCE
PROBLEMS DESCRIPTION
Low CPU Speed (MHz) If Yes is displayed, the maximum speed of the virtual CPU, which is dependant on
the speed of the underlying processor on the ESXi host, has been exceeded.
No Hyperthreading (Ignore If Yes is displayed, the system checks to determine if hyperthreading is enabled on
if running on non-Intel the ESXi host.
processor) Note: This parameter can only report on Intel processors that support
hyperthreading. It cannot report on AMD or other non-Intel processors.
vMotion occurred If Yes is displayed, the system checks to determine if vMotion is enabled on the
ESXi host.
Low CPU Reservation If Yes is displayed, the guaranteed minimum allocation of CPU resources for this
(MHz) virtual machine has been exceeded.
Low Memory Reservation If Yes is displayed, the guaranteed minimum allocation of memory resources for
(MB) this virtual machine has been exceeded.
101
Mitel Virtual Appliance Deployment
Mitel application performance issues when CPU and memory requirements are
satisfied
There can be cases where, even though sufficient CPU and memory resources are available
for Mitel virtual machines, Mitel applications appear to be performing poorly, with problems
such as poor quality voice streaming, message playback disruptions, or dropped calls.
Probable cause
Resolution
102
Known issues (VMware infrastructure)
2. Examine host processor, BIOS, and related processor settings, to ensure that they meet
the requirements listed in Table 4, “Minimum hardware requirements,” on page 9. Correct
any settings or move the Mitel applications to fully qualified hosts.
3. 'CPU Ready' values that consistently exceed 100ms (<0.5%) of delay indicate onset of
CPU resource starvation which may impact the real time operation of voice and video
streaming services. Lower values of 'CPU Ready' may still be impacted. Ensure that the
CPU resource reservations are correctly configured and applied. Ensure that the resources
are not oversubscribed.
Mitel Virtual Appliance performance is sluggish, possibly after upgrade. Depending on the virtual
appliance, this could appear as poor voice quality.
Probable cause
Snapshots were created before upgrade, and were not deleted before restarting the Mitel Virtual
Appliance. The more snapshots there are, the more the performance will be degraded.
Resolution
Mitel Applications may cease to function if MSL switches the file system to "read-only" in an
effort to prevent file system corruption. Before concluding that a read-only file system is the
problem, confirm that MSL has switched to read-only by selecting the Mitel virtual machine,
and clicking Launch Virtual Machine Console. Observe the output, which may indicate that
the file system is now read-only. The console is the only place where you will be able to see
the read-only trigger, since logs cannot be written to disk in the read-only state.
Probable cause
Very high disk latency between the ESXi host and the virtual machine's data store over short
time period can trigger file system protection and the read-only state in MSL.
Resolution
To restore service immediately, right-click on the Mitel virtual machine, select Power and
Restart Guest. Use the Virtual Machine Console to confirm that system shut-down and start-up
occur before confirming that service is restored.
To prevent the problem from recurring, determine whether a SAN connection disruption
occurred, and examine Virtual Machine Disk Configuration and Statistics. Address any
failures or exceptions.
It is very important to make sure the system virtual disk can sustain the I/O throughput
requirement and the disk latency requirement. Because the physical storage is shared,
performance of storage could be momentarily reduced or degraded as the workload increases.
103
Mitel Virtual Appliance Deployment
Other external factors can also affect storage performance. For example, congestion in the
storage network induces network latency that, in turn, increases the I/O latency on storage
devices. In general, storage device average latency should not be more than 30 ms.
Performance of virtual machine will be affected as storage latency increases.
If SSD is used as part of the storage device, the maximum acceptable average latency should
be less. Disk I/O latency can be monitored by VMware vCenter Server. vCenter Server can
trigger actions (send notification to system administrators, for example) if it sees a degraded
or abnormal latency on storage devices. For example, vCenter server can be set up to send
warnings if disk latency exceeds 1000 ms for one minute, and send alerts if disk latency exceeds
3000 ms for one minute. The following procedure is required to enable Storage I/O monitoring
in VMware vCenter Server from vSphere Client.
Note: The threshold values are just examples. The administrator of the virtual
infrastructure must determine the actual values suitable within the environment. The
steps that follow are based on vSphere 5.5 and may be slightly different in other versions.
104
Known issues (VMware infrastructure)
System administrators should be notified when storage performance degradation begins. This
should allow enough time for them to correct the problem before storage performance falls to
an unacceptable level.
A Mitel Virtual Machine consumes CPU reaching near 100% continuously, and becomes
unresponsive. Memory consumption may also increase to near 100%.
Probable cause
The SAN volume hosting the Mitel virtual machine's vmdk is no longer available, possibly due
to network connectivity or SAN system failure.
Resolution
Confirm that the SAN volume is no longer accessible from the ESXi host running the Mitel
Virtual Machine. This should be reported next to the data store name in the vSphere client,
such as:
datastore1 (inactive)
or
datastore1 (inaccessible)
Shut down (Shut Down Guest) the Mitel virtual machine, if it is responsive. If it is not responsive,
plan and execute a restart of the ESXi host. Investigate and address the issue with the SAN
volume.
Large file transfers to or from a Mitel virtual machine may have very slow throughput, or
streaming media between multiple Mitel virtual machines may be poor quality or interrupted.
Probable cause
Large Receive Offload (LRO) may be enabled for the VMware vmxnet3 kernel module within
MSL. The vmxnet3 kernel module is a software driver used to support all VMxnet3 virtual
network cards used within a virtual machine.
Some Linux-based operating systems such as MSL do not support LRO. MSL will ignore LRO
packets and they will require re-transmission.
105
Mitel Virtual Appliance Deployment
Resolution
For up-to-date information on this issue, check the VMware knowledge base at
http://kb.vmware.com/; for example http://kb.vmware.com/kb/1027511.
Mitel virtual machines are blocked during power up due to checking file systems
unexpected inconsistency error
After deploying a Mitel virtual appliance and powering on the Mitel virtual machine created, the
operating system boot does not finish successfully. When examining the Virtual Machine
Console, the Checking filesystems command will indicate FAILED and report an error similar
to the following output will be seen:
Probable cause
The ESXi host that the Mitel virtual appliance was deployed to does not have the correct system
time set. The incorrect time is pushed from the ESXi host to the virtual machine on power up.
This time is earlier than the last one recorded in the MSL file system. This causes interactive
file maintenance to be performed, however, this requires a root password which is not yet set
if this is the first boot of the Mitel Virtual Machine.
Resolution
1. Shut down (Shut Down Guest) the Mitel virtual machine.
2. Change the system time of the ESXi server hosting the Mitel Virtual Machine to the current
time.
3. Power on the Mitel virtual machine and observe MSL boot-up in the Virtual Machine
Console.
106
Known issues (VMware infrastructure)
On Power On of a Mitel virtual machine that was shut down with a Power Off, startup delays,
file system repairs, or startup failures occur.
Probable cause
The Power Off request disrupted the running applications or system state in the Mitel virtual
machine. The virtual machine disk may be corrupted or in an inconsistent state
Resolution
Always use Shut Down Guest to turn off the virtual machine. This is a safe, graceful shutdown.
Do not use Power Off, since this effects an immediate shutdown, equivalent to sudden powering
off of a physical server.
Mitel resiliency solution is not exhibiting the correct resiliency behavior when an ESXi host or
its dependent resources become unavailable.
Probable cause
Mitel virtual machines configured for resiliency must reside on different ESXi hosts to ensure
that one ESXi-related outage does not affect all Mitel virtual machines in a resilient configuration.
Resolution
Ensure that Mitel virtual machines configured for resiliency never run on the same ESXi server.
If deploying a resilient solution to one ESXi cluster, ensure that Distributed Resources Scheduler
(DRS) anti-affinity rules are set up for the ESXi cluster to ensure that the Mitel Virtual Machines
are distributed among different ESXi hosts in the cluster. Also see “Anti-affinity rules” on page 34
and “Other virtual appliance management considerations” on page 34.
Mitel applications appear to be performing poorly with respect to message record and playback,
prompt playing, and so on, with problems such as poor quality voice record or playback,
message playback disruptions, or similar.
Probable cause
This can often be attributed to high disk latency or slow storage IOPs, or intermittent disk access
disruptions. This could include issues with the overall storage environment, or issues in
configuration of the VM's virtual disk.
Resolution
107
Mitel Virtual Appliance Deployment
Lazy Zeroed or Thick Provision Eager Zeroed. It should NOT be set to Thin Provisioned.
If Thin Provisioned is set, conversion to one of the Thick provisioning options is recom-
mended. Convert to a Thick Provision method by moving the virtual machine to another
data store:
a. Select Migrate, Change Datastore option.
b. Change Select a virtual disk format to Thick Provision Lazy Zeroed.
c. Select a different data store.
d. Complete the wizard to start the migration.
2. Select the Mitel virtual machine, open Snapshot Manager and ensure that there are no
snapshots. If there are snapshots listed, verify that all snapshots are no longer required;
then click Delete All to remove them.
3. Select the Mitel Virtual Machine, click the Performance tab, select the Advanced button
and investigate the Datastore and Disk statistics for latency and read/write throughout.
(Write Latency, Read Latency and/or Highest Latency).
4. Examine these values over the time period when issues were experienced. Values should
be consistently near 0 ms latency. Consistent latency responses above 30 ms or peak
values over 80 ms can result in many performance problems. Read/Write throughput should
meet or exceed requirements for the application if running at full load (see Table 8 on
page 21). If high latency or low throughput is observed, it is highly recommended that you
locate and migrate the virtual machine to a data store that is performing better.
Mitel logs files show missing interrupts, high clock skew, hardware watchdog warnings or similar.
Probable Cause
Host(s) may not meet minimum processor requirements or not properly configured for high
performance.
Resolution
Examine host processor, BIOS and related processor settings, to ensure they meet
requirements listed in Table 4, “Minimum hardware requirements,” on page 9. Correct any
settings or move to fully qualified hosts.
108
Appendix A
GLOSSARY
Mitel Virtual Appliance Deployment
110
Glossary of terms and acronyms
111
Mitel Virtual Appliance Deployment
112
Glossary of terms and acronyms
113
Mitel Virtual Appliance Deployment
114
Glossary of terms and acronyms
115
Mitel Virtual Appliance Deployment
116
Glossary of terms and acronyms
117
Mitel Virtual Appliance Deployment
118
Appendix B
VMWARE VSPHERE STORAGE SETUP
GUIDELINES
Mitel Virtual Appliance Deployment
120
VMware vSphere storage setup guidelines for Mitel virtual appliances
This chapter describes some ESXi storage options and recommends some storage
configurations that can be optimized for running Mitel virtual appliances on ESXi hosts. Various
options and considerations are discussed first, with recommendations to follow. Other practices
and more advanced areas specific to VMware and real-time, latency sensitive workloads are
also mentioned for considerations.
Note: This chapter provides information about VMware and Hyper-V applications and
requirements that were added for previous releases, and which may not be up-to-date.
Always refer to your current VMware or Hyper-V documentation for the latest information.
Storage virtualization
ESXi hosts provide storage virtualization (also called virtual disk), which logically abstracts the
physical storage layer from virtual machines. ESXi virtual machines use virtual disks the same
way a physical machine would use a local disk drive. However, a virtual disk can be mapped
to one large physical file, a physical disk, or a set of files that are located either locally on the
ESXi host or remotely.
A virtual machine uses a virtual SCSI controller to access virtual disks. The virtual SCSI
controllers can be any of the following:
• BusLogic Parallel
• LSI Logic Parallel
• LSI Logic SAS
• VMware Paravirtual
The latest Windows operating systems use LSI Logic SAS by default, and most of the advanced
Linux operating systems use LSI Logic Parallel by default. Whether the storage devices are
being accessed through parallel SCSI, iSCSI, network, Fiber Channel, or FCoE, they are
transparent to the guest operating system and applications running on the virtual machine.
121
Mitel Virtual Appliance Deployment
Traditional hard disk drives use rotating mechanical disks as electromechanical devices to store
data. Since it takes time to move a read/write head to the specific region on a rotating disc to
access information, access time is slower when compared to a solid state disk drive. Hard disk
drives have moving parts: the rotating disk and the moving read/write head. The overall average
access latency is affected by the speed of spindle rotation and the read/write head access time.
The data interface is one of SATA, SAS, or Fiber Channel.
Solid State disks use semiconductors as the storage medium rather than rotating disks. Since
there are no moving parts, SSDs provide much faster access to data than conventional hard
disk drives. Like mechanical hard disk drives, SSDs come with different interfaces: SATA, SAS,
and Fiber Channel. One of the disadvantages of using SSD is its shorter service life. Storage
cells inside a SSD do not survive as many write cycles as mechanical disks do. Operators must
carefully monitor how intensively the SSD device is utilized to calculate its estimated lifetime.
Types of storage
There are two major types of physical storage.
• Local Storage: stores information on disks that are directly connected to the host, either
internally or externally.
• Networked Storage: stores information on external disks or arrays that attached to the
host through a high-speed network.
Local storage
Local storage is directly connected to ESXi hosts, either internally or externally. It can be
accessed by only one host and cannot be shared across multiple hosts. Storage access is fast
because it is accessible only by the attached host. It also creates a single point of failure because
storage access is impossible when the host fails.
Networked storage
Networked Storage consists of external storage systems connected to ESXi hosts through a
high-speed storage network. The networked storage can be shared among ESXi hosts; it can
be accessed by multiple hosts concurrently, and is therefore eliminated as a single point of
failure for local storage.
Networked Storage can be implemented by using the following network storage technologies:
• Fiber Channel (FC)
• iSCSI
• NFS
• Shared Serial Attached SCSI (SAS)
• Virtual SAN
122
VMware vSphere storage setup guidelines for Mitel virtual appliances
FC SAN is a very high performance type of storage device. The ESXi host must be equipped
with Fiber Channel host bus adapters (HBA) and Fiber Channel switches to connect to the FC
SAN. It uses Fiber Channel protocol and a Fiber Channel network to transport SCSI traffic.
A less expensive alternative is to use FCoE (Fiber Channel over Ethernet) adapters on an
Ethernet network as the physical network to transport SCSI traffic using Fiber Channel protocol.
iSCSI
iSCSI packages SCSI traffic into the TCP/IP protocol and it travels through a standard TCP/IP
network instead of the specialized FC network. ESXi host serves as iSCSI initiator that
communicates with iSCSI targets in remote iSCSI storage systems. ESXi host s two types of
iSCSI connections:
• Hardware iSCSI: available through a third-party adapter capable of off-loading iSCSI and
network processing.
• Software iSCSI: a software-based iSCSI initiator provided by VMware. It uses VMkernel
to connect to the storage without any specialized third-party adapter.
NFS
ESXi hosts can use Network File System (NFS) version 3 to communicate with NAS or NFS
servers for accessing virtual disks. While SCSI is a block-level type of access, NFS is a file-level
type of access. In theory, the performance is lower when compared to block-based access
types.
Shared SAS is a newer technology that depends on hardware vendors. It has the advantage
of cost, performance, and simplicity over other storage technologies in localized environments.
Virtual SAN
Virtual SAN is a distributed layer of software that runs natively as part of the ESXi hypervisor.
Virtual SAN aggregates local or direct-attached disks of a host cluster and creates a single
storage pool that is shared across all hosts in the cluster. Virtual SAN offers some of the benefits
of SAN without the need to purchase a real SAN.
Mitel virtual appliances support deployment on Virtual SAN. However, Virtual SAN has its own
limitations and requires additional effort to setup and configure. Consult the VMware
documentation to make sure that Virtual SAN will meet your needs before deploying it.
RAID levels
RAID is a data storage virtualization technology that combines multiple disk drives into a single
logical unit for the purposes of data redundancy and/or performance improvement. Data can
be distributed across the drives in one of several ways, referred to as RAID levels. Each RAID
level provides a different balance between reliability, availability, performance and capacity.
Each RAID level inherits a specific write penalty. The write penalty is a measure of the extra
123
Mitel Virtual Appliance Deployment
waiting time for parity info or redundant info to be written onto the array of disks. A list of some
popular RAID configurations is shown in the following sections.
RAID 0
RAID 0 provides no data redundancy or fault tolerance. It improves performance through the
use of parallel read and write operations across multiple disk drives. RAID 0 has no error
detection mechanism, so failure of one disk causes data loss on the array. Since there is no
parity information to calculate, and only the data is written onto the disk, this type of RAID array
has a write penalty of 1.
RAID 1
RAID 1 uses mirroring. Data is written identically to two or more drives, producing a mirrored
set. Each read request is served by the disk with the lowest seek latency and rotational latency.
Write performance is degraded, compared to RAID 0 because all drives being mirrored must
be updated. However, RAID 1 provides redundancy and the disk array continues to operate as
long as at least one drive is functioning. Because data is written to two disks (requires two write
operations), RAID 1 has a write penalty of 2.
RAID 10
RAID 10 is a nested RAID level of RAID 1 and RAID 0. It creates a striped set from a series of
mirrored drives. The array can sustain multiple drive losses, as long as no mirror loses all of
its drives. Since RAID 0 carries a write penalty of 1 (no penalty), RAID 10 has the same write
penalty as RAID 1, which is 2.
RAID 5
RAID 5 uses block-level striping with distributed parity. Parity information is distributed among
the drives. It requires all drives but one to be present to operate. Upon a single drive failure
subsequent reads can be calculated from the distributed parity. RAID 5 requires a minimum of
three disks, and it is seriously affected by array rebuild time and the chance of failure during
rebuild. RAID 5 has a write penalty of 4.
RAID 6
RAID 6 uses block-level striping with double distributed parity. Double parity provides fault
tolerance for up to two failed drives. RAID 6 has a write penalty of 6.
Independent persistent
In this mode, changes are written and remain on the disk (persistent), providing the best
performance. However, snapshots cannot be taken on an independent persistent disk. Since
most backup software needs to create a snapshot of a virtual disk to do backups, operators
need to handle backups on independent persistent disks differently.
124
VMware vSphere storage setup guidelines for Mitel virtual appliances
Independent non-persistent
In this mode, changes are appended to a redo log. The redo log is erased when the virtual
machine is powered off or reverts to a snapshot, causing changes to be discarded. Because
the redo log tracks the changes in the file system of the virtual machine, performance is not as
high as for disks in independent persistent mode.
Dependent
In this mode, changes are appended to a redo log that persists between power cycles. Similar
to Independent non-persistent mode, disk performance of dependent disks are lower than for
independent persistent mode. This is the Mitel default.
Thick-provisioned
Thick provisioned disks have all their space allocated at creation time. Because the space is
pre-allocated it takes more time to create. There are two types of thick-provisioned disk.
Eager-zeroed
Eager-zeroed disks have all the space pre-allocated and zeroed out at the time of creation. It
takes more time to create the disk, but the pre-zeroing action results in the best disk
performance.
Lazy-zeroed
Lazy-zeroed disks have all the space pre-allocated at creation time, but blocks are zeroed only
on first write. It reduces creation time but reduces performance the first time a block is written.
Thin-provisioned
Thin-provisioned disks have space allocated and zeroed upon the first write, as opposed to
upon creation. For the first write on any unwritten file block this method carries a higher I/O
cost. However, subsequent writes on the same blocks have the same performance as
eager-zeroed thick disks.
Partition alignment
Earlier versions of ESX/ESXi create VMFS3 partitions with 64 KB alignment. The 64 KB
alignment is retained even after the file system is upgraded to VMFS5. The vSphere Client
automatically aligns VMFS3 and VMFS5 partitions along the 1 MB boundary at creation. To
obtain the better 1 MB alignment, a 64 KB aligned partition must be deleted and recreated using
vSphere Client and an ESXi 5.0 or later host.
125
Mitel Virtual Appliance Deployment
SAN multipathing
SAN Multipathing is a fault-tolerance and performance enhancement technique. There are
multiple physical and logical paths between the host and the storage device through different
controllers and switches.
Paths are accessed based on the configured path policies. SAN path policies can have a
significant effect on storage performance. For most Active/Passive storage arrays, ESXi uses
the Most Recently Used (MRU) path policy by default. Fixed path policy is not recommended
for Active/Passive arrays because it can cause frequent and rapid path switching resulting in
slow LUN access.
For most Active/Active storage arrays, ESXi defaults to use of the Fixed path policy. For optimal
performance, the Round Robin path policy can be considered.
Round Robin path policy can improve storage performance in some environments by cycling
I/O requests through all active paths. However, not all storage arrays support the Round Robin
path policy. Check the storage array documentation or the storage vendor to make sure the
Round Robin path policy is supported and recommended for your storage array and
configuration.
File systems
VMware ESXi host s two different file systems for virtual machines: VMFS and NFS.
VMFS
VMFS is the VMware native high-performance clustered file system, supported either locally
or network-attached.
NFS
The NFS distributed file system has been in use for nearly 20 years. NFS is strictly
network-attached and uses Remote Procedure Calls to access remote files.
Raw Disk
VMware vSphere supports storing virtual machine guest files on a raw disk. Raw device
mapping (RDM) provides a mechanism for a virtual machine to have direct access to a LUN
on the physical storage subsystem (iSCSI or Fiber Channel only). RDM is a mapping file in a
separate VMFS volume that acts as a raw physical storage device. It has some benefits, but it
also has limitations, and is not suitable in every situation. Consult the VMware documentation
to thoroughly understand Raw Device Mapping before using it.
Recommendations
Storage network
Either iSCSI or NFS would be suitable. iSCSI is a block-based protocol operating at a lower
level, while NFS is a file-based protocol operating at a higher level. However, iSCSI has the
126
VMware vSphere storage setup guidelines for Mitel virtual appliances
advantage of being slightly more efficient. The following implementation is recommended for
both iSCSI and NFS.
• For better iSCSI performance, enable jumbo frames when possible. Jumbo frames reduce
packet-processing overhead, thus improving CPU efficiency of storage I/O. Use of jumbo
frames requires end-to-end hardware support from iSCSI adapters, storage arrays, and
hardware network switches. The default Ethernet frame size is 1500.
• It is recommend that both iSCSI and NFS have their own dedicated physical network to
minimize network interference from other packet sources. If this is not possible, logical
separation using VLAN can be used.
RAID level
RAID 10 provides the most balanced performance, redundancy, cost, and write penalty, so it
is the recommended RAID level. Other RAID levels may still be used to build virtual disks to
better fit specific workload requirements.
Mitel supports the default Dependent virtual disk mode. Dependent disk mode allows temporary
disk snapshot, which Mitel s during backup and upgrade. Virtual disks of Mitel virtual appliances
must not contain snapshots during normal operation.
Mitel applications are real-time applications and require high system performance during
operation. It is preferred that all disk spaces be allocated during system installation so that they
are ready to roll during operation. Therefore, the thin-provisioned disk type is not recommended.
Thick Provisioned with lazy-zeroed virtual disk type (the default) is recommended because it
has a more balanced disk performance and OVA deployment time.
Partition alignment
Use the latest version of VMware software to perform disk partitioning and formatting, for
optimum disk drives preparation.
SAN multipathing
File system
In general, VMware VMFS is recommended for use as the underlying file system because
VFMS is native to VMware and it is designed specifically for the VMware environment.
127
Mitel Virtual Appliance Deployment
Other considerations
CPU utilization
Since a higher level of CPU resources is needed to process higher levels of networking traffic,
it is important to ensure sufficient CPU resources are available, without limiting the maximum
network throughput. It is important to monitor CPU utilization of high-throughput workloads to
ensure that the CPU resource is sufficient.
If a physical NIC is shared by multiple virtual machines, each virtual machine could affect the
performance of other virtual machines. For best network performance, use separate physical
NICs for heavy networking I/O virtual machines and latency-sensitive virtual machines. 10 Gbps
NICs have multi-queue support, so this recommendation typically does not apply, as long as a
sufficient number of queues is available.
Jumbo frames
ESXi supports the use of Jumbo Frames with iSCSI. ESXi allows Jumbo Frames with MTU
size up to 9000 bytes. When a Jumbo Frame is used for iSCSI traffic, the following must be true:
• The network must support Jumbo Frames of that MTU size end-to-end to be effective.
• Network equipment and storage vendors must be consulted to ensure that the physical
NICs, iSCSI HBAs, and network switches support Jumbo Frames from end to end.
• Network switches have been set up and verified for Jumbo Frames, to ensure proper
operation on the network.
Flash Read Cache accelerates virtual machine performance through the use of host-resident
flash-based storage devices as a cache. VMware Virtual Flash configured as a cache can be
used for virtual machine migration, read caching, and write caching. VMware Distributed
Resource Scheduling (DRS), and vSphere High Availability support Virtual Flash.
The functionality of hardware acceleration enables an ESXi host to integrate compliant storage
arrays and off-load specific virtual machine and storage management operations to the storage
hardware. Off-loading tasks to storage hardware means that the storage consumes less CPU,
memory and storage fabric bandwidth on the ESXi hosts. Storage hardware acceleration works
only when appropriate hosts and storage arrays are combined. Verify with the hardware vendors
to make sure that storage hardware acceleration is supported.
VMware-specific optimizations
When using ESXi with a SAN, use the following tips to avoid SAN problems.
• Put only one VMFS data store on each LUN. Multiple VMFS data store on one single LUN
is a configuration that is not recommended.
128
VMware vSphere storage setup guidelines for Mitel virtual appliances
• Do not change the path policy the system sets unless the implications of the change are
fully understood.
• Ensure there is no single point of failure from the SAN to the host HBAs.
• I/O device latency should not be consistently greater than 20-30 ms (for more information,
refer to this VMware vSphere Blog post:
http://blogs.vmware.com/vsphere/2012/06/troubleshooting-storage-perfor-
mance-in-vsphere-part-2.html).
• Tune the virtual network adapter (VMXNET3). Some of the default parameters may not suit
your environment (retries and time out values, for example). In that case, fine tuning of
parameters may be required.
• If the virtual machine is using an LSI Logic vHBA, the reqCallThreshold value could be
adjusted.
The lower the reqCallThreshold value, the lower the time the I/O requests are likely to stay
in the vHBA queue. If reqCallThreshold is set to 1, it means the I/O will be dispatched to
the lower layer even there is only one I/O request in the LSI Logic vHBA queue. The default
reqCallThreshold value is 8, and it is a system-wide configuration on a ESXi host.
VMware vSphere employs the vSphere Pluggable Storage Architecture (PSA) in its storage
subsystem. PSA is essentially a collection of plugins inside the VMkernel layer of the ESXi
host. The top-level plugin in the PSA is the Multipathing Plugin (MPP). MPP defines how
vSphere manages and accesses storage, including load balancing, path selection, and fail over.
VMware provides the Native Multipathing Plugin (NMP) as its MPP, but storage vendors may
also provide their own MPP implementations.
Within each MPP, including VMware NMP are two sub-level plugins: the Storage Array Type
Plugin (SATP) and the Path Selection Plugin (PSP). SATP handles path fail over and PSP
handles load balancing.
Depending on the type of storage arrays used, the default PSP policy could be:
• Fixed: The default policy for most active-active arrays.
• Most Recently Used: The default policy for most active-passive arrays.
• Round Robin: The default policy for some active-active and active-passive arrays.
For best performance, check with your storage vendor to obtain the appropriate plugins, if
available.
Reference
VMware Publication: vSphere Storage (for ESXi 5.5 and vCenter Server 5.5)
http://pubs.vmware.com/vsphere-55/topic/com.vmware.ICbase/PDF/vsphere-esxi-vcenter-se
rver-55-storage-guide.pdf
129
Mitel Virtual Appliance Deployment
130
Appendix C
VMWARE VCLOUD DIRECTOR
Mitel Virtual Appliance Deployment
vCloud Director
Note: This chapter provides information about VMware applications and requirements
that were valid for previous releases, and which may not be up-to-date. Always refer to
your current VMware documentation for the latest information.
This chapter describes the general behavior of all Mitel virtual applications when deployed in
a VMware vCloud Director environment (vCD). This includes deployment methods, runtime
behavior, and networking-related considerations.
vCloud Director requires vSphere (ESXi + vCenter Server), and vCloud Networking and Security
(vCNS) as an underpinning, at a compatible version level.
Note: vCD 5.1 and above only, are compatible for use with Mitel virtual appliances.
132
vCloud Director
vCloud Director operates as an administration layer on top of VMware vSphere, using vSphere
resources, and providing an abstracted view of vSphere resources to users. vCloud Director
uses VMware vCNS to create secured organization networks and other networking constructs;
this is largely invisible to the user.
Table 28 below outlines key concepts for understanding how vCloud Director operates.
133
Mitel Virtual Appliance Deployment
Only information specific to Mitel virtual appliances is described in this guide; for full information
and instructions, refer to the VMware documentation. For specifics related to using the vCloud
Director Web Admin interface, refer to the VMware vCloud Director User Guide.
This section describes how to create and publish Mitel virtual appliances as vCloud Director
Catalog Template items from OVA packages. vCloud Director releases before 5.5 do not directly
support deployment of VAs or vApps packaged in OVA format, which is Mitel standard packaging
and distribution format.
If you are using vCloud Director 5.1 or earlier, which support uploading of VA and vApps only
when packaged in OVF format, (the unpacked file hierarchy making up the VA or vApp
definition), then you must use the following procedures. The OVA is essentially a TAR-packed,
one-file packaging of the OVF directory tree, so Mitel OVAs must first be unpacked to OVF, or
otherwise prepared to be uploaded to vCloud Director Catalogs. Multiple methods are available.
Note: Catalog items can also be created as ISO images in vCloud Director. However,
Mitel VAs and vApps do not need this capability, since they are packaged as OVAs.
While it is possible, deploying from an ISO is not recommended practice for
OVA-packaged applications, and is not covered in this guide.
The following subsections provide descriptions of three different procedures to create Catalog
items from OVA distribution files. The procedures described here are:
• “Unpack and publish OVA in Linux” on page 135
• “Unpack and publish OVA in Microsoft® Windows®” on page 135
• “Import VM from vSphere” on page 137
134
vCloud Director
If you are using Linux desktop-based administration, the Linux desktop may be the best way
to unpack and publish the OVA. Both command line and GUI options are described. The GUI
method described assumes CentOS/Red Hat.
The Linux method is preferred to the others discussed in this section because it may be
somewhat easier to perform than the Windows desktop method.
To unpack and publish from OVA format file in a Linux environment, use the following steps.
3. Create a destination directory for the unpacked OVF. 3. In the Archive Manager:
# mkdir app-ovf • Click Extract.
• In the folder drop down menu, select Extract >
Other > New Folder.
• Enter a name for the destination folder, app-ovf, for
example.
• Click Extract.
4. Unpack the OVA using the TAR command. 4. Close the Archive Manager
# tar -xf app.ova --directory app-ovf/
The files are unpacked into the specified destination
directory.
Upload the OVF to the Catalog
1. Open the vCloud Director Web Admin for the Organization as Organization Administrator or Catalog Author.
2. Open the target Catalog:
3. Click the Upload icon.
4. Browse to the OVF folder you created in the previous step, and select app.ovf., or the name of the destination
folder you created.
5. Click Upload.
6. Certificate exceptions may pop up during progress. Click Accept on each one.
7. Confirm progress and final success.
If you are using mostly Windows desktop-based administration, the Windows desktop approach
may be the best way to unpack and publish the OVA.
135
Mitel Virtual Appliance Deployment
This approach involves extra steps, compared to the Linux method, and the use of additional
tools, so the Linux method is preferable, if it is available.
As an alternative to the above procedure, VMware also provides a tool called OVF Tool, which
has a command line interface. Since this tool is more difficult to install and use, is considerably
slower, and can alter file names as provided in the original OVA packaging, it is not generally
136
vCloud Director
recommended. However, this tool may be useful where scripting is used to automate OVF
conversion and importing.
This method requires that you have both vCloud Director system administrator level access
and vCenter administrator access. This cannot be used by normal Organization level
administrators Organization users.
This method may be most appropriate if you wish to set up and publish master Catalogs for
your customer Organizations to use.
To accomplish this, you can create a separate special Organization specifically for the purpose
of importing and publishing the OVA-based VAs and vApps with a Public Catalog. Users of this
special Organization are given system-level administrator access to vCloud Director and to
vCenter, allowing them to do the procedures required to populate the Catalog. The special
Organization's Public Catalog can then be accessed by normal customer Organizations for the
purpose of deploying the VAs and vApps.
2. Create a new, separate resource pool and/or a folder at the vSphere level to manage the
new VMs you will deploy. This step is optional, but recommended to make organizing your
VMs easier.
3. Deploy the application OVA as a VM in vSphere vCenter using normal VMware procedures.
Leave the VM powered off in vSphere.
Note: You must select valid network connections for each vNIC on the VM in vSphere.
If you do not, the Import operation (described in the next procedure) will fail in vCloud
Director. The connections you enter do not have to be valid for vCloud Director use, but
they must be valid in vSphere.
Note: You must be logged in as vCloud Director System Administrator. Normally, vCloud
Director Organization-level administrators do not have such access.
2. Navigate to the target Organization and the specific public or private Catalog to import the
VA or vApp into. Select the Catalog.
137
Mitel Virtual Appliance Deployment
OR: Click the Actions drop-down menu and select Import from vSphere.
OR: Right-click in the Template list and select Import from vSphere.
Note: It is recommended that you not use Copy VM. This is the default setting; Copy
VM copies the VM to vCloud Director space, leaving the original in place in vSphere.
This can waste space and be more difficult to manage.
9. Optional: Select Gold Master. See “General settings for Catalog Template items” on
page 138 for details.
10. Click OK to exit the Import dialog box, and start the import operation.
11. In the Catalog List view, watch the progress of the import operation and confirm that it
finishes successfully. This process can be quite long, depending on the size of the VM (or
OVA), network performance, and the total load on the hosts involved.
After Catalog Template items are uploaded or imported, and after they appear in the vApp
Templates List view, you can modify the Template settings.
To open a template, select the template. Right-click and select Properties, or select Properties
from the Actions drop-down menu.
138
vCloud Director
• Set VM startup order. This is not applicable to Mitel VAs, and other single VM virtual
appliances.
Note: If you are installing multiple Mitel vApps, it is strongly recommended that you
deploy them one at a time. This will avoid having the guest properties for all of the vApps
displayed at the same time, which can cause confusion in completing the deployment.
Note: You cannot modify the runtime and storage lease duration for the VA at this stage,
but you can modify and renew the lease at a later time. See “vCloud Director Lease and
Lease Expiry” on page 151.
c. Configure the network mappings: For each network interface on the VM, select the
network to connect the interface to, and the IP Allocation type, For Mitel VAs, select
Static - Manual IP Allocation. Static - Manual is the only IP Allocation type currently
available for Mitel VAs.
For VAs with more than one vNIC, the network and IP Assignment must be done for
each vNIC separately.
Note: For IP Assignment type, three options are available. Each of these enable address
and related configuration to be pushed into the VA when it is deployed. Which one to
use depends on desired outcome, and compatibility of the specific VA being deployed.
However, current release Mitel VAs do not support either IP Pool or DHCP-based auto
configuration of addresses. Always select Static – Manual allocation.
d. Configure Properties. Not all Mitel product releases support the modification of custom
Properties; these Mitel VAs will display the message, “There are no user configurable
properties.”
When the custom properties screen appears, fill in all fields, For more information about
the Custom properties for each Mitel VA, see “Deploying Mitel OVAs” on page 64.
139
Mitel Virtual Appliance Deployment
Note: Configure Networks and Custom Properties settings are not available for use
with current MSL-based VAs.
e. The Ready to Complete screen appears. Check the settings displayed to ensure that
they are all correct, and click Finish.
5. Navigate to My Cloud > vApps view to watch progress and ensure success.
One important difference between vCloud Director-based deployment and deployment directly
on vSphere is that the Configuration Options dialog is not presented in vCloud Director, as it is
in vSphere. Therefore the administrator cannot choose the specific deployment profile to use;
Small Business or Enterprise, for example. In vCloud Director, the vApp is deployed in its default
configuration.
While this issue is created when the Virtual Appliance is uploaded or imported to a vApp
Template in the Catalog, it manifests only when the Template is deployed.
Unless the vApp is appropriately modified to correspond to the desired deployment profile, its
resourcing may be incorrect; it may be too small for the desired capacity, or unnecessarily large.
Refer to section “Modifying vApp Resourcing for Deployment Configuration” on page 146 for
details on how to work around this issue.
Post-deployment configuration
For current Mitel VAs other than MiVoice Business Express, after the Mitel VA has been created,
the VA is then configured as for deployment of the VA under vSphere vCenter. In the case of
MiVoice Business Express, initial configuration is performed using the Mitel Initial Configuration
Wizard at deployment time (see “Deploying and configuring virtual applications from catalogs”
on page 139).
This procedure varies, depending on Mitel VA specifics; the basic steps are as follows:
1. Right-click the VA name, and select Start.
2. Double-click the Console icon to open the console on the VA.
140
vCloud Director
3. Configure the administrator password, the host name, the IP address, IP mask, and DNS.
4. Select the network interfaces, and so on, as required.
Note: With the exception of MiVoice Business Express, Mitel VAs are not pre-configured
with IP address as set in the address auto-configuration step above. The MiVoice
Business Express Initial Configuration Wizard guides you through the initial system
setup at the time you first login to the MiVoice Business Express web administration.
Refer to MiVoice Business Express documentation.
There are, however, large potential differences in the way resources are allocated to the
run-time VAs in vCloud Director environments. Many Mitel VAs have CPU and memory
reservations built into their OVA packaging. vCloud Director makes modifications to internal
resource reservations of the underlying OVA/OVF when the VA is deployed. vCloud Director
does not modify other aspects, such as disk size and number of vCPUs, for example.
In the vCloud Director environment, the allocation model assigned to the Organization vDC
into which virtual appliances are being deployed affects how resource reservations built into
the virtual appliance OVA are applied. Table 30 below summarizes these differences.
141
Mitel Virtual Appliance Deployment
Notes:
1. VA # vCPU - number of vCPUs assigned to the VA in OVA/OVF
2. VA memory - amount of memory (memory size) assigned to VA in OVA/OVF
3. VA CPU reservation - CPU resource reservation (GHz) applied to VA in OVA/OVF
4. VA memory reservation - memory resource reservation (GB) applied to VA in OVA/OVF
5. VA limit - resource limit (CPU or memory) applied to the VA in OVA/OVF
6. VA shares - resource shares (CPU or memory) applied to VA in OVA/OVF
7. vCPU speed - vCPU speed value assigned to Pay-As-You-Go, as set in Organization vDC allocation
policies (see following subsections)
8. Guarantee % - CPU or Memory resource guarantee value, as set in the Organization vDC allocation
Note: In early testing, Mitel found discrepancies in the way Allocation Pool reservations
are set, relative to the formulas in VMware documentation. It is important to check actual
applied values in the customer vCloud Director environment by inspecting actual
reservations of the deployed VMs at vSphere level while the VM is running.
In vCloud Director, each allocation model allows different controls for how resource over-commit
is managed. For example, as can be seen in Table 30 above, both Pay-As-You-Go and
Allocation Pool allocation models provide a “guarantee percent” control, that sets the
reservation at run time, based on adjustment of the virtual appliances built-in reservations.
Even these two models apply the allocation percent control somewhat differently, however.
Only the Reservation Pool allocation model fully preserves the virtual appliances reservations.
Further details and discussion about each allocation model are provided in the following
sub-sections.
The Pay-As-You-Go allocation model is not recommended for real-time voice applications that
provide their own strict built-in reservations, since it cannot provide the required CPU
reservation guarantees.
Pay-As-You-Go allocation, with suitable tuning, could be a good choice in some circumstances,
such as large hosting providers, or try-and-buy agreements.
As shown in Table 30, this model effectively ignores any CPU GHz reservation built into the
OVA file for the VA, instead basing reserved resources only on the number of vCPUs and the
vCPU speed, as modified by the percent guarantee. Similarly, memory reservation is based on
the VA's built-in memory size, as modified by the percent guarantee, and ignoring VA built-in
reservations. Both CPU and memory shares are always forced to Normal, independent of any
built-in VA setting.
142
vCloud Director
Recommendations:
• The Pay-As-You-Go allocation model is not recommended for use with real-time-sensitive
UCC applications (MiVoice Business Virtual, MBG Virtual, MiCollab Virtual), and is not
recommended for deployments in which voice quality guarantees are part of the SLA.
• If you are using the Pay-As-You-Go allocation model, customers with a strict SLA can be
provided with a different Organization vDC, specifically for their real time applications, using
a different allocation model (Reservation Pool, for example). This is the preferred approach.
• The Pay-As-You-Go allocation model can be a viable choice in larger hosting providers,
where there are plenty of resources, overall, and there are large numbers of Organizations
using the same allocation approach.
• As guaranteed percent values approach 100%, this model becomes increasingly close to
the Reservation Pool model, with the exception that it is based on set vCPU speed in the
Organization vDC (rather than the VA's CPU reservations), and on the VA's set memory
size (rather than memory reservations).
• If voice quality is part of the SLA provided to the Organization, then allocation guaranteed
percent values should be set relatively high for the Organization vDC, especially when the
CPU reservation is to be used for real time voice.
• If voice quality is not an important part of the SLA provided to the Organization, then
allocation guaranteed percent values can be set somewhat lower for the CPU. However,
memory percent guarantees should still be set fairly high, to avoid slow response time.
For detailed information on configuring the Pay-As-You-Go allocation model, refer to VMware
documentation.
The Allocation Pool allocation model is not available for real-time voice applications that provide
their own strict built-in reservations, since it cannot provide the required CPU or memory
reservation guarantees.
Allocation Pool allocation, with adequate and guaranteed resourcing at the Organization vDC
level, can be a good choice in some circumstances, such as small systems running alongside
other low-demand workloads.
In the Allocation Pool model, resources are committed to the Organization vDC (drawn from
the overall Provider vDC) when the Organization vDC is created. In effect, this allows the
customer Organization to control their own over-commit levels. If the CPU guaranteed percent
143
Mitel Virtual Appliance Deployment
value remains set at 0 (the default) larger numbers of virtual appliances and vApps may be run
simultaneously, with the effect that they all slow down, but they are still able to run.
VAs with built-in reservations are then modified when deployed, as shown in Table 30, and
draw their resources from the overall Organization pool. This effectively reduces the Mitel built
in reservations, which may affect voice quality.
Recommendations:
• The Allocation Pool allocation model is not recommended for use with real-time-sensitive
UCC applications (MiVoice Business Virtual, MBG Virtual, MiCollab Virtual, etc), and is not
recommended for deployments were voice quality guarantees are part of the SLA.
• Alternatively, the customer with the strict SLA could be provided with a different Organization
vDC, specifically for their real time applications, using a different allocation model (Reser-
vation Pool, for example). This is the preferred approach.
• The Allocation Pool allocation model can be a good choice, under some circumstances,
especially when the customer Organization can afford lots of overall resourcing for their
Organization vDC.
• In this model, it would primarily be up to the customer Organization to ensure they provide
enough resources to ensure voice quality. This, in effect, provides customer-controlled SLA.
• If it is known that the customer will be deploying voice or other real-time sensitive virtual
appliances or vApps, the hosting provider should set CPU guaranteed percent values
relatively high. This will ensure that the Organization vDC overall gets adequate CPU and
memory resources for their needs, including the real-time applications, even when there is
high overall load across the host's Provider vDC.
For detailed information about configuring the Allocation Pool allocation model, refer to VMware
documentation.
The Reservation Pool allocation model is the only allocation model that is fully recommended
for real-time voice applications that provide their own strict built-in reservations (like Mitel voice
applications).
Where the Organization is using a different model for general applications, you can configure
a separate Organization vDC using the Reservation Pool allocation model to run only the
real-time applications.
144
vCloud Director
• Allow resources to be allocated and guaranteed at the Organization vDC level, as well as
at the individual VA level, with individual VAs drawing on resources from the Organization
vDC's overall reservation as they are run.
• Reservation Pool is the only allocation model that allows Administrators (at either Organi-
zation or System level) to access and modify resource reservations on VMs; the other
allocation models do not allow this.
In the Reservation Pool model, resources are fully committed to the Organization vDC (drawn
from the overall Provider vDC) when the Organization vDC is created. Built in OVA reservations
are not modified, as shown in Table 30. This allows the customer Organization to control their
own over-commit levels, within their vDC. If resources are over-committed (that is, total
reservations becomes higher than the total reservation for the Organization vDC), then VAs
will fail to deploy.
Recommendations:
• The Reservation Pool allocation model is recommended for Mitel VAs, which have built in
resource reservations.
• The Reservation Pool allocation model is ideal for real-time applications, where both CPU
and memory reservations are important to guarantee voice quality SLAs. This comes at
expense, however, of zero over-commit available to the provider.
• In this model, it may be primarily up to the customer Organization to ensure that they provide
enough resources to ensure voice quality. In effect, SLA is customer-controlled, using
reservations built into the VAs as they are deployed.
• Alternatively, if the hosting provider is also supplying the VAs the customer Organization is
deploying, (they could be selected from a Catalog provided by the host Organization, for
example), then specific reservations can be set in advance for customer use. This allows
flexibility for setting strict reservations where important, and not setting them where it is not
important.
• This allocation policy can be an ideal solution as a voice (or other real-time) application-spe-
cific Organization vDC, to supplement other vDCs used by the Organization for less critical
needs. For example, the customer can run non-critical applications in a vDC using Allocation
Pool or Pay-As-You-Go, and run only the real-time application in a Reservation Pool vDC.
For detailed information on configuring the Reservation Pool allocation model, refer to VMware
documentation.
Aside from the allocation model, other underlying resources used to create the Organization
vDC can also have large effects on the performance of the deployed virtual appliances.
Storage
145
Mitel Virtual Appliance Deployment
vCloud Director 5.1 adds the concept of “storage profiles”. These must be configured at the
vSphere level for vCloud Director, which uses them to automatically select storage for vDCs,
based on the specified profile.
The names you assign to the storage should reflect its performance and/or level of protection.
For example, a high performance store, with full network RAID 10 and replication protection
might be labeled “Platinum”, the same but with no replication might be “Gold”, mid-performance
with RAID 10 protection, “Silver”, and low performance with no RAID 10, “Bronze”.
Note: Due to vMotion requirements, storage for vCloud Director must be shared storage
(SAN, vSAN, NFS, VMware VSA), and must be available on all hosts supporting the
vCloud Director-provided vDC cluster(s). Host-local storage cannot be used.
Network
The same performance considerations apply within vCloud Director as in the underlying
vSphere layer.
Note that vCloud Director requires the use of virtual distributed switches (dvSwitch) in the
underlying vSphere networks. Regular vSwitches are not compatible.
As noted in “Deployment differences relative to vSphere” on page 140, when vApps are
deployed on vCloud Director, the Configuration Options dialog box is not presented as it is
in direct vSphere deployments. Therefore, the administrator cannot choose the specific
deployment profile to use (Small Business or Enterprise, for example), and resourcing for the
vApp may be incorrect for the desired capacity (too small, or unnecessarily large).
Note: Methods described here can only be fully implemented when the VA or VA
Template (Catalog item) are deployed in an Organization vDC configured to use the
Reservation Pool allocation model.
To modify the VA resourcing to match the appropriate profile, two methods are available, as
follows:
In this approach, the VA is brought into a vApp Template. Then it is deployed in a Reservation
Pool based vDC and modified. Finally, the VA is added back into the Catalog as a new vApp
Template. This may be done by either Organization or System level administrators.
The Templates resourcing can now be inspected (but not modified), by selecting and
opening the vApp, selecting the desired contained VM, and opening the Properties dialog
146
vCloud Director
box on the VM. The Hardware tab shows the number of vCPUs, memory, and disk
assignments; the Resource Allocation tab shows CPU and memory reservations, shares,
and limits.
Note: The Resource Allocation tab is available only when the vApp Template is stored
on a Reservation Pool vDC.
2. Deploy the vApp Template to My Cloud, as described earlier. See “Deploying and config-
uring virtual applications from catalogs” on page 139.
Note: IMPORTANT: During deployment, at the Network Mapping step, the VM's built-in
network interface Source name will be modified to match whatever network is selected
to connect the source to. For example, the Source name “LAN” or “WAN” will be
overwritten to become whatever the name of the selected destination network is: “LAN”
will become “My Organization-internal” (or whatever is selected).
Therefore, it is important to connect to network names that are easily understood when
the modified vApp Template is re-deployed at later stages. Preferably, Organization
network names should be configured to match the Source names of the VMs to be
modified (the network name is also “LAN” or “WAN,” for example).
Organization network names can be modified by either Organization or System-level
administrators, under the Organization's Administration / Org vDC Networks tab.
Note: IMPORTANT: Do not run the deployed vApp (or its VMs) at this stage or configure
its operating system (MSL) or any attributes other than resource profile. Any
configuration, such as OS-level host name, network configuration, and so on, will be
captured in the modified vApp Template Catalog item in later steps, and are likely to be
incorrect for the deployment of the modified vApp Template. In this case, the MSL initial
configuration will not run (because MSL is already configured).
3. Open the deployed vApp (double-click or right-click Open), and open the desired contained
VM Properties dialog box (double-click or right-click Properties).
4. In the General tab, rename the VM appropriately.
5. In the Hardware tab (see Figure 2 below), modify the Number of vCPUs, Total memory,
or increase Disk size as appropriate to the desired profile. If required, additional disks may
be added in this step.
147
Mitel Virtual Appliance Deployment
Note: The disk size of an already deployed disk may only be increased, not decreased.
Where Table 8 indicates two disks, a new disk must be added to the VM when it is modified.
For example, when modifying from MiCollab Virtual or MiVoice Border Gateway Virtual
Small Business profile to the corresponding Enterprise profile, an additional disk of the
appropriate size must be configured at this step.
If memory is increased, and the corresponding reservation is also to be increased, then the
dialog box must be closed (Click OK), and then re-opened for the new value to be saved.
Do this before moving to the next step (Resource Allocation).
148
vCloud Director
6. In the Resource Allocation tab (see Figure 3 below), modify the vCPU Reservation, Pri-
ority (shares) and Limit, as well as Memory Reservation, Priority (shares) and Limit, as
appropriate for the desired profile. Disk shares and I/O limits cannot be modified.
Note: The Resource Allocation tab is available only for vApps deployed in a vDC using
the Reservation Pool allocation model.
7. Add the modified vApp into the Catalog as a vApp Template. Right-click Add to Catalog
or select the vApp, click Tools, and select Add to Catalog.
149
Mitel Virtual Appliance Deployment
In this approach, the VA is first deployed at vSphere level, then into vCloud Director as a vApp
Template Catalog item. While the number of vCPUs, memory, and disk are preserved under
all allocation models, only Reservation Pool-based vDC fully preserves vCPU and memory
reservations, as described in Table 30, “Resource Settings vs. Allocation Model,” on page 141
above.
Note: This method is available only to system level administrators. Organization level
administrators do not have permissions for Import from vSphere.
150
vCloud Director
When the VA is deployed from the Catalog, the administrator or user can also modify the lease
time assigned to the VA, up to the maximums configured for the Organization. After deployment,
the administrator or user must periodically reset the leases for the VA to avoid expiry, unless
the lease is set to Never Expires (recommended for mission-critical VAs). This also applies to
Catalog Templates (vApp template lease).
When the vApp or VA run-time lease expires, the VA will be shut down, potentially interrupting
service. The VA can still be re-started from its storage, as long as the vApp storage has not yet
expired.
For mission critical VAs—and most Mitel VAs are mission-critical—you must either:
• Set the run-time and storage leases to Never Expires, or
• Set up automated processes to monitor and track expiry dates, and automate lease renewal
to avoid interruptions due to expired leases.
Robust processes for preventing unplanned and unexpected lease expiry are strongly
recommended.
151
Mitel Virtual Appliance Deployment
Un-deploying a VA does not affect other VAs running in the same Organization. Resources are
relinquished, allowing other VAs to be started, or granted a larger share of the resources
allocated to the Organization vDC.
152
Networking and security considerations
Allows the application to connect directly to an external network, such as an enterprise LAN
(private cloud), extended customer WAN (public cloud with MPLS or other VPN support),
or public Internet (public cloud or enterprise private cloud).
• Network Address Translation (NAT)-routed external network
During Organization network creation, the vCloud Director Create Organization Network Wizard
provides some typical setup options. Advanced options are also available, which allow you to
configure the same set up steps in finer detail.
Organization networks are created when Organizations are initially configured or later, when
additional networks are created for an existing Organization. These options are discussed in
the following subsections.
Mitel applications can readily be deployed in this topology, by treating the direct-connected
external network as if it were an enterprise LAN. In this case, it is assumed that no MBG-based
protection of the Mitel applications is required. Mitel applications are deployed to directly
connect to the external network, as is usual in a LAN environment.
153
Mitel Virtual Appliance Deployment
MiVoice Business Express does not fit this network topology. This network topology is
appropriate only for stand-alone applications with no MBG protection, and where there is some
form of firewall protection provided between the vCloud Director networks and the Internet (or
other untrusted network).
Recommendations:
• The direct-connected, flat Organization network topology is recommended for Mitel virtual
applications in cases where MBG protection of the applications is not required. Typically,
this would be in an Enterprise or SMB private cloud deployment, where the enterprise has
direct control over the entire network. In the hosting, public cloud case, the external network
must be an extension of the customer WAN, through MPLS or other VPN, for example.
• This topology is not suitable for MiVoice Business Express.
• In many circumstances this topology is not available for deployment due to hosting provider
restrictions, for example. When this is the case, another direct-connected (layered) topology
or a NAT-routed topology should be used.
154
Networking and security considerations
Mitel applications can readily be deployed in this topology, by treating the direct-connected
external network as if it were the Internet. It may be any of the following:
• Live Internet connection (private cloud DMZ, or public cloud, hosting-provided Internet)
• Enterprise LAN (private cloud)
• The direct connection may map to an external-facing network for a particular customer
(public cloud/hosting). In this case, the external network may be an extension of the cus-
tomer WAN, for example, through MPLS or other VPN.
In this case, MBG Virtual is deployed in server-gateway mode, bridging between the
direct-connected external network and the internal network. This provides the necessary
security to the other Mitel applications, which are deployed on the isolated internal network.
External devices such as IP Phones are configured in teleworker mode, following normal MBG
deployment rules. Other MBG services, such as web proxy, may also be enabled following
normal MBG deployment rules.
155
Mitel Virtual Appliance Deployment
Alternatively, MiVoice Business Express may be deployed, with its WAN interface connected
to the external network, and its LAN interfaces connected to the isolated internal network.
An example vCloud Director configuration for this is shown in Figure 6 below. This is a screen
capture from vCloud Director Organization > My Cloud > VM view.
Recommendations:
• Direct-connected, layered Organization network topology is recommended approach for
Mitel virtual applications for which MBG protection is required.
• This is preferred over the NAT-routed topology (described in the next section), because it
requires reduced configuration overhead for the vCNS component (which is not required
in this topology), and reduced risk of vCNS adding to media flow latency and jitter (and
potentially reduced voice quality).
• In some circumstances, this topology may not be available for deployment, due to hosting
provider restrictions, for example. In that case, the NAT-routed topology should be used.
156
Networking and security considerations
In the NAT-routed topology, Mitel applications are deployed behind the vCNS NAT/Firewall,
however, some may not operate properly due to the address translation. Operation of Mitel
applications behind NAT is application-specific, and the same as the behavior of the application
behind a physical NAT device. Refer to the application-specific documentation for details.
For this network topology, MiVoice Border Gateway Virtual is deployed between the
external-facing network and the isolated internal network. In effect, MBG Virtual is in series
with a vCNS NAT/Firewall.
157
Mitel Virtual Appliance Deployment
An example vCloud Director configuration for this is shown in Figure 8 below. This screen
capture is taken from the vCloud Director Organization > My Cloud > VM view.
Recommendations:
• The NAT-routed/layered Organization network topology is recommended approach for Mitel
virtual applications in which MBG protection is required.
• The direct connected/layered Organization network topology described in “Deployment in
direct-connected, layered Organization network” on page 155 would generally be preferred,
if possible, due to added complexity in configuration imposed by adding NAT in the path
and the risk of reduced voice quality due to latency and jitter possibly added by the Edge
component (considered low risk).
158
Appendix D
SECURITY IN MITEL/VMWARE VIRTUAL
NETWORKS
Mitel Virtual Appliance Deployment
160
Security in Mitel VMware deployments
When setting up your network with virtual servers, security considerations are the same as for
traditional servers, and all of the same security measures, including firewalls, DMZ,
authentication, encryption, and SSL-protected connections, apply.
In addition, there are measures you can take specific to your virtualized environment; these
are described in this chapter. For detailed VMware product information, see the VMware
documentation.
Note: This chapter provides information about VMware and Hyper-V applications and
requirements that were valid for previous releases, and which may not be up-to-date.
Always refer to your current VMware or Hyper-V documentation for the latest information.
Security components of vCNS were formerly known as the VMware product suite.
Additional networking functionality is included in vCNS, but is not directly relevant to Mitel
deployments, so it is not described or discussed here. Please refer to VMware documentation
for details.
VSHIELD MANAGER
vShield Manager provides centralized management for the other vCNS/vShield security
products. One instance of vShield Manager is installed and configured for each vCenter in the
overall environment.
vShield Manager is accessed through a plug-in to vCenter, which appears in the vSphere Client
the administrator uses to manage the vCenter, or through a Web administration interface on
the vShield Manager application.
VCNS
vCNS provides gateway services at the network edge. The basic architecture is similar to that
of a NAT/Router in the physical world, as shown in Figure 9 below.
161
Mitel Virtual Appliance Deployment
vCNS is deployed as one or more virtual appliances within the virtual infrastructure. vCNS
instances may be deployed between sections of the data center network, at the external facing
edge of the data center network, in a virtual DMZ.
Capabilities include:
• Network Address Translation (NAT) with external and internal address mapping capability,
inbound and outbound access rules, and so on.
• Firewall, with 5-tuple inbound and outbound access rules.
• vCNS High Availability, with two vCNS instances in active-passive configuration, allowing
stateful fail-over (vCNS Advanced only).
Note: While the overall service will recover, there could be some delay to vCNS fail-over,
so there would be loss of voice or other media flows for that period and signaling delays
as well.
Note: This fail-over of vCNS instances is not equal or equivalent to the vSphere HA
feature, which provides automatic restart of VMs in case of VM failure, or failure of the
host on which the VM resides.
• vCNS is available at different performance levels: Compact, Large and X-Large, allowing
scaling up of the solution. In-place vCNS appliances can be transformed to the larger
capacity configuration, but cannot be changed to a lower performance level.
162
Security in Mitel VMware deployments
• Site-to-site VPN.
• Load balancing across multiple external-facing interfaces.
• DHCP address configuration on the internal-facing network.
• DNS relay service for the internal network.
vCNS, on its own, does not provide protection against network-internal threats (threats from
inside the NAT boundary). Nor does it provide isolation of network-internal concerns, as may
be required for regulatory compliance (like Sarbanes Oxley or Payment Card Industry (PCI)
Data Security Standard, for example).
163
Mitel Virtual Appliance Deployment
Many other approaches are feasible, including use of other vCNS components such as vShield
App, depending on overall solution requirements and architecture. The topologies described
in the following sections have been tested and verified, so they are supported and
recommended.
Note: Due to the possibility of reduced voice quality, use of vCNS Compact sizing is not
recommended or supported, except where no UCC traffic will flow through the vCNS
appliance.
• Deploy all security related components (vCNS, MBG Virtual, MiVoice Business Express)
on clusters with vSphere DRS/vMotion and HA enabled. Security related components
should be configured in HA for high restart priority.
• Enable vCNS High Availability on all vCNS appliances. This ensures maximum continuity
should one of the active-passive pair of Edge appliances fail. Note that this feature is not
the same as vSphere HA, and that there are some serious limitations for VoIP applications.
See “vCNS” on page 161 for more information.
164
vCNS Security Topologies
In the diagrams, VMware elements are shown in blue shading; Mitel elements are shown in
red shading.
This topology is suitable for enterprise private cloud type deployments, where the enterprise
administrator has full control over their network and virtual infrastructure. It is assumed that
network-internal threats and strong isolation between internal networks are not major concerns.
Normally, the virtualized infrastructure would be on premise (in the corporate data center).
However, it may be feasible for the infrastructure to be hosted externally by a provider as an
IaaS deployment, using a VLAN technology such as MPLS to extend the enterprise network
to inside the host data center. vCNS site-to-site VPN capability could also be used. Such an
extended enterprise network would connect on the “internal side” nominally at the same place
MBG Virtual internal interface connects (VLAN-External-Facing in Figure 10).
165
Mitel Virtual Appliance Deployment
TOPOLOGY DESCRIPTION
The vCNS controlled Private Cloud - Parallel topology consists of the following elements:
• Some number of remote networks, such as branch or home offices and/or Teleworker users.
These access the private cloud across Internet (or other non-trusted network). These are
typically protected by a NAT/firewall device (not shown). All endpoints in these remote
networks are assumed to be in Teleworker mode of operation.
• Internet public network.
• Access Router, providing access to and from the untrusted network, usually the Internet.
• (Optional) In practice, there may also be a physical firewall in the Internet path, behind or
as part of the access router. This is quite likely in large enterprise case, for protection of
non-virtualized resources inside the enterprise network.
• DMZ network layer, where machines meant to span between untrusted and trusted network
domains reside. The external side of the DMZ attaches to the untrusted network (VLAN-In-
ternet in the diagram), typically using public IP addressing. The internal side attaches to
the trusted internal network (VLAN-External-Facing in the figure). Machines in this layer
are virtualized, and may be running on the same host servers as other non-DMZ VMs.
Access to the DMZ VM vNICs is strictly controlled through the use of VLANs.
• MBG Virtual, residing in the DMZ network layer, and configured in server-gateway mode,
provides firewall, NAT handling, and access, specifically to and from the Mitel UC applica-
tions (MiVoice Business Virtual, and so on). The MBG Virtual may also be used for the SIP
Service Provider gateway function (SBC), if required. The external-facing “WAN” interface
of the MBG Virtual attaches to the external untrusted network (VLAN-Internet), and the
internal-facing LAN interface attaches to the internal trusted network
(VLAN-External-Facing).
• vCNS, residing in the DMZ network layer, and providing NAT/firewall service to general
purpose VMs in the overall solution. The external-facing interface of the vCNS instance
attaches to the external untrusted network (VLAN-Internet), and the internal-facing interface
attaches to the internal trusted network (VLAN-External-Facing).
Note: In this topology, vCNS is not directly required to support Mitel UCC applications.
vCNS and MBG do not directly interact, only co-exist. If only UCC applications are
included in the virtual environment, the vCNS component is not required.
• UCC applications network containing the Mitel UCC application VAs, other than the MBG
Virtual gateway (VLAN-UCC in the diagram). These are configured as they would be in a
traditional LAN environment. Note that MBG Virtual may still be present, if required for
internal services such as Secure Recording Connector (SRC).
• One or more “Other” internal networks. In the private cloud case, this is likely to include
separate networks for end user devices such as PCs, a voice network for VoIP devices
such as IP Phones, and possibly several more. Some of these other networks may exist
inside the virtual infrastructure (rather than external to it, as shown here). Devices used to
manage the virtual infrastructure (administrator PCs, for example) may reside on one or
more of these networks.
166
vCNS Security Topologies
For information about configuring vCloud Networking and Security (formerly vShield Edge),
MBG Virtual, the IP Phones (and other VoIP end-points), and the external firewall (if one exists
in the deployment), see Table 32, “Network component configuration by Topology type,” on
page 180.
For a comparison of the advantages and disadvantages, and the recommended applications
of each topology, see Table 31, “Topology comparisons and recommendations,” on page 176.
167
Mitel Virtual Appliance Deployment
This topology is similar to the topology described in the previous section, but moves the MBG
Virtual function “inside”, behind the vCNS NAT/firewall.
This topology is appropriate for enterprise private cloud type deployments, where the enterprise
administrator has full control over their network and virtual infrastructure. It is assumed that
network-internal threats and strong isolation between internal networks other than UCC are not
major concerns. Additional layering of security is provided for UCC applications, so security
from internal threats on UCC is stronger.
TOPOLOGY DESCRIPTION
The vCNS controlled Private Cloud - Series topology consists of the following elements. Only
elements different from the previous topology are described here.
• MBG is moved to the internal network, behind vCNS and the (optional) physical firewall.
MBG Virtual is in series with the vCNS NAT/firewall.
• MBG Virtual is configured in Server-Gateway mode, providing firewall, NAT handling and
access to and from the Mitel UCC applications (MiVoice Business Virtual, and so on).
• The MBG Virtual may also be used for SIP Service Provider gateway function, if required.
• The external-facing WAN interface of the MBG Virtual attaches to an internal trusted network
(VLAN-External-Facing in the diagram), and the internal-facing LAN interface attaches to
a different internal trusted network (VLAN-UCC in the diagram).
• vCNS, residing in the DMZ network layer, provides address and port mapping for Mitel UCC
applications via MBG Virtual, as well as provides NAT/firewall service to general purpose
VMs in the overall solution.
• Specific vCNS configuration is required to support Mitel UCC/MBG Virtual interface.
For information about configuring vCloud Networking and Security (vCNS), MBG Virtual, the
IP Phones (and other VoIP end-points), and the external firewall (if one exists in the deployment),
see Table 32, “Network component configuration by Topology type,” on page 180.
For a comparison of the advantages and disadvantages, and the recommended applications
of each topology, see Table 31, “Topology comparisons and recommendations,” on page 176.
For MiVoice Business Express, two IP addresses are used on each of the external-facing
connections (VLAN-External-Facing), and on the internal-facing connection (VLAN-UCC). This
168
vCNS Security Topologies
is required for MSL/MiCollab/MBG external addressing and for MCA proxy external address,
as well as the MSL/MiCollab/MBG internal address, and the MiVoice Business internal address.
This topology is suitable for enterprise private cloud deployments, where the enterprise
administrator has full control over the network and virtual infrastructure. It is assumed that
network-internal threats, and strong isolation between internal networks other than UCC are
not major concerns. Additional layering of security is provided for UCC applications by the MBG
component of MiVoice Business Express, so security from internal threats on UCC is stronger
than in the previous two topologies described. In addition, use of MiVoice Business Express
simplifies deployment.
TOPOLOGY DESCRIPTION
The vCNS controlled Private Cloud - MiVoice Business Express topology consists of the
following elements. Only elements different from the previous topologies are described here.
• MiVoice Business Express is deployed between the trusted side of the DMZ and the UCC
network. In effect, the MBG component of MiVoice Business Express is in series with the
vCNS NAT/firewall.
• In MiVoice Business Express, the MBG component is configured in Server-Gateway mode,
providing firewall, NAT handling and access to and from the other MiCollab applications.
169
Mitel Virtual Appliance Deployment
• The MBG component may also be used for SIP Service Provider gateway function, if
required.
• The external-facing WAN interface of MiVoice Business Express attaches to an internal
trusted network (VLAN-External-Facing in the diagram), and the internal-facing LAN inter-
face attaches to a different internal trusted network (VLAN-UCC in the diagram). Two fixed
IP addresses are assigned to each interface.
• vCNS, residing in the DMZ network layer, provides address and port mapping for MiVoice
Business Express, and provides NAT/firewall service to general purpose VMs in the overall
solution. For MiVoice Business Express, two externally visible fixed IP addresses are
mapped to MiVoice Business Express WAN addresses.
For information about configuring vCNS, MBG Virtual, the IP Phones (and other VoIP
end-points), and the external firewall (if one exists in the deployment), see Table 32, “Network
component configuration by Topology type,” on page 180.
For a comparison of the advantages and disadvantages, and the recommended applications
of each topology, see Table 31, “Topology comparisons and recommendations,” on page 176.
170
vCNS Security Topologies
This topology is well-suited for hosted multi-tenant public cloud deployments, providing UCC
as a service (UCaaS), possibly along with other services. The hosting provider has full control
over their network and virtual infrastructure; however the customer is assumed to have limited
control. It is assumed that network-internal threats within a particular customer are not a
concern. Strong isolation between customers is a very high concern.
171
Mitel Virtual Appliance Deployment
Note: This topology maps very closely to VMware vCloud Director NAT routed
Organization Networks, where deployment of vCNS appliances to isolate customer
networks is automated by vCloud Director. See “vCloud Director” on page 97.
If only UCC applications are being hosted, then the vCNS layer is not required for UCC
functionality. (It may, however, be required by the host provider for other reasons.) In this case
the MBG Virtual could be connected directly to the Internet, as a normal Server Gateway
deployment, without vCNS in the path. This variation is not discussed in detail in this guide;
refer to the MiVoice Border Gateway documentation for more information.
TOPOLOGY DESCRIPTION
The Public Cloud / vCNS topology is largely an expansion of the previous topology (“Topology
2: vCNS Based Private Cloud - MBG Virtual Behind vCNS” on page 167), by completely
replicating the topology across multiple customers. vCNS is used to isolate customers from
each other. Within each customer network, MBG Virtual performs the NAT traversal proxy
function, as well as providing an additional firewall layer in front of the UCC applications.
Optionally, a shared MBG Virtual (or cluster of MBG Virtual instances), acting as SIP
Proxy/Session Border Controller (SBC) for multiple customers, is also introduced.
Consolidation of capacity is provided for multiple customers. All trunks come in to one location,
and calls are routed to the appropriate customer's UCC environment (MiVoice Business Virtual
or MiVoice Business Express). This has potential advantages, but at the cost of forcing customer
networks to be not completely identical to each other.
• May enable reduction of total resource reservations across the overall hosting solution to
provide the SIP SBC services across multiple customers, due to statistical multiplexing
(peak loads for multiple customers are unlikely to occur at the same time). It could also
reduce the resource reservations required per customer MBG instance, while creating a
reservation for the shared MBG function that is lower than the total—reduced across the
customer base. This potential advantage can be expected to be higher where there are
large numbers of smaller end customers involved.
• Potential cost savings if SIP services are purchased in bulk by the hosting provider.
• Increased simplicity for bandwidth management. All SIP trunking becomes a single band-
width management zone.
• Increased resiliency of the SIP SBC functions, using MBG clustering.
In a public cloud environment, it can be assumed that an external firewall will usually also be
in place, outside of the virtualized environment.
172
vCNS Security Topologies
within the customer network to be pre-configured with respect to addressing when the
customer is deployed.
• Optionally, a shared MBG Virtual (or cluster of MBG Virtual instances), acting as SIP
Proxy/SBC for multiple customers, can also be deployed. This connects to each individual
customer internal network, nominally at the same point as the MBG Virtual internal interface.
This is the expected scenario in most hosting environments.
Note: If a shared SIP SBC is used, then customer internal networks can not be identical
to each other. Each customer network must have a different UCC network address
range/subnet, so that the shared SIP SBC can route customer-specific SIP traffic
correctly.
• If a shared MiVoice Border Gateway (MBG) is provided, then the customer MBG Virtual is
used for Teleworker access and Secure Recording Connector only. If there is no shared
MBG, then the customer MBG Virtual is also used as SIP Proxy/SBC for the individual
customer, as required.
• The per-customer MBG Virtual can also be used for remote administration and other pur-
poses, as per normal MBG operation. Refer to the MiVoice Border Gateway documentation
for more information.
For information about configuring vCNS, MiVoice Border Gateway, the IP Phones (and other
VoIP end-points), and the external firewall (if one exists in the deployment), see Table 32,
“Network component configuration by Topology type,” on page 180.
For a comparison of the advantages and disadvantages, and the recommended applications
of each topology, see Table 31, “Topology comparisons and recommendations,” on page 176.
173
Mitel Virtual Appliance Deployment
This topology is appropriate for hosted multi-tenant public cloud deployments, providing UCC
as a service (UCaaS), possibly along with other services, if required. The hosting provider has
full control of the network and virtual infrastructure; however, the customer is assumed to have
limited control. It is assumed that network-internal threats within each customer are not a
concern. Strong isolation between customers is a very high concern. Use of MiVoice Business
Express simplifies deployment and initial configuration.
174
vCNS Security Topologies
Optionally, a shared MiVoice Border Gateway (or cluster of MBG Virtual instances), acting as
SIP Proxy/SBC for multiple customers, can also be introduced. A discussion of advantages
and disadvantages of the shared MBG are discussed in Topology 4.
Note: This topology maps very closely to the VMware vCloud Director NAT-routed
Organization Networks. This deployment of vCNS appliances to isolate customer
networks is automated by vCloud Director. For more information, see “vCloud Director”
on page 97.
TOPOLOGY DESCRIPTION
The Public Cloud/vCNS with MiVoice Business Express topology is largely an expansion of the
previous “Topology 3: vCNS Based Private Cloud - MiVoice Business Express” on page 168,
completely replicating the topology across multiple customers. vCNS is used to isolate
customers from each other. Within each customer network, MiVoice Business Express MBG
component performs the NAT traversal proxy function, as well as providing an additional firewall
layer in front of the UCC applications.
Optionally, a shared MBG Virtual (or cluster of MBG Virtual instances), acting as SIP Proxy/SBC
for multiple customers, is also introduced.
In a public cloud environment, it can be assumed that an external firewall will also be in place,
outside of the virtualized environment.
The Public Cloud/vCNS MiVoice Business Express topology consists of the following elements.
Only elements different from the previous topology are described here.
• MiVoice Business Express is deployed between the vCNS internal interface and the UCC
network. This is functionally equivalent to the previous scenario.
For information about configuring vCNS, MBG Virtual, the IP Phones (and other VoIP
end-points), and the external firewall (if one exists in the deployment), see Table 32, “Network
component configuration by Topology type,” on page 180.
For a comparison of the advantages and disadvantages, and the recommended applications
of each topology, see Table 31, “Topology comparisons and recommendations,” on page 176.
175
Mitel Virtual Appliance Deployment
176
vCNS Security Topologies
TOPOLOGY 3
VCNS BASED PRIVATE CLOUD - MIVOICE BUSINESS EXPRESS
• MiVoice Business Express • Additional configuration • Recommended for simplified
simplified deployment and initial overhead on vCNS (minimal). deployment and configuration,
configuration. • Risk that vCNS could impact using MiVoice Business Express.
• Simple configuration and voice/video quality under load, • Private cloud/enterprise
management adding only two 1:1 due to added delay and jitter in deployments, where admins
address mapping rules at vCNS. media streams. This is a have full control of the internal
• Additional security specifically for relatively lower risk than in network and virtual environment,
applications behind the MBG Topology 2, due to the relatively and network-internal threats are
component of MiVoice Business small scale of MiVoice Business not a strong concern.
Express (on UCC VLAN). Express, unless the vCNS • Networks where additional
Appliance is under high load for protection from network-internal
• Similar to traditional VLAN-based
other reasons (non-UCC loads). threats for UCC applications is
approach. Easily understood and
familiar to administrator. • No protection from desired.
network-internal threats, other
than MiVoice Business Express
VLANs.
• No isolation of concerns
(separated trust zones,
compliance zones, tenants)
177
Mitel Virtual Appliance Deployment
178
vCNS Security Topologies
179
Mitel Virtual Appliance Deployment
180
vCNS Security Topologies
181
Mitel Virtual Appliance Deployment
182
vCNS Security Topologies
183
Mitel Virtual Appliance Deployment
184
© Copyright 2018, Mitel Networks Corporation. All Rights Reserved. The Mitel word and logo are trademarks of Mitel Networks
Corporation, including itself and subsidiaries and authorized entities. Any reference to third party trademarks are for reference only and Mitel
mitel.com makes no representation of ownership of these marks.