BP 2009 Metro Availability PDF
BP 2009 Metro Availability PDF
BP 2009 Metro Availability PDF
Copyright
Copyright 2020 Nutanix, Inc.
Nutanix, Inc.
1740 Technology Drive, Suite 150
San Jose, CA 95110
All rights reserved. This product is protected by U.S. and international copyright and intellectual
property laws.
Nutanix is a trademark of Nutanix, Inc. in the United States and/or other jurisdictions. All other
marks and names mentioned herein may be trademarks of their respective companies.
Copyright | 2
Metro Availability
Contents
1. Executive Summary.................................................................................5
2. Introduction.............................................................................................. 7
2.1. Audience.........................................................................................................................7
2.2. Purpose.......................................................................................................................... 7
5. Operational Scenarios...........................................................................24
5.1. Establishing Metro Availability..................................................................................... 24
5.2. Planned Failover.......................................................................................................... 25
5.3. Network Outage Between the Nutanix Clusters.......................................................... 28
5.4. Site Failure................................................................................................................... 31
5.5. Site Recovery...............................................................................................................35
5.6. Metro Availability Witness-Specific Workflows.............................................................38
5.7. Operational Scenarios Summary................................................................................. 41
5.8. Asynchronous Snapshot Workflows.............................................................................43
3
Metro Availability
9. Conclusion..............................................................................................62
Appendix..........................................................................................................................63
About the Author................................................................................................................. 63
About Nutanix...................................................................................................................... 63
List of Figures................................................................................................................ 64
List of Tables.................................................................................................................. 66
4
Metro Availability
1. Executive Summary
IT departments now regularly face demanding service-level agreements (SLAs) that necessitate
building resiliency into all aspects of the datacenter. While traditional solutions provide
redundancy at the hardware layer, their administrators must still manage downtime due to
infrastructure failures, site maintenance, or natural disasters.
Most solutions that exist today are expensive, complex, and built on legacy principles and
infrastructure. As a result, many application workloads are not protected—or are underprotected
—and vulnerable to outage. To help address these concerns, Nutanix has developed a
continuous availability solution called Metro Availability. Metro Availability creates a global file
system namespace across Nutanix clusters and uses synchronous replication. Combining the
Nutanix hyperconverged infrastructure with a continuous availability solution limits downtime
and preserves all data, even during a complete site failure. Further, Metro Availability enables
workload mobility for disaster avoidance or planned maintenance scenarios.
Metro Availability allows administrators to leverage hypervisor clustering technologies across
datacenters. We call this type of configuration a stretched cluster, and it helps to minimize
downtime during unplanned outages. Metro Availability also supports the migration of virtual
machines (VMs) across sites, using technologies such as vMotion, which means that you have
zero downtime while transitioning workloads between datacenters.
With Metro Availability, Nutanix now delivers an entire spectrum of solutions to provide the
data and application protection you need to fulfill your SLAs and meet your recovery point and
recovery time objectives (RPOs and RTOs). Nutanix has built safeguards into the platform for
everything from minor events, such as individual VM deletion, to major ones, including unplanned
datacenter failure. Setup and management are simple, intuitive, and controlled from the Prism UI.
With Prism, enterprises have, for the first time, a consumer-grade management experience for
handling disaster recovery and high availability.
The following diagram shows the Nutanix features as they align with RPO and RTO requirements
for minor or large incidents.
1. Executive Summary | 5
Metro Availability
1. Executive Summary | 6
Metro Availability
2. Introduction
2.1. Audience
We wrote this best practices document for those responsible for architecting, designing,
managing, and supporting Nutanix infrastructures. Readers should already be familiar with
VMware vSphere and the Nutanix enterprise cloud software, which includes Acropolis and Prism.
We have organized this document to address key items for enabling successful design,
implementation, and transition to operation.
2.2. Purpose
This document presents an overview of Nutanix and the Metro Availability feature, and we
discuss deployment considerations and general best practices around Metro Availability
functionality in a VMware vSphere environment. At the conclusion of this document, the reader
should be comfortable architecting and deploying a Metro Availability–based solution on Nutanix.
Version
Published Notes
Number
1.0 February 2015 Original publication.
2.0 April 2016 Updated for AOS 4.6.
3.0 January 2017 Updated for AOS 5.0.
3.1 March 2018 Updated platform overview.
3.2 May 2018 Updated Nutanix overview and deduplication recommendations.
3.3 September 2018 Updated container sizing and network configuration guidance.
3.4 August 2019 Updated for AOS 5.11.
3.5 June 2020 Updated for AOS 5.15 and AOS 5.17.
2. Introduction | 7
Metro Availability
The Nutanix platform supports Metro Availability in conjunction with other data management
features, including compression, deduplication, and tiering. Metro Availability also allows you
to enable compression for the synchronous replication traffic between the Nutanix clusters.
When you configure the remote site, enabling replication traffic compression reduces the total
bandwidth required to maintain the synchronous relationship.
Active Containers
Active containers are fully read and write and process all VM I/O operations. Local read
operations for a VM running against an active container function similarly to non-Metro
environments. All local write operations, including both random and sequential writes, go
through the oplog of the CVM with Metro Availability enabled. When local writes occur, data
replicates locally based on the container’s replication factor. In parallel, the system sends remote
writes to the oplog of a CVM in the cluster maintaining the standby container. The standby
container’s replication factor determines the number of writes required to complete the remote
write operation against the peer cluster. After all replication factor writes are processed for both
the active and standby containers, the VM receives acknowledgement that the write is complete.
The following diagram depicts this process.
Standby Containers
Standby containers receive synchronous write updates from the active container. Standby
containers mount to the appropriate hypervisor instances in the Nutanix cluster, so you can run
VMs locally to the hypervisor nodes that own the container and datastore. While the standby
container appears available, it does not process VM I/O directly. The system forwards all VM
read or write I/O targeted for a standby container to a CVM in the cluster that owns the active
container before returning data or acknowledgement to the VM. The following figures depict the
behavior of these remote read and write operations.
Failure Handling
A network failure between Nutanix clusters in a Metro relationship first halts any writes
against the active container, and the replication state reports “remote unreachable.” A standby
container becomes unavailable and inactive to the hypervisor host while the active container is
unreachable. The failure handling setting offers three options for disabling Metro Availability to
allow writes to continue against the active container: manual, automatic resume, or witness.
With failure handling set to manual, VM writes do not resume until either network connectivity
is restored between the clusters or an administrator manually disables Metro Availability. Use
this option in environments that require strict synchronous replication at all times between sites.
If network connectivity is restored, the replication state immediately returns to enabled and
synchronous replication continues between the clusters; otherwise, writes are held indefinitely.
If an administrator manually disables replication, a “disabled” state is reported and writes
resume, but they only propagate to the active container. Following a manual disable operation,
replication to the standby container does not resume until network connectivity is restored and
the administrator manually performs a reenable operation.
With failure handling set to automatic resume, if network connectivity is restored within the
timeout specified under the automatic replication break setting, the replication state immediately
returns to enabled and writes continue to replicate synchronously between the clusters. The
default automatic replication break timeout is 10 seconds. If the timeout period expires, the
system automatically disables Metro Availability and writes resume, but they only propagate to
the active container. Using this option, temporary network outages between sites do not impact
applications running against the active container. Following the automatic replication break,
replication to the standby container does not resume until network connectivity is restored and
the administrator manually performs a reenable operation.
The Nutanix Availability witness provides an automatic mechanism for disabling Metro Availability
for the primary protection domain and promoting the standby protection domain. With failure
handling set to witness, when network connectivity between the Nutanix clusters is interrupted,
both sites attempt to obtain a lock against the witness VM for each protection domain. A cluster
attempts to obtain the witness lock against primary protection domains after 10 seconds and
against standby protection domains after 120 seconds. If the primary protection domain obtains
the lock, the system automatically disables Metro Availability and writes resume, but they only
propagate to the active container. If the primary protection domain fails to obtain the lock, all I/O
is held for its respective container until either network connectivity is restored or an administrator
disables the protection domain. If the standby protection domain obtains the lock, the system
automatically promotes it to active. If the standby protection domain fails to get the lock, it
becomes unavailable and inactive to the hypervisor host until either network connectivity is
restored or an administrator manually performs the promotion. The Operational Scenarios section
contains more details regarding automatic failure handling when using the witness.
Once you have set the failure handling option during the protection domain creation process, you
can modify it from the Nutanix command line interface (nCLI) or the Prism UI. The Utilities and
Alerting section provides an example of how to modify this setting using the nCLI.
Reenable Replication
When replication has been disabled, either manually by an administrator or automatically based
on the failure handling setting, once network connectivity is restored, you can issue a reenable
command from the original active cluster to resume replication. When you select “reenable,” the
system creates a snapshot that references the last snapshot taken as a part of a previous enable
or reenable operation. While enabled, Metro Availability takes a snapshot automatically (once
every four hours or once every six hours with AOS 5.17 and later) to use as a reference point.
The data that represents the differences between the current and the last reference snapshot
then replicates to the standby container. The amount of data replicated may be more than the
incremental amount of change since the network or site outage and could even represent all data
if the containers were empty at the time of the last reference snapshot.
You can also perform the reenable option against the standby cluster to carry out planned
failovers or to reestablish Metro Availability following an unplanned event. The direction of
replication after choosing reenable depends on the cluster from which you selected it: Whichever
cluster you use to issue the reenable command becomes the active site and the source for
replication. If you use the standby cluster to reenable Metro Availability, replication for the
container is essentially reversed; the previous standby container is now active, and the original
active container is now standby.
The following flowcharts provide an overview of the replication states from the perspective of the
active protection domain in the next figure, and the standby protection domain in the subsequent
figure. Please note that these flowcharts do not include details about deleting a protection
domain, which is an additional option and affects states. The Operational Scenarios section
contains additional information on which steps to perform in each event.
• Shut down: Shuts down all VMs on the isolated host, assuming VMware tools are installed. If a
shutdown is not successful, a power-off command is issued within five minutes.
“should run on” rules. Must run on rules force VMs to always reside on a member of a specific
host group. Should run on rules attempt to place VMs on members of the specified host DRS
group, but they can be overridden based on certain conditions, including host failure.
Antiaffinity between a group of VMs and a group of hosts relies on “must not run on” and “should
not run on” rules. Must not run on rules prevent VMs from ever running on members of a
specified host group. Should not run on rules attempt to prevent VMs from running on members
of a specified host group, but can be overridden based on certain conditions, including host
failure.
Metro Availability offers flexibility when deciding whether to use must or should affinity or
antiaffinity rules. Using should rules can help automate restarting VMs across Nutanix clusters
during a site failure. If you don’t want automated restart across Nutanix clusters, you can use
must rules instead.
An ideal Metro Availability configuration places the vCenter Server into a fault domain (generally
a third site) separate from the active and standby Nutanix clusters. This structure allows
environment management if either the active or standby site fails.
Alternatively, you can place the vCenter Server in a container used for Metro Availability and
replicated to the standby site. In such a configuration, you can protect and recover the vCenter
Server with VMware HA or as a part of disaster recovery procedures between the Nutanix
clusters.
5. Operational Scenarios
Metro Availability provides several methods for managing replication states, including the Prism
UI, REST API, the nCLI, and PowerShell. The following sections outline nondisruptive planned
workload migration, manual unplanned failover, and automated unplanned failover. The Prism
Web Console Guide contains additional details on how to manage Metro Availability from Prism.
For scenarios including the Metro Availability witness, you must install the witness in a
failure domain separate from the Nutanix clusters, then register the witness with the clusters
participating in the Metro relationship.
An administrator can form a single VMware cluster that contains all the nodes from both Nutanix
clusters. You can configure VMware HA and DRS rules to manage VM placement policies.
5. Operational Scenarios | 24
Metro Availability
Nutanix recommends keeping VMs running locally to active containers. You can use VMware
DRS affinity rules to ensure that VMs are running against cluster nodes that own the active
containers. Should run on rules keep VMs running against a specific group of hosts but also
allow failover if a site fails. You can use must run on rules to prevent VMs from restarting against
certain cluster nodes even in the event of a site failure.
We provide specific steps for creating containers, remote sites, and Metro Availability protection
domains in the Prism Web Console Guide. The following figure summarizes the general steps.
5. Operational Scenarios | 25
Metro Availability
Reenabling replication following a promotion provides the shortest window for resuming the
synchronous relationship. Waiting to reenable replication allows you to validate VM operations
prior to marking the previous active container as standby. Additionally, if the scenario is simply to
test VM failover and not persist the data in the standby site, you can skip reenabling replication
from the standby site.
5. Operational Scenarios | 26
Metro Availability
Planned failover requires a force-promote of the targeted standby container. The force-promote
allows a standby container to become active and enables reads and writes to occur against the
local cluster. All VMs running against a given container must migrate to the secondary cluster
prior to promoting the standby container. The force-promote is nondisruptive to VMs running in
the standby container. When you force-promote a standby container, the current active container
goes into a decoupled state, making the datastore read-only.
Following promotion, if failure handling is configured for manual or automatic resume, the
formerly active site needs to be disabled. Following the disable operation, reenabling Metro
Availability from the standby site reverses the replication relationship. With AOS 5.11 and later,
witness failure handling does not require you to disable the formerly active site. While the
formerly active site is decoupled, you can reenable it directly from the promoted standby site.
The original active container is marked as standby and any new writes that occur overwrite its
contents.
The next figure outlines the planned failover process with vMotion when you have configured the
failure handling setting for either manual or automatic resume mode.
If you are using asynchronous snapshots in combination with Metro, you may need to suspend
those schedules prior to issuing a disable command:
ncli pd suspend-schedules name=<PDName>
Figure 13: Nondisruptive Migration with Manual or Automatic Resume Failure Handling
The next figure outlines the planned failover process with vMotion when you have configured
the failure handling setting for witness mode. Manage any asynchronous snapshot schedules as
previously noted.
5. Operational Scenarios | 27
Metro Availability
Tip: When using AOS 5.11 and witness failure handling, you don’t need to perform
the disable operation during a planned failover.
5. Operational Scenarios | 28
Metro Availability
lack of site communication also causes datastore heartbeating to fail for partitioned servers
from the perspective of the active datastores. With both host communication and datastore
heartbeating failed, the leader VMware HA agent attempts to restart VMs.
Before attempting to restart failed VMs, VMware HA ensures that the cluster has resources
available to support the VMs, such as network port groups and datastores. When standby
datastores have an inactive state, VMware HA can’t attempt to restart VMs against those
containers. VMware HA retry counters don’t apply, as the servers are no longer on the VMware
HA compatibility list. Thus, VMware HA continually looks to update its compatibility list so it can
restart the VMs once the standby containers become available again.
Any VM running against a standby container when this kind of network outage occurs loses
access to its datastore (one of the reasons Nutanix recommends running VMs locally to their
active containers). You can restart these failed VMs on the opposite site, against the active
container, as a part of the VMware HA failure detection and VM restart process.
While VMware HA does not require VMware vCenter to be available to restart VMs, certain
management operations, such as modifying VMware HA or DRS settings (including affinity rules),
are affected when the vCenter Server is unavailable. Depending on the location of the vCenter
Server instance, options for managing the cluster in a degraded state may be limited.
5. Operational Scenarios | 29
Metro Availability
VMs running against active containers continue to operate but the network failure may affect
them, depending on the Metro Availability failure handling setting and the duration of the outage.
5. Operational Scenarios | 30
Metro Availability
5. Operational Scenarios | 31
Metro Availability
state (see the following figure). If required, VMware HA reelects an HA leader that attempts to
restart VMs that are offline because of the site outage.
Metro Availability replication enters a degraded state (remote unreachable), and the standby
datastores become unavailable and enter an inactive state. The inactive state prevents VMware
HA from attempting to restart VMs against those containers. VMware HA retry counters do not
apply, as the servers are no longer on the VMware HA compatibility list. This means that VMware
HA continually looks to update its compatibility list so it can restart the VMs when the standby
containers become available again.
When the remote site has failed, you can promote any standby containers in the surviving
site. Once promoted, the containers become active again to the VMware HA cluster. VMware
HA updates its compatibility list and powers on VMs that reside in that container. VMware HA
automatically overrides VMware DRS should run on affinity rules, and VMs governed by those
rules can restart in the surviving site. Must run on affinity rules are enforced, and you must
update them to allow VMs covered by those rules to restart in the surviving site.
Any VM running against a standby container when this kind of network outage occurs loses
access to its datastore (one of the reasons Nutanix recommends running VMs locally to their
active containers). As the opposite site has failed, these VMs can only resume when the standby
container is promoted.
While VMware HA doesn’t require VMware vCenter to be available to restart VMs, certain
management operations, such as the modification of VMware HA or DRS settings (including
affinity rules), are affected when the vCenter Server is unavailable. Depending on the location of
the vCenter Server instance, options for managing the cluster in a degraded state may be limited.
5. Operational Scenarios | 32
Metro Availability
VMs running against active containers continue to operate, but the site failure may affect them
if the Metro Availability failure handling setting is set to manual, as we discussed in an earlier
section.
5. Operational Scenarios | 33
Metro Availability
prevent migration attempts if the failed site and cluster return to service. The next figure shows
the general process.
5. Operational Scenarios | 34
Metro Availability
Starting with the AOS 5.15 release, Metro uses VMCP APD implementation to handle storage-
only failures. In case of storage-only failure on the primary site, Metro Availability detects the
APD condition and automatically fails over the VMs on the affected site to the secondary site
after the storage is promoted on the secondary site. This scenario makes the storage on the
primary (affected) site unavailable for reads or writes.
Tip: You must enable APD on the ESXi hosts for automatic VM failover to work in
a Metro Availability configuration. For more information on how to configure VMCP,
refer to VM Component Protection in VMware’s documentation.
5. Operational Scenarios | 35
Metro Availability
Figure 20: Initial Cluster State Following Site Recovery with Existing Configuration
Prior to recovering a cluster in a Metro Availability relationship that has been offline, ensure that
DRS settings, including affinity rules, do not cause unwanted VM movement. If you recover a
site and cluster, it is possible for legacy DRS settings to attempt to move VMs to unwanted hosts
that have stale data. To prevent this unwanted movement, set DRS to manual or update rules to
enforce must not run on requirements temporarily while replication is reenabled.
To restart Metro Availability once you have recovered a site, you must disable active protection
domains in the recovered cluster if you use manual or automatic resume mode failure handling.
You can then issue a reenable command. When you use AOS 5.11 or later and witness failure
handling, you do not need to disable the active protection domains in the recovered cluster in
order to reenable them from the promoted site. Ensure that you issue the reenable operation
from the appropriate Nutanix cluster, as the site chosen becomes the sole active copy and source
5. Operational Scenarios | 36
Metro Availability
of replication. For the example given in the previous figure, the reenable command comes from
Site B. The following figure outlines the general process.
If you are using asynchronous snapshots in combination with Metro, you may need to suspend
those schedules before you issue a disable command:
ncli pd suspend-schedules name=<PDName>
5. Operational Scenarios | 37
Metro Availability
Witness Failure
In the context of the Metro relationship, the witness is passive; it isn’t required for cluster
availability or replication when Metro is in a healthy state. A witness can fail or sites can lose
communication with the witness and Metro replication continues without interruption. Similarly, if
a failed witness returns to operation, you don’t need to take any additional steps to recover the
environment.
The witness only queries lock status and determines whether to allow an automatic disable or
promotion of protection domains when communication between the Nutanix clusters fails. If the
witness is unavailable at this time, the automatic decision making fails, affecting VM availability in
much the same way as manual failure handling.
If the witness is permanently lost, you must associate the protection domains with a new witness
or new control protection domain states as needed. With a healthy Metro Availability relationship,
you can change the failure handling of the protection domain from witness to another option,
such as automatic resume, to accomplish this “unwitness” operation.
5. Operational Scenarios | 38
Metro Availability
you to locally unwitness protection domains. A local unwitness operation disassociates protection
domains in that cluster from the existing witness, allowing you to manually disable active
protection domains and promote standby protection domains. Use the nCLI to perform the local
unwitness function:
ncli pd update-failure-handling name=<PDName> failure-handling=Automatic local-only=true
Only perform a local unwitness operation following site loss if you expect the existing witness to
be unavailable for an extended period. The next figure depicts the general process for recovery
where both Site A and the witness site fail.
5. Operational Scenarios | 39
Metro Availability
5. Operational Scenarios | 40
Metro Availability
Rolling Failure
A rolling failure entails communication loss first between Site A and Site B followed by the
complete loss of Site A. During the initial failure between Site A and Site B, the cluster that owns
the active protection domain attempts to acquire the witness lock. If the active protection domain
successfully acquires the lock, Metro Availability is disabled automatically and I/O resumes.
During this time, the cluster that owns the standby protection domain also detects the network
interruption and, after 120 seconds (by default), attempts to obtain the witness lock. As the
lock was already acquired by the active site, the standby site fails to obtain it, and the standby
container becomes inactive.
The rolling failure continues and Site A is then lost. Because Site B was initially unable to obtain
the witness lock, the standby protection domains remain in an inactive state, unable to serve
VMs. You must locally unwitness any standby protection domains in order to promote them to
active. Once you have promoted these domains, VMs can restart via the HA recovery process.
Perform the local unwitness operation as described in the Primary Site and Witness Loss section.
5. Operational Scenarios | 41
Metro Availability
Automatic Resume
Failure Scenario Witness Mode Manual Mode
Mode
Automatic protection
An administrator must An administrator must
Site A outage or domain promotion
promote protection promote protection
complete network in Site B. VMs
domains in Site B for domains in Site B for
failure in Site A automatically restart in
VMs to restart. VMs to restart.
Site B.
VMs continue to run VMs continue to run VMs are paused and
Site B outage or
on Site A following an on Site A following an an administrator must
complete network
automatic disable of automatic disable of disable protection
failure in Site B
protection domains. protection domains. domains in Site A.
VMs continue to run VMs continue to run VMs are paused and
Connection loss
on Site A following an on Site A following an an administrator must
between Site A and
automatic disable of automatic disable of disable protection
Site B
protection domains. protection domains. domains in Site A.
No impact while Metro
Witness failure N/A N/A
relationship is healthy.
Connection loss
No impact while Metro
between witness and N/A N/A
relationship is healthy.
Site A
Connection loss
No impact while Metro
between witness and N/A N/A
relationship is healthy.
Site B
Connection loss
No impact while Metro
between witness and N/A N/A
relationship is healthy.
both Site A and Site B
VMs on Site A are
Connection loss paused. Automatic VMs continue to run VMs are paused and
between Site A and protection domain on Site A following an an administrator must
Site B and between promotion in Site B. automatic disable of disable protection
witness and Site A VMs automatically protection domains. domains in Site A.
restart in Site B.
5. Operational Scenarios | 42
Metro Availability
Automatic Resume
Failure Scenario Witness Mode Manual Mode
Mode
VMs on Site A
are paused. An
Connection loss VMs continue to run VMs are paused and
administrator must
between all sites on Site A following an an administrator must
unwitness in Site A
including the witness automatic disable of disable protection
and disable protection
(Site A recovery) protection domains. domains in Site A.
domains to resume
VMs.
An administrator
Connection loss must unwitness in An administrator must An administrator must
between all sites Site B and promote promote protection promote protection
including the witness protection domains domains in Site B for domains in Site B for
(Site B recovery) in Site B for VMs to VMs to restart. VMs to restart.
restart.
An administrator
must unwitness in An administrator must An administrator must
Rolling failure: Metro
Site B and promote promote protection promote protection
failure followed by Site
protection domains domains in Site B for domains in Site B for
A failure
in Site B for VMs to VMs to restart. VMs to restart.
restart.
All VMs pause until All VMs pause until
you fix the storage you fix the storage
The witness promotes
outage or manually outage or manually
the storage on Site B.
promote the storage promote the storage
The storage on Site A
Storage-only outage on Site B. After you on Site B. After you
becomes inaccessible.
on Site A promote storage promote storage
If you enabled VMCP
on Site B, manually on Site B, manually
for APD, VMs fail over
migrate VMs by migrate VMs by
to Site B.
generating a VMware generating a VMware
HA event. HA event.
5. Operational Scenarios | 43
Metro Availability
5. Operational Scenarios | 44
Metro Availability
Snapshots for Metro Availability–enabled protection domains occur at the container level, which
means that each snapshot operation captures all files in the container. By default, the system
takes a snapshot every four hours to create a checkpoint used for incremental resynchronization
if the Metro relationship becomes disabled. The default snapshot interval with AOS 5.17 and later
is six hours.
Snapshot Scheduling
Administrators can take snapshots intended for backup and restoration manually or configure
them within a schedule. Schedules are configured against the cluster that has the Metro
Availability protection domain in the active state. You can use Prism to establish schedules that
create local snapshots in both the source and target Metro clusters. Optionally, you can select a
third remote site to retain snapshots outside of the Metro relationship. The following figure shows
these options.
Snapshot Restore
Snapshots are restored either within the active Metro protection domain, or, in three-site
scenarios, against the asynchronous protection domain. You cannot restore snapshots within
the cluster that hosts the standby Metro protection domain. A snapshot contains all files in the
5. Operational Scenarios | 45
Metro Availability
protection domain. Because the restoration process recovers all files, use a redirected restore.
A redirected restore recovers the files to a subdirectory in a container, preventing you from
overwriting any active files.
Perform restores from the nCLI using the following steps:
• Obtain the snapshot ID. For example, to get the snapshot ID for a protection domain called
metropd (using a bash shell):
ncli pd ls-snaps name=metropd | grep "Id\|Create Time"
• Following a restore using the above example, a /temp directory is available in the metropd
datastore in the ESXi cluster. You can copy any folders and files required for recovery to other
locations if necessary.
• Register the restored VMs by adding the configuration files to inventory.
• If you restore snapshots and register the VMs to a new location, ensure that the VMDK paths
for the VM point to the path specified as part of the restore operation.
• Delete all remaining folders and files that were not needed as a part of the restore.
5. Operational Scenarios | 46
Metro Availability
5. Operational Scenarios | 47
Metro Availability
• Optional: Because Metro clusters have been offline, synchronize changes to shorten migration
time.
pd add-one-time-snapshot name=metropd remote-sites=SiteA retention-time=86400
• Optional but recommended: Perform the initial full replication to shorten migration time.
pd add-one-time-snapshot name=metropd remote-sites=SiteA retention-time=86400
5. Operational Scenarios | 48
Metro Availability
5. Operational Scenarios | 49
Metro Availability
6.1. Requirements
• 5 ms round-trip time latency maximum between active and standby Nutanix clusters.
⁃ Validated by the system when configuring Metro Availability protection domains.
• Active and standby container names must be identical between clusters.
⁃ Validated by the system when configuring Metro Availability protection domains.
• Datastore names must be the same as the container name.
• Ensure that all virtual disks associated with a given VM reside on a container enabled for
Metro Availability.
6.2. Interoperability
• The NX-2000 series is not supported.
• You can create local and remote snapshots between the Metro Availability–enabled
containers. You can configure a third Nutanix cluster to be the target for remote asynchronous
replication.
• Cluster configuration:
⁃ Cluster models can be different between active and standby sites.
⁃ vMotion may require Enhance vMotion Compatibility (EVC).
⁃ Cluster resource sizes can be different.
⁃ Redundancy factor can be different.
• Container configuration settings:
⁃ Container specific settings, such as compression and replication factor, can be different.
• Do not enable a proxy on remote sites that you use with a Metro Availability protection
domain.
• When linked clones are in a Metro Availability container, the gold image must reside in the
same container.
6.3. Limitations
Prior to AOS 5.17, Metro Availability performed a snapshot every four hours, which is supported
in nodes with up to 80 TB HDD tiers. With AOS 5.17 and later, the default snapshot interval is six
hours, which enables support for nodes with up to 120 TB HDD tiers. If you use a version before
5.17, you can modify the default snapshot schedule if needed to support denser nodes.
Before you enable Metro Availability on a container with VMs in an async DR protection domain,
delete the async DR protection domain or remove the VMs associated with the Metro container
from the protection domain.
Latency on vDisks might increase during the synchronization between two clusters.
If you disable Metro Availability, I/O from the standby site can’t be forwarded to the primary site.
Commands run on the hypervisors on the standby site may take additional time to run because
the hypervisors can’t access the underlying storage.
Restoring snapshots of a Metro protection domain with overwrite is not supported.
Symbolic links and hard links are not supported.
You can’t host VMs on the secondary cluster during the Metro enable operation for the same
container.
You can’t host VMs on the primary cluster during promotion of the secondary site for the same
container.
• Create no more than 50 protection domains in a Metro Availability cluster pair configured for
witness failure handling.
• 200 ms round-trip latency or less between the witness and the Nutanix clusters participating in
Metro Availability.
• Use Nutanix alerts to monitor Metro Availability replication health.
• Enable remote site compression in bandwidth-limited environments.
• Manually take a snapshot of the protection domain before you disable Metro Availability to
ensure the most efficient synchronization when you reenable Metro Availability.
Metro Availability replication to the standby site does not maintain VM-specific data locality
to a particular node in the standby cluster. When failing over a VM to the standby cluster and
promoting the standby container, any node in that cluster can service local reads. Over time, data
becomes resident on the node that owns the running VM. Therefore, you do not have to target a
specific node for a VM to use on failover to a promoted standby cluster.
Restart Priorities
You can configure VM restart priorities to control the order in which VMware HA restarts VMs.
Setting these priorities can help in correctly restarting multitiered applications that run across
multiple VMs. It is important to understand that the restart priority is based on the task of
powering on the VM and is not linked to application availability within the VM. Because of this
basis, it is possible for lower-priority VMs to be operational before higher-priority VMs.
Isolation Response
The isolation response setting for a Metro Availability environment should only be relevant when
individual nodes fail, not when a cluster is partitioned. When a cluster is partitioned (as with a
network outage between the sites), the nodes local to a site can communicate with each other
and respond to VMware HA election traffic. In this case, host isolation does not occur.
Host isolation does occur when an individual node fails because of a network outage for that
node, and management traffic, including response to either election traffic or the VMware HA
leader, also fails. Datastore heartbeating is likely to fail as well, considering the converged
networking of the Nutanix platform. Because of this behavior, we generally recommend
configuring the isolation response to shutdown for any user VMs running on that server. Always
set Nutanix CVMs to leave powered on or disabled, depending on the version of vSphere you
use.
Datastore Heartbeating
We recommend configuring datastore heartbeating against containers that mount to all nodes
in a VMware HA cluster. When using Metro Availability, this means you configure datastore
heartbeating against the stretched containers enabled for replication. Given that the single
VMware HA leader in the cluster could be operating in either site, a perceived host failure in
the opposite site could be validated against the replicated container configured for datastore
heartbeating. The additional remote replication traffic caused by enabling datastore heartbeating
is minimal and should not be a performance concern.
⁃ create | add
⁃ remove | rm
⁃ metro-avail-enable (includes reenable option)
⁃ metro-avail-disable
⁃ promote-to-active
⁃ update-break-replication-timeout
⁃ update-failure-handling
Example:
protection-domain update-break-replication-timeout name=DSHB_PD timeout=10
• Promote-NTNXProtectionDomainStretchCluster
• Start-NTNXProtectionDomainStretchCluster
⁃ Maps to enable
⁃ Includes reenable option
• Stop-NTNXProtectionDomainStretchCluster
⁃ Maps to disable
• Update-NTNXProtectionDomainStretchTimeout
• Update-NTNXStretchFailureHandling
8.5. Alerts
The following Prism alerts are generated with Metro Availability.
• Information
⁃ If a reenable operation requires full resynchronization.
• Warning
⁃ When latency between sites is more than 5 ms for 10 seconds.
⁃ When the number of entities in a Metro Availability–protected container exceeds 3,600.
⁃ When the total number of entities in live containers and associated remote snapshots
exceeds 50,000.
Note: Exceeding 50,000 entities, including snapshots, also prevents you from
enabling Metro Availability.
• Critical
⁃ When a protection domain is in a remote unreachable state.
⁃ When a protection domain is in a decoupled state.
⁃ When a protection domain fails to automatically disable.
• Licensing
⁃ Metro Availability requires the Ultimate software edition. If the Ultimate license is not
enabled, you see a license warning while configuring Metro Availability and in the Prism UI
after configuration.
9. Conclusion
Metro Availability represents two industry firsts for hyperconverged platforms: the first continuous
availability solution and the first synchronous storage replication solution. Like all Nutanix
features, management of Metro Availability is simple, intuitive, and built directly into the
software included with Acropolis. The simplified management and operation of Metro Availability
stands out compared to other more complex solutions offered with legacy three-tiered storage
architectures.
9. Conclusion | 62
Metro Availability
Appendix
About Nutanix
Nutanix makes infrastructure invisible, elevating IT to focus on the applications and services that
power their business. The Nutanix enterprise cloud software leverages web-scale engineering
and consumer-grade design to natively converge compute, virtualization, and storage into
a resilient, software-defined solution with rich machine intelligence. The result is predictable
performance, cloud-like infrastructure consumption, robust security, and seamless application
mobility for a broad range of enterprise applications. Learn more at www.nutanix.com or follow us
on Twitter @nutanix.
Appendix | 63
Metro Availability
List of Figures
Figure 1: Nutanix Data Protection Spectrum, Including Metro Availability......................... 6
Figure 13: Nondisruptive Migration with Manual or Automatic Resume Failure Handling. 27
Figure 20: Initial Cluster State Following Site Recovery with Existing Configuration....... 36
64
Metro Availability
65
Metro Availability
List of Tables
Table 1: Document Version History................................................................................... 7
66