CLI Guide

Table Of Contents
Description
During an inter-cluster link failure, you can permit I/O to resume at one of the two clusters: the winning cluster.
I/O remains suspended on the losing cluster.
When the inter-cluster link heals, the winning and losing clusters re-connect, and the losing cluster discovers that the winning
cluster has resumed I/O without it.
Unless explicitly configured otherwise (using the auto-resume-at-loser property), I/O remains suspended on the losing
cluster. This prevents applications at the losing cluster from experiencing a spontaneous data change.
The delay allows the administrator to shut down applications.
After stopping the applications, you can use this command to:
Resynchronize the data image on the losing cluster with the data image on the winning cluster,
Resume servicing I/O operations.
You can then safely restart the applications at the losing cluster.
Without the --force option, this command asks for confirmation before proceding, since its accidental use while applications
are still running at the losing cluster could cause applications to misbehave.
Examples
VPlexcli:/clusters/cluster-2/consistency-groups/TestCG> resume-at-loser
This may change the view of data presented to applications at cluster cluster-2. You
should first stop applications at that cluster. Continue? (Yes/No) Yes
In the following example:
The ls command shows consistency group cg1 as suspended, requires-resume-at-loser on cluster-2 after cluster-2 is
declared the losing cluster during an inter-cluster link failure.
The resume-at-loser command restarts I/O on cluster-2.
The ls command displays the change in operational status:
VPlexcli:/clusters/cluster-1/consistency-groups/cg1> ls
Attributes:
Name Value
------------------- ----------------------------------------------------------
active-clusters [cluster-1, cluster-2]
cache-mode synchronous
detach-rule no-automatic-winner
operational-status [(cluster-1,{ summary:: ok, details:: [] }),
(cluster-2,{ summary:: suspended, details:: [requires-resume-at-
loser] })]
passive-clusters []
read-only false
recoverpoint-enabled false
storage-at-clusters [cluster-1, cluster-2]
virtual-volumes [dd1_vol, dd2_vol]
visibility [cluster-1, cluster-2]
Contexts:
advanced recoverpoint
VPlexcli:/clusters/cluster-1/consistency-groups/cg1> resume-at-loser -c cluster-2
This may change the view of data presented to applications at cluster cluster-2. You
should first stop applications at that cluster. Continue? (Yes/No) Yes
VPlexcli:/clusters/cluster-1/consistency-groups/cg1> ls
Attributes:
Name Value
------------------- ----------------------------------------------------------
active-clusters [cluster-1, cluster-2]
cache-mode synchronous
detach-rule no-automatic-winner
operational-status [(cluster-1,{ summary:: ok, details:: [] }),
(cluster-2,{ summary:: ok, details:: [] })]
passive-clusters []
read-only false
recoverpoint-enabled false
Commands
109