Building Disaster Recovery Serviceguard Solutions Using Metrocluster with Continuous Access for P9000 and XP A.11.00

NOTE: In Figure 47 (page 144), the writes to the disk at DC2 are not accepted till the Delta Resync
pair is re-synchronized when using 3DC CAJ/CAJ replication. So, in the event of 3DC DR software
failing to resync the Delta Resync pair, the application package fails to come up. In this case,
though the DC1-DC2 device group pair is in SSWS state at DC2, the writes to the disk are rejected.
You must fix the issue and manually resync the Delta Resync pair using the command pairresync
-g <device group name> swaps before restarting the package.
Recovery at the third data center (DC3)
Recovery at the third data center (DC3) is automated, and is manually triggered using the
Continentalclusters command cmrecovercl.
The Three Data Center software issues the horctakeover command at DC3. This command
results in a swap takeover only if the there is continues access synchronous replication is configured
between DC1 and DC2 and the status of this sync pair at the site in the primary cluster, where the
journal device group is configured, is either SVOL-SSWS or PVOL-PSUS. In all other situations, the
horctakeover command results in an SVOL takeover.
NOTE: HP Storage XP 3DC CAJ/CAJ replication requires that the Active-CAJ pair to be deleted
to allow an application to use the data copy at DC3. As part of recovery, cmrecovercl issues the
pairsplit –S command to delete the Acive-CAJ pair on both ends of the replication (i.e. at DC3 and
either at DC2 or DC1). If it fails, then cmrecovercl issues the pairsplit –R command to delete the
Acive-CAJ pair only on the DC3 side.
Performing Recovery Group Rehearsal in the 3DC DR Solution
The Continentalclusters feature DR Rehearsal is supported in the 3DC DR Solution except when
using the Site Aware Disaster Tolerant Architecture (SADTA).
To configure and perform disaster recovery rehearsal in a 3DC environment, see Disaster Recovery
Rehearsal in Continentalclusters white paper available at http://www.hp.com/go/
hpux-serviceguard-docs —> HP Serviceguard Continentalclusters.
Failback Scenarios
Failback from Data Center 3 (DC3)
After restoring the primary cluster from the disaster, you can move the application packages back
to the primary cluster. Complete the following steps to move the application package back to DC1:
1. Verify that all the nodes in DC1, DC2 and DC3 are up and running.
2. Start DC1-DC2 cluster if it is not running.
3. Start the RAID Manager instance on each node in DC1, DC2 and DC3.
4. Verify that all the Continuous Access links are up.
5. Halt the 3DC package if it is running on DC3.
6. Recover the latest data from DC3.
7. Change the Cluster ID of all LVM and SLVM volume groups managed by the package. For
LVM volume groups, run the following commands from a node in the primary cluster to change
the cluster ID:
# vgchange -c n <vg_name>
# vgchange -c y <vg_name>
For SLVM volume groups, run the following commands from a node in the primary cluster to
change the cluster ID:
# vgchange -c n -S n <vg_name>
# vgchange -c y -S y <vg_name>
Recovery at the third data center (DC3) 145