Serviceguard NFS Toolkit A.11.11.06, A.11.23.05 and A.11.31.05 Administrator's Guide HP-UX 11i v1, v2, and v3
Table Of Contents
- Serviceguard NFS Toolkit A.11.11.06, A.11.23.05 and A.11.31.05 Administrator's Guide
- Table of Contents
- 1 Overview of Serviceguard NFS
- Limitations of Serviceguard NFS
- Overview of Serviceguard NFS Toolkit A.11.31.05 with Serviceguard A.11.18 (or later) and Veritas Cluster File System Support
- Overview of the Serviceguard NFS Modular Package
- Overview of the NFS File Lock Migration Feature
- Overview of NFSv4 File Lock Migration Feature
- Overview of Serviceguard NFS with Serviceguard A.11.17 Support
- Integrating Support for Cluster File Systems into Serviceguard NFS Toolkit
- Overview of Cluster File Systems in Serviceguard NFS Toolkit
- Limitations and Issues with the current CFS implementation
- Supported Configurations
- How the Control and Monitor Scripts Work
- 2 Installing and Configuring Serviceguard NFS Legacy Package
- Installing Serviceguard NFS Legacy Package
- Before Creating a Serviceguard NFS Legacy Package
- Configuring a Serviceguard NFS Legacy Package
- Copying the Template Files
- Editing the Control Script (nfs.cntl)
- Editing the NFS Control Script (hanfs.sh)
- Editing the File Lock Migration Script (nfs.flm)
- Editing the NFS Monitor Script (nfs.mon)
- Editing the Package Configuration File (nfs.conf)
- Configuring Server-to-Server Cross-Mounts (Optional)
- Creating the Cluster Configuration File and Bringing Up the Cluster
- Configuring Serviceguard NFS Legacy Package over CFS Packages
- 3 Installing and Configuring Serviceguard NFS Modular Package
- Installing Serviceguard NFS Modular Package
- Before Creating a Serviceguard NFS Modular Package
- Configuring a Serviceguard NFS Modular Package
- Configuring Serviceguard NFS Modular Package over CFS Packages
- 4 Migration of Serviceguard NFS Legacy Package to Serviceguard NFS Modular Package
- 5 Sample Configurations for Legacy Package
- Example One - Three-Server Mutual Takeover
- Example Two - One Adoptive Node for Two Packages with File Lock Migration
- Cluster Configuration File for Adoptive Node for Two Packages with File Lock Migration
- Package Configuration File for pkg01
- NFS Control Scripts for pkg01
- NFS File Lock Migration and Monitor Scripts for pkg01
- Package Configuration File for pkg02
- NFS Control Scripts for pkg02
- NFS File Lock Migration and Monitor Scripts for pkg02
- Example Three - Three-Server Cascading Failover
- Example Four - Two Servers with NFS Cross-Mounts
- 6 Sample Configurations for Modular Package
- Index

There is a short time, after one package has failed over but before the cmmodpkg command has
executed, when the other package can fail over and the host will adopt it. In other words, if two
packages fail over at approximately the same time, a host may adopt both packages, even though
the package control option is specified.
See “Example Two - One Adoptive Node for Two Packages with File Lock Migration” (page 77)
for a sample configuration using the package control option.
NOTE: The NFS specific variables have been moved to NFS specific control script in Serviceguard
NFS Toolkit with the A.11.11.02 and A.11.00.05 releases. See section “Editing the NFS Control
Script (hanfs.sh) ” (page 31) for the details.
Editing nfs.cntl for NFS Toolkit A.11.00.04 and A.11.11.01 (or lower)
For Serviceguard NFS Toolkit A.11.00.04 or lower for HP-UX 11.0 or Serviceguard NFS Toolkit
A.11.11.01 or lower for HP-UX 11i v1 using Serviceguard A.11.09 (or below) framework.
Example steps:
1. Create a separate VG[n] variable for each volume group.
VG[0]=/dev/vg01 VG[1]=/dev/vg02 ...
2. Create a separate LV[n] and FS[n] variable for each volume group and file system that
will be mounted on the server:.
LV[0]=/dev/vg01/lvol1;FS[0]=/ha_root
LV[1]=/dev/vg01/lvol2;FS[1]=/users/scaf
LV[2]=/dev/vg02/lvol1;FS[2]=/ha_data
3. Create a separate XFS[n] variable for each NFS directory to be exported. Specify the directory
name and any export options.
XFS[0]="/ha_root" XFS[1]="/users/scaf" XFS[2]="-o ro /ha_data"
Do not configure these exported directories in the /etc/exports file. When an NFS server
boots up, it attempts to export all file systems in its /etc/exports file. If those file systems
are not currently present on the NFS server node, the node cannot boot properly. This
happens if the server is an adoptive node for a file system, and the file system is available
on the server only after failover of the primary node.
4. Specify the IP address for the package and the address of the subnet to which the IP address
belongs.
IP[0]=15.13.114.243 SUBNET[0]=15.13.112.0
The IP address you specify is the relocatable IP address for the package. NFS clients that
mount the file systems in the package will use this IP address to identify the server. You
should configure a name for this address in the DNS or NIS database, or in the /etc/hosts
file.
5. If you want to run the NFS monitor script , set the NFS_SERVICE_NAME variable to the
value of the SERVICE_NAME variable in the package configuration file. Each package must
have a unique service name.
NFS_SERVICE_NAME[0]=nfs1.monitor
If you do not want to run the NFS monitor script, comment out the NFS_SERVICE_NAME
and NFS_SERVICE_CMD variables:
# NFS_SERVICE_NAME[0]=nfs.monitor
# NFS_SERVICE_CMD[0]=/etc/cmcluster/nfs/nfs.mon
By default, the NFS_SERVICE_NAME and NFS_SERVICE_CMD variables are commented
out, and the NFS monitor script is not run.
You do not have to run the NFS monitor script. If your NFS package configuration file
specifies PKG_SWITCHING_ENABLED YES and NET_SWITCHING_ENABLED YES (the
30 Installing and Configuring Serviceguard NFS Legacy Package