HP XC System Software Installation Guide Version 3.2
NOTE: If a compute node did not boot up, the spconfig utility configures the node as
follows:
Configured unknown node n14 with 1 CPU and 1 MB of total memory...
After the node has been booted up, re-run the spconfig utility to configure the correct
settings.
3. If the system is using a QsNet
II
interconnect, ensure that the number of node entries in the
/opt/hptc/libelanhosts/etc/elanhosts file matches the expected number of
operational nodes in the cluster. If the number does not match, verify the status of the nodes
to ensure that they are all up and running, and re-run the spconfig utility. If the number
does not match, it is possible that the QsNet
II
network card on the missing node is not fully
operational.
4. Follow this procedure to add customizations to the SLURM configuration file:
a. Go to Appendix J (page 229) to determine the type of customizations that are available
or required. For instance, if you installed and configured SVA, SVA requires certain
SLURM customizations.
b. Use the text editor of your choice to edit the SLURM configuration file:
/hptc_cluster/slurm/etc/slurm.conf
c. Use the information in Appendix J (page 229) to customize the SLURM configuration
according to your requirements.
d. If you make changes to the slurm.conf file, save your changes and exit the text editor.
e. Update the SLURM daemons with this new information:
# scontrol reconfig
If some nodes are reported as being in the down state, see “Troubleshooting SLURM” (page 176)
for more information.
3.19.2 Perform LSF Postconfiguration Tasks
Follow this procedure to set up the LSF environment and enable LSF failover (if you assigned
the resource_management role to two or more nodes). Omit this task if you did not configure
LSF.
1. Begin this procedure as the root user on the head node.
2. Set up the LSF environment by sourcing the LSF profile file:
# . /opt/hptc/lsf/top/conf/profile.lsf
3. Verify that the LSF profile file has been sourced by finding an LSF command:
# which lsid
/opt/hptc/lsf/top/6.2/linux2.6-glibc2.3-x86_64-slurm/bin/lsid
This sample output was obtained from an HP ProLiant server. Thus, the directory name
linux2.6-glibc2.3-x86_64-slurm is included in the path (the string x86_64 signifies
a Xeon- or Opteron-based architecture). The string ia64 is included in the directory name
for HP Integrity servers. The string slurm exists in the path only if LSF-HPC with SLURM
is configured.
Remainder Applies to LSF-HPC with SLURM:
The remainder of this procedure applies to LSF-HPC with SLURM. If standard LSF is
configured, omit the remaining steps.
4. If you assigned two or more nodes with the resource_management role and want to
enable LSF failover, enter the following command; otherwise, proceed to step 5.
108 Configuring and Imaging the System