HP XC System Software Administration Guide Version 3.1
The lshosts command reports the following resource information:
• ncpus
The total number of available processors within the SLURM lsf partition.
This value is calculated as the minimum value between the number of processors in all available
nodes in the lsf partition and the number of licensed cores. If total number of usable cores is 0, LIM
sets the value of ncpus to 1 and closes the host.
• maxmem
The minimum value of configured SLURM memory for all nodes.
This value represents the maximum value of memory on the node with the least memory. Running
a job on all nodes must account for this value.
This value is calculated as a minimal value over all the compute nodes.
• maxtmp
The minimum value of configured SLURM TmpDisk space for all nodes.
This value represents the maximum value of disk space on the node with the least amount of disk
space. Running a job on all nodes must account for this value.
This value is calculated as a minimal value over all the compute nodes.
The following is an example of the LSF lshosts command.
$ lshosts
HOST_NAME type model cpuf ncpus maxmem maxswp server RESOURCES
lsfhost.loc SLINUX6 Opteron8 16.0 22 2048M - Yes (slurm)
The lshosts command reports a hyphen (-) for all the other load index and resource information. Initially,
SLURM is not configured with any memory or temporary disk space, so LIM reports the default value of
1 MB for each index.
For more information on the lshosts command, see lshosts(1) .
The lsload command reports the load index, that is, the number of current login users on the LSF execution
host.
The following is an example of the LSF lsload command.
$ lsload
HOST_NAME status r15s r1m r15m ut pg ls it tmp swp mem
lsfhost.localdo ok - - - - - 1 - - - -
For more information on the lsload command, see lsload(1) .
In these examples, 22 processors on this HP XC system are available for use by LSF-HPC with SLURM.
You can verify this information, which is obtained by LSF-HPC with SLURM, with the SLURM sinfo
command:
$ sinfo --Node --long
NODELIST NODES PARTITION STATE CPUS MEMORY TMP_DISK ...
xc5n[1-10,16] 11 lsf idle 2 2048 1 ...
The output of the sinfo command shows that 11 nodes are available, and that each node has 2 processors.
The LSF lshosts command and the SLURM sinfo command both report the memory for each node as
2,048 MB. This memory value is configured for each node in /hptc_cluster/slurm/etc/slurm.conf;
it is not obtained directly from the nodes. See the SLURM documentation for more information on
configuring the slurm.conf file.
15.12 LSF-HPC with SLURM Monitoring
LSF-HPC with SLURM is monitored and controlled by Nagios using the check_lsf plug-in.
188 Managing LSF