HP XC System Software Administration Guide Version 3.0

Your HP XC system is configured initially with all compute nodes in a single SLURM partition, called lsf.
In some situations, you might want to remove some nodes from the lsf partition and manage them directly
with SLURM, submitting jobs to those nodes with the srun --partition=partition-name command.
LSF manages only one partition. If present, the LSF partition must have these characteristics specified:
PartitionName=lsf
RootOnly=YES
Shared=FORCE
Assigning YES to the RootOnly= characteristic means that only the superuser (root) can create allocations
for normal user jobs.
The Shared=FORCE characteristic ensures that more than one job can run on the same node. LSF-HPC uses
this facility to support preemption and scheduling multiple serial jobs on the same node (node sharing). The
FORCE value makes all nodes in the partition available for sharing without user means of disabling it.
Do not configure the MaxNodes, MaxTime, or MinNodes parameters in an LSF partition; these parameters
conflict with LSF scheduling decisions.
Note
The root-only lsf partition is provided for submitting and managing jobs through an interaction of SLURM
and LSF. If you intend to use SLURM independently from LSF, consider configuring a separate SLURM partition
for that purpose.
Table 12-2. describes the SLURM partition characteristics available on HP XC systems.
Table 12-2. SLURM Partition Characteristics SLURM Partition Characteristics
DescriptionCharacteristic
List of nodes that constitute this partition.Nodes
A comma-separated list of group IDs that are allowed to use the partition.
The default is ALL, which allows all users to access the partition.
AllowGroups
A text string, which must be either YES or NO.
The value YES indicates that the corresponding partition is the default when
users submit jobs without specifying a partition.
Default
A text string, which must be either YES or NO.
The text string YES specifies that only the superuser (root) can use this
partition.
RootOnly
Specifies the maximum count of nodes that can be allocated to any single
job.
The default is unlimited or -1.
MaxNodes
Specifies the maximum time limit (in minutes) allowed for jobs in this partition.
The default is unlimited or -1.
MaxTime
Specifies the minimum number of nodes that can be allocated to any single
job.
The default is 1.
MinNodes
A text string that indicates whether node sharing for jobs is allowed:Shared
YES The node may be shared or not, depending on the allocation.
FORCE The node is always available to be shared.
NO The node is never available to be shared.
The state of the partition.
The possible values are UP or DOWN.
State
Consider a system that has 128 nodes. The following line in the /hptc_cluster/slurm/etc/slurm.conf
file indicates that partition lsf controls all 128 nodes:
PartitionName=lsf RootOnly=yes Shared=FORCE Nodes=n[1-128]
Node Characteristics 105