HP-MPI Version 2.3.1 for Linux Release Note
Table Of Contents
- HP-MPI V2.3.1 for Linux Release Note
- Table of Contents
- 1 Information About This Release
- 2 New or Changed Features in V2.3.1
- 3 New or Changed Features in V2.3
- 3.1 Options Supported Only on HP Hardware
- 3.2 System Check
- 3.3 Default Message Size Changed For -ndd
- 3.4 MPICH2 Compatibility
- 3.5 Support for Large Messages
- 3.6 Redundant License Servers
- 3.7 License Release/Regain on Suspend/Resume
- 3.8 Expanded Functionality for -ha
- 3.8.1 Support for High Availability on InfiniBand Verbs
- 3.8.2 Highly Available Infrastructure (-ha:infra)
- 3.8.3 Using MPI_Comm_connect and MPI_Comm_accept
- 3.8.4 Using MPI_Comm_disconnect
- 3.8.5 Instrumentation and High Availability Mode
- 3.8.6 Failure Recover (-ha:recover)
- 3.8.7 Network High Availability (-ha:net)
- 3.8.8 Failure Detection (-ha:detect)
- 3.8.9 Clarification of the Functionality of Completion Routines in High Availability Mode
- 3.9 Enhanced InfiniBand Support for Dynamic Processes
- 3.10 Singleton Launching
- 3.11 Using the -stdio=files Option
- 3.12 Using the -stdio=none Option
- 3.13 Expanded Lightweight Instrumentation
- 3.14 The api option to MPI_INSTR
- 3.15 New mpirun option -xrc
- 4 Known Issues and Workarounds
- 4.1 Running on iWarp Hardware
- 4.2 Running with Chelsio uDAPL
- 4.3 Mapping Ranks to a CPU
- 4.4 OFED Firmware
- 4.5 Spawn on Remote Nodes
- 4.6 Default Interconnect for -ha Option
- 4.7 Linking Without Compiler Wrappers
- 4.8 Locating the Instrumentation Output File
- 4.9 Using the ScaLAPACK Library
- 4.10 Increasing Shared Memory Segment Size
- 4.11 Using MPI_FLUSH_FCACHE
- 4.12 Using MPI_REMSH
- 4.13 Increasing Pinned Memory
- 4.14 Disabling Fork Safety
- 4.15 Using Fork with OFED
- 4.16 Memory Pinning with OFED 1.2
- 4.17 Upgrading to OFED 1.2
- 4.18 Increasing the nofile Limit
- 4.19 Using appfiles on HP XC Quadrics
- 4.20 Using MPI_Bcast on Quadrics
- 4.21 MPI_Issend Call Limitation on Myrinet MX
- 4.22 Terminating Shells
- 4.23 Disabling Interval Timer Conflicts
- 4.24 libpthread Dependency
- 4.25 Fortran Calls Wrappers
- 4.26 Bindings for C++ and Fortran 90
- 4.27 Using HP Caliper
- 4.28 Using -tv
- 4.29 Extended Collectives with Lightweight Instrumentation
- 4.30 Using -ha with Diagnostic Library
- 4.31 Using MPICH with Diagnostic Library
- 4.32 Using -ha with MPICH
- 4.33 Using MPI-2 with Diagnostic Library
- 4.34 Quadrics Memory Leak
- 5 Installation Information
- 6 Licensing Information
- 7 Additional Product Information
4.18 Increasing the nofile Limit
The nofile limit on large Linux clusters needs to be increased in /etc/security/
limits.conf
* soft nofile 1024
For larger clusters, HP recommends a setting of at least:
• 2048 for clusters of 1900 cores or fewer
• 4096 for clusters of 3800 cores or fewer
• 8192 for clusters of 7600 cores or fewer
4.19 Using appfiles on HP XC Quadrics
To use appfiles on HP XC Quadrics clusters, set MPI_USESRUN=1. The appfile can only
differ in host name and rank count.
4.20 Using MPI_Bcast on Quadrics
On Quadrics interconnected clusters, the repeated use of MPI_Bcast within a tight
loop can cause an application to fail with the following Elan trap message:
ELAN TRAP - 0 0 CPROC - Bad Trap
Status=lbb40005 CommandProcSendTransExpected Command=200000201
Setting the environment variable LIBELAN_GROUP_SANDF=0 disables the latest “Store
and Forward” broadcast optimization from Quadrics while preserving all the other
optimized collectives.
4.21 MPI_Issend Call Limitation on Myrinet MX
Some earlier versions of Myrinet MX have a known resource limitation involving
outstanding MPI_Issend() calls. If more than 128 MPI_Issend() calls are issued
and not yet matched, further MX communication can hang. The only known
workaround is to have your application issue less than 128 unmatched MPI_Issend()
calls at a time. This limitation is fixed in versions 1.1.8 and later.
4.22 Terminating Shells
When a foreground HP-MPI job is run from a shell window, if the shell is terminated,
the shell sends signal SIGHUP to the mpirun process and its underlying ssh processes,
thus killing the entire job.
When a background HP-MPI job is run and the shell is terminated, the job might
continue depending on the actual shell used. For /bin/bash, the job is killed. For
/bin/sh and /bin/ksh, the job continues. If nohup is used when launching the job,
only background ksh jobs can continue. This behavior might vary depending on your
system.
4.18 Increasing the nofile Limit 31