...various WARNINGS from CMS that don't bother me reported from each processor (16 repeats in the test shown here, only 2 lines shown) followed by an mpi error with changing process names and exit codes (e.g. below):
WARNING: The fill value you have entered in nest_x.nml is not the one used by cms (1.2676506E30), is that what you intend?
WARNING: The fill value you have entered in nest_x.nml is not the one used by cms (1.2676506E30), is that what you intend?
-------------------------------------------------------
Primary job terminated normally, but 1 process returned
a non-zero exit code.. Per user-direction, the job has been aborted.
-------------------------------------------------------
--------------------------------------------------------------------------
mpirun detected that one or more processes exited with non-zero status, thus causing
the job to be terminated. The first process to do so was:
Process name: [[17771,1],4]
Exit code: 2
compute-1-10.26319ipath_userinit: assign_context command failed: Network is down
--------------------------------------------------------------------------
PSM was unable to open an endpoint. Please make sure that the network link is
active on the node and the hardware is functioning.
Error: Could not detect network connectivity
--------------------------------------------------------------------------
--------------------------------------------------------------------------
It looks like MPI_INIT failed for some reason; your parallel process is
likely to abort. There are many reasons that a parallel process can
fail during MPI_INIT; some of which are due to configuration or environment
problems. This failure appears to be an internal failure; here's some
additional information (which may only be relevant to an Open MPI
developer):
PML add procs failed
--> Returned "Error" (-1) instead of "Success" (0)
--------------------------------------------------------------------------
[[47159,1],0]: A high-performance Open MPI point-to-point messaging module
was unable to find any relevant network interfaces:
Module: OpenFabrics (openib)
Host: node925
Another transport will be used instead, although this may result in
lower performance.
--------------------------------------------------------------------------