I was doing more tests and nothings works, I'm a little desperate with this problem... :(
Thanks Samy for your answer but my problem exists with on single node as well as 4 nodes (64 cores). Running:
singularity exec -B /home ../../of6/openfoam6.MPIx.img mpiexec -n 16 simpleFoam -case /home/carmen/test_singularity/OpenFOAM/pruebaOF5_16cores
I can work on 16 cores correctly, but I understand that I am running the container's OMPI, likewise, I want to launch it on 64 cores for performance and for that (I understand) that I should launch it like this (as I have done with SGE):
mpirun -n $SLURM_NTASKS singularity exec ../../of6/openfoam6.x.img simpleFoam -parallel -case /home/carmen/test_singularity/OpenFOAM/pruebaOF5_16cores
And it doesn't work.
Fatih :)
xx True
xx Not at all. I can run basic programs in parallel using slurm without singularity
xx True. Only fails with Slurm and Singularity
I've tried so many combinations and cases, so I'm sure I'm forgetting some of them.
Background:
- I was working with SGE and everything worked perfectly (recently the cluster in which I work has migrated to Slurm, -is a supercomputing center, not a particular mini-cluster-)
- This same container worked with up to 64 cores without problems.
Some tests:
- Connecting directly to a node by ssh, OF works correctly with 16 cores (w/ 32 too, but having only 16 cores creates oversubscription)
- Doing salloc -N1 -n16 (or salloc -N2 -n32), and then run the program, fails.
- Both with 16 cores and with 32, executing my script (sbatch myscript.sh), fails.
- A simple "hello world" with Slurm doesn't work either, e.g. running:
mpirun -n $SLURM_NTASKS singularity exec -B /home container.img /home/carmen/test_singularity/mpi_slurm/hello_world
I attach the recipe of the last container that I created.
Thanx,
Carmen
----