Hello Community,
I want to run a
large simulation with around 170 Million cells. For that I have made a
Linux Cluster with four work stations. All four work stations are
identical, each have 24 cores and 48 processors, 128 GB RAM (total 512
GB). All four work stations have FDS 6.7.6 and MPI Version 3.1
installed.
I have already updated the stack size to unlimited on all four work stations.
I have created simplified three test models with 96 Meshes (1 mesh on 1 core).
Model 1 with 24 million cells run perfectly with maximum 15 GB RAM consumption on each work station.
Model
2 with 68 millions cells crashed after few minutes. At the time of
crash, the work stations had maximum RAM consumption around 52 GB
(around 40%) with the error:
Model 3 with 98 million cells crashes as soon as with the same error.
I
would like to ask if the error raised due to running out of memory,
then why the RAM consumption is only 40%. While I have already updated
the stack size to unlimited.
I am running the job using command:
mpiexec -hosts w1,w2,w3,w4 -n 96 -print-all-exitcodes fds job.fds
I would be grateful if you share your view to solve this problem. I want
to run actual simulation with around 170 million cells, it would be
great if you share your expert advices regarding handling of this kind
of huge models. The simplified fds file is in the attachment.
Regards
Rohan