[slurm-users] Distribute the node resources in multiple partitions and regarding job submission script

390 views
Skip to first unread message

Purvesh Parmar

unread,
Apr 10, 2022, 7:58:25 AM4/10/22
to slurm...@lists.schedmd.com
Hello,

I have been using slurm 21.08. 

Q1 : I have 8 nodes with 2 gpus each and 128 cores with 512 GB RAM. I want to distribute the node resources in 2 partitions so that "par1" partition  will have 2 gpus with 64 cores and 256 GB ram of the node and the other partition "par 2" will have the remaining  64 cores and remaining 256 gb ram of the same node. 

par1 should be the default partition. 

I have used MaxCPUsPerNode and also listed each node in both par1 and par2 .However, while job submission, if i give par2 as the partition name and use gres:gpu, still the job is getting submitted and is going for run (in spite of par2 not having gpus).

slurm.conf (something like this)

########################
NodeName=comp1,comp2......comp8 Sockets=1 CPUs=64 CoresPerSocket=64 ThreadsPerCore=1 Gres=gpu:2
PartitionName=par1 State=UP Nodes=comp1,comp2......comp8 MaxCPUsPerNode=64
PartitionName=par1 State=UP Nodes=comp1,comp2......comp8 MaxCPUsPerNode=64
PartitionName=par2 State=UP Nodes=comp1,comp2......comp8 MaxCPUsPerNode=64
########################

Where are the things going wrong?

Q2 : How to save the job scripts permanently? I have given
SlurmdSpoolDir=/usr/local/slurm/var/spool/slurmd
AccountingStorageEnforce=safe
AccountingStoreFlags=job_script,job_env

Regards,
Purvesh

Purvesh Parmar

unread,
Apr 12, 2022, 5:49:13 AM4/12/22
to slurm...@lists.schedmd.com
Hello,

I am using slurm 21.08. I am stuck with the following. 

Q1 : I have 8 nodes with 2 gpus each and 128 cores with 512 GB RAM. I want to distribute each node's resources in 2 partitions so that "par1" partition  will have 2 gpus with 64 cores and 256 GB ram of the node and the other partition "par 2" will have the remaining  64 cores and remaining 256 gb ram and no gpus of the same node. 

Ozeryan, Vladimir

unread,
Apr 12, 2022, 6:06:20 AM4/12/22
to Slurm User Community List

1.       I don’t see where you specifying a “Default” partition (DEFAULT=yes)

2.       In “NodeName=* ” you have Gres=gpu:2 (All nodes on that line have 2 GPUs.) Create another “NodeName” line below and list your non-gpu nodes there without the GRES flag.

 

From: slurm-users <slurm-use...@lists.schedmd.com> On Behalf Of Purvesh Parmar
Sent: Tuesday, April 12, 2022 5:49 AM
To: slurm...@lists.schedmd.com
Subject: [EXT] [slurm-users] Distribute the node resources in multiple partitions and regarding job submission script

 

APL external email warning: Verify sender slurm-use...@lists.schedmd.com before clicking links or attachments

 

Reply all
Reply to author
Forward
0 new messages