Hi Matthieu & John
this is the backtrace of slurmctld during the slowdown
(gdb) bt
#0 0x00007fb0e8b1e69d in poll () from /lib64/libc.so.6
#1 0x00007fb0e8617bfa in sss_cli_make_request_nochecks () from /lib64/libnss_sss.so.2
#2 0x00007fb0e86185a3 in sss_nss_make_request () from /lib64/libnss_sss.so.2
#3 0x00007fb0e8619104 in _nss_sss_getpwnam_r () from /lib64/libnss_sss.so.2
#4 0x00007fb0e8aef07d in getpwnam_r@@GLIBC_2.2.5 () from /lib64/libc.so.6
#5 0x00007fb0e9360986 in _getpwnam_r (result=<optimized out>, bufsiz=<optimized out>, buf=<optimized out>, pwd=<optimized out>, name=<optimized out>) at uid.c:73
#6 uid_from_string (name=0x1820e41 "g2bottin", uidp=uidp@entry=0x7fff07f03a6c) at uid.c:111
#7 0x000000000043587d in get_group_members (group_name=0x10ac500 "g2") at groups.c:139
#8 0x000000000047525a in _get_groups_members (group_names=<optimized out>) at partition_mgr.c:2006
#9 0x0000000000475505 in _update_part_uid_access_list (x=0x7fb03401e650, arg=0x7fff07f13bf4) at partition_mgr.c:1930
#10 0x00007fb0e92ab675 in list_for_each (l=0x1763e50, f=f@entry=0x4754d8 <_update_part_uid_access_list>, arg=arg@entry=0x7fff07f13bf4) at list.c:420
#11 0x000000000047911a in load_part_uid_allow_list (force=1) at partition_mgr.c:1971
#12 0x0000000000428e5c in _slurmctld_background (no_data=0x0) at controller.c:1911
#13 main (argc=<optimized out>, argv=<optimized out>) at controller.c:601
As Matthieu said it seems something related to SSS daemon.
However we don't notice any slowdown due to SSSd in our environment.
As I told you before, we are just testing SLURM on a small 100 nodes cluster before going into production with about 6000 nodes next Wednesday.
At present the other nodes are managed by PBSPro and the 2 PBS servers are running on the same nodes as the SLURM controllers.
PBS queues are also configured with users/groups ACLs and we never noticed any similar slowdown.
Moreover, only 3 SLURM partitions have the AllowGroups ACL
[root@mgmt01 slurm]# grep AllowGroups slurm.conf
PartitionName=bdw_fua_gwdbg Nodes=r040c03s0[1,2] Default=NO DefMemPerCPU=3000 DefaultTime=00:30:00 MaxTime=00:30:00 State=UP QOS=bdw_fua_gwdbg DenyQos=bdw_qos_special AllowGroups=g2
PartitionName=bdw_fua_gw Nodes=r040c03s0[1,2] Default=NO DefMemPerCPU=3000 DefaultTime=00:30:00 MaxTime=48:00:00 State=UP QOS=bdw_fua_gw DenyQos=bdw_qos_special AllowGroups=g2
PartitionName=bdw_fua_gwg2 Nodes=r040c03s0[1,2] Default=NO DefMemPerCPU=3000 DefaultTime=00:30:00 MaxTime=168:00:00 State=UP QOS=bdw_fua_gwg2 DenyQos=bdw_qos_special AllowGroups=g2
So why does the UID-GID mapping take so long?
@John: we defined many partitions on the same nodes but in the production cluster they will be more or less split across the 6K nodes.
thank you very much