Something to watch out for is the number of threads being spawned, sometimes the "molpro" start script and the clusters job scheduler get confused about how many MPI/OpenMP threads should be launched, should the threads be pinned to specific CPU cores/NUMA nodes, etc.
Once I have seen a situation where all MPI threads got pinned to 1 core by SLURM, or each MPI thread spawned 128 OpenMP threads...neither was good for performance.
I can also recommend checking how much time the molpro process spends in state "D". In HPC environments it is common that the disk system (especially the default /home or general-purpose data storage partition) is relatively low-performance, and not well suited for quantum chemistry programs that write and read from the disk intensely. Usually there is either some node-local storage, or a high-speed scratch file system on the network that should be used for Molpro.
In short: Check that the expected number of MPI and OpenMP threads are
running, check that no more than 1 Molpro thread is pinned to the same
physical core, make sure that you dont have a disk bottleneck.
Best,
Tibor Győri