I've been a recent hire to a company that employs accelerad for the calculation of irradiance for ray-tracing in their simulations. However, a lot of my colleagues have been noticing some differences between the GPU and CPU calculations of the illumination matrices retrieved from rfluxmtx. From validation, we have concluded that the CPU provides the correct results while the GPU provides higher values than expected as seen as in the provided figure, where red represents the GPU calculated global irradiance and blue the CPU

To provide more technical context, we command I use to call upon rfluxmtx is as follows:
accelerad_rfluxmtx: running: accelerad_rcontrib -fo+ -lw 0.0001 -ab 5 -ad 10000 -I+ -y 4032 -faa -c 1 -bn 1 -b 'if(-Dx*0-Dy*0-Dz*1,0,-1)' -m groundglow -f reinhartb.cal -p MF=1,rNx=0,rNy=0,rNz=-1,Ux=0,Uy=1,Uz=0,RHS=+1 -bn Nrbins -b rbin -m skyglow '!oconv -f -i tmp/oct_frame_-50.oct ./core/ray_tracing/skyglow_m1.rad'
Note that this the code I use when using reinhard's sky subdivision 1 and when using a different subdivision I use lw = 0.0000083 and ad = 120000.
Initially I thought it might be due to the size of the simulated plant, as noted that it could cause an issue. However, when reducing to only a size that covers 10m diameter we still got the same results.
I have just started to use this software, so I'm not yet comfortable just messing around with and I hope there are some ideas as to why this is happening.
Thanks in advance.