The implementation works on modern GPUs at current state. The main limitation is the required support of double precision floating point operations.
We expect a total speedup up to a factor of 20/40 at DP/SP with small code customizations on a more recent GPU which has full native support of DP calculations like e.g., the Nvidia Tesla C2050/C2070.
- How is the accuracy of the result impacted by the GPU runs (due to SP)?
- Is there a way to take advantage of the DP performance of Tesla GPUs, as is stated in Huntemann et al. 2011, in order to speed up the calculations significantly?
- Is there a way to utilize two (2) or more GPUs in one (1) Motherboard, in order to speed up the process?
Not at the moment. There is a separate issue for that -
https://github.com/adda-team/adda/issues/185 , but that seems
rather complicated to implement. It may be possible to run two
instances of ADDA in parallel (two independent runs), each
choosing a different GPU. But that is probably not what you want.
Maxim.
P.S. This answer has been forwarded to your e-mail address for
your convenience. However, if you want to continue the discussion
please reply to the group's e-mail. You are also advised to check
the corresponding discussion thread at
http://groups.google.com/group/adda-discuss and/or subscribe to it
to receive automatic notifications, since other answers/comments
will not be necessarily forwarded to your e-mail.