> E-mail:
t...@itu.dk
Hej Tobias
Bulk jobs in the form of packing multiple jobs inside a single job
slot on the resources is an experimental feature and it never really
gained traction in production, so I don't know for sure if it still
works. The MiG design relies on a single job only per resource exe
node, so the bulk job implementation makes a number of hacks to work
around that assumption. The proper way to implement it would be a
redesign to remove the single job limitation and let each resource exe
node take compatible jobs until it is full, but we haven't had the
resources to work on that so far. Please also refer to the "Python-
only resources" ticket in our issue tracker:
http://code.google.com/p/migrid/issues/detail?id=61
For LRMS resources (e.g. those with PBS), a similar but more tested
behavior can be achieved by creating multiple exe nodes for the
resource. In that way you can decide how many concurrent jobs you want
to allow while leaving the actual concurrency management to PBS. If
two jobs fit at the same time PBS will run them concurrently, and
otherwise the MiG scheduler will decide if one should get locally
queued based on the queue-delay mechanism. This is of course
suboptimal compared to bulk fill by MiG but in real life it works well
enough. With the MiG queue-delay feature it is still possible to
compensate somewhat for queue load on the PBS resource. You can find
an example queue-delay script for PBS in the mig/resource/pbs-scripts
folder.
If you want to support multiple concurrent jobs even from different
MiG users you can create multiple local users on the PBS resource and
assign them to individual exe nodes. In that way you get a more
general multi-job concurrency than the MiG bulk setup where only jobs
from the same user can be bulk-executed.
We do not have much documentation on the server internals apart from
the (Wiki
http://code.google.com/p/migrid/wiki/DeveloperTopics),
README (
http://code.google.com/p/migrid/source/browse/trunk/README)
and inline comments in the code, but feel free to ask here!
Cheers, Jonas