You could increase 'spark.akka.askTimeout' to something higher than 10seconds.
I use 30 seconds in some of our expts ...
Been meaning to dump these someplace for a while, this is as good as any !
The other options I have are :
-Dspark.tasks.schedule.aggression=ANY
This is slightly specific to our jobs - but bottomline is, if you
have a rack aware cluster, set this to RACK_LOCAL for some non trivial
bump in performance (would be job characterstics specific though).
-Dspark.tasks.revive_periodic.interval=200
How soon to 'revive' offers - makes sense when customizing property above.
-Dspark.speculation=true
To overcome stragglers.
-Dspark.speculation.multiplier=1.2
-Dspark.speculation.quantile=0.5
These customize speculative execution and are slightly specific to our
jobs - but you might be able to play with it too.
-Dspark.storage.memoryFraction=0.25
Fraction of available memory to use for memory - our persistance is
memory and disk since we deal with RDD's which cannot be completely
hosted in available memory.
-Dspark.akka.frameSize=100
-Dspark.akka.askTimeout=30
You already have these set.
-Djava.io.tmpdir=<my_temp>
In case temp directory fills up ...
Hope these help ...
Regards,
Mridul
> --
> You received this message because you are subscribed to the Google Groups
> "Spark Users" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to
spark-users...@googlegroups.com.
> For more options, visit
https://groups.google.com/groups/opt_out.