Imran Rashid
unread,Jan 8, 2013, 12:31:02 AM1/8/13Sign in to reply to author
Sign in to forward
You do not have permission to delete messages in this group
Either email addresses are anonymous for this group or you need the view member email addresses permission to view the original message
to spark...@googlegroups.com
Every so often I want to abort some long-running computation in spark, but I would like to leave the spark context & cached data alone. Eg., sometimes I've waited a while for a bunch of data to get loaded in memory, and some intermediate calculations to happen, but then I run something that either (a) takes a really long time due to a coding error on my part or (b) fails, but takes a long time for spark to kill the job (or maybe spark never realized the stage has died). What is the right way for me to kill just that one running computation?
of course I'd need to have some error handling in my code to deal w/ this, but ideally there would be some JobKilledException or something which I could then handle, eg.
val sc = ...
val myBigRdd = ...
try{
//run unsafe code here
} catch {
case JobKilledException => ...
}
Anyway to do this or something similar?
thanks,
Imran