13/08/23 12:12:42 INFO cluster.TaskSetManager: Loss was due to java.lang.ClassNotFoundException: Job$Client
at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:270)
at java.io.ObjectInputStream.resolveClass(ObjectInputStream.java:623)
at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1610)
at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1515)
at java.io.ObjectInputStream.readArray(ObjectInputStream.java:1661)
at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1342)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
at spark.Utils$.deserialize(Utils.scala:34)
at spark.SparkContext$$anonfun$objectFile$1.apply(SparkContext.scala:390)
at spark.SparkContext$$anonfun$objectFile$1.apply(SparkContext.scala:390)
at scala.collection.Iterator$$anon$21.hasNext(Iterator.scala:440)
at scala.collection.Iterator$class.foreach(Iterator.scala:772)
at scala.collection.Iterator$$anon$21.foreach(Iterator.scala:437)
at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:102)
at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:250)
at scala.collection.Iterator$$anon$21.toBuffer(Iterator.scala:437)
at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:237)
at scala.collection.Iterator$$anon$21.toArray(Iterator.scala:437)
at spark.RDD$$anonfun$1.apply(RDD.scala:461)
at spark.RDD$$anonfun$1.apply(RDD.scala:461)
at spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:613)
at spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:613)
at spark.scheduler.ResultTask.run(ResultTask.scala:77)
at spark.executor.Executor$TaskRunner.run(Executor.scala:98)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:724)
13/08/23 12:12:42 INFO cluster.TaskSetManager: Starting task 1.0:0 as TID 4 on executor 0: ip-10-234-17-229.ec2.internal (preferred)
13/08/23 12:12:42 INFO cluster.TaskSetManager: Serialized task 1.0:0 as 1579 bytes in 1 ms
13/08/23 12:12:42 INFO cluster.TaskSetManager: Lost TID 4 (task 1.0:0)
13/08/23 12:12:42 INFO cluster.TaskSetManager: Loss was due to java.lang.ClassNotFoundException: Job$Client [duplicate 1]
13/08/23 12:12:42 INFO cluster.TaskSetManager: Starting task 1.0:0 as TID 5 on executor 0: ip-10-234-17-229.ec2.internal (preferred)
13/08/23 12:12:42 INFO cluster.TaskSetManager: Serialized task 1.0:0 as 1579 bytes in 0 ms
13/08/23 12:12:42 INFO cluster.TaskSetManager: Lost TID 5 (task 1.0:0)
13/08/23 12:12:42 INFO cluster.TaskSetManager: Loss was due to java.lang.ClassNotFoundException: Job$Client [duplicate 2]
13/08/23 12:12:42 INFO cluster.TaskSetManager: Starting task 1.0:0 as TID 6 on executor 0: ip-10-234-17-229.ec2.internal (preferred)
13/08/23 12:12:42 INFO cluster.TaskSetManager: Serialized task 1.0:0 as 1579 bytes in 0 ms
13/08/23 12:12:42 INFO cluster.TaskSetManager: Lost TID 6 (task 1.0:0)
13/08/23 12:12:42 INFO cluster.TaskSetManager: Loss was due to java.lang.ClassNotFoundException: Job$Client [duplicate 3]
13/08/23 12:12:42 INFO cluster.TaskSetManager: Starting task 1.0:0 as TID 7 on executor 0: ip-10-234-17-229.ec2.internal (preferred)
13/08/23 12:12:42 INFO cluster.TaskSetManager: Serialized task 1.0:0 as 1579 bytes in 1 ms
13/08/23 12:12:43 INFO cluster.TaskSetManager: Lost TID 7 (task 1.0:0)
13/08/23 12:12:43 INFO cluster.TaskSetManager: Loss was due to java.lang.ClassNotFoundException: Job$Client [duplicate 4]
13/08/23 12:12:43 ERROR cluster.TaskSetManager: Task 1.0:0 failed more than 4 times; aborting job
13/08/23 12:12:43 INFO scheduler.DAGScheduler: Failed to run collect at Job.scala:26
[error] (run-main) spark.SparkException: Job failed: Task 1.0:0 failed more than 4 times
spark.SparkException: Job failed: Task 1.0:0 failed more than 4 times
at spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:642)
at spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:640)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:60)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
at spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:640)
at spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:303)
at spark.scheduler.DAGScheduler.spark$scheduler$DAGScheduler$$run(DAGScheduler.scala:364)
at spark.scheduler.DAGScheduler$$anon$1.run(DAGScheduler.scala:107)
13/08/23 12:12:43 INFO network.ConnectionManager: Selector thread was interrupted!
java.lang.RuntimeException: Nonzero exit code: 1
at scala.sys.package$.error(package.scala:27)
[error] {file:/root/toolbox/template/}default-376a53/compile:run: Nonzero exit code: 1
[error] Total time: 12 s, completed Aug 23, 2013 12:12:43 PM
Thank you.