val job: WordCountJob = new WordCountJob(arguments)
val flow = job.buildFlow
flow.complete()
}
The ArchiBuntu-E5550 is my local machine.
I have a CDH5 cluster running, all nodes are active
When I run the code as a scala app, this is the error:
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/home/ken/.ivy2/cache/ch.qos.logback/logback-classic/jars/logback-classic-1.1.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/home/ken/.ivy2/cache/org.slf4j/slf4j-log4j12/jars/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See
http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [ch.qos.logback.classic.util.ContextSelectorStaticBinder]
08:14:38.020 [main] INFO o.a.h.conf.Configuration.deprecation -
fs.default.name is deprecated. Instead, use fs.defaultFS
08:14:38.419 [main] INFO o.a.h.conf.Configuration.deprecation -
fs.default.name is deprecated. Instead, use fs.defaultFS
08:14:38.425 [main] INFO c.flow.hadoop.util.HadoopUtil - resolving application jar from found main method on: scala.App$class
Exception in thread "main" java.lang.NullPointerException
at org.apache.hadoop.util.ClassUtil.findContainingJar(ClassUtil.java:43)
at org.apache.hadoop.mapred.JobConf.setJarByClass(JobConf.java:533)
at cascading.flow.hadoop.planner.HadoopPlanner.initialize(HadoopPlanner.java:221)
at cascading.flow.FlowConnector.connect(FlowConnector.java:457)
at com.twitter.scalding.ExecutionContext$class.buildFlow(ExecutionContext.scala:47)
at com.twitter.scalding.ExecutionContext$$anon$1.buildFlow(ExecutionContext.scala:94)
at com.twitter.scalding.Job$$anonfun$buildFlow$1.apply(Job.scala:230)
at com.twitter.scalding.Job$$anonfun$buildFlow$1.apply(Job.scala:230)
at scala.util.Success.flatMap(Try.scala:231)
at com.twitter.scalding.Job.buildFlow(Job.scala:230)
at WordCountJobRunner$.delayedEndpoint$WordCountJobRunner$1(WordCountJob.scala:27)
at WordCountJobRunner$delayedInit$body.apply(WordCountJob.scala:17)
at scala.Function0$class.apply$mcV$sp(Function0.scala:34)
at scala.runtime.AbstractFunction0.apply$mcV$sp(AbstractFunction0.scala:12)
at scala.App$$anonfun$main$1.apply(App.scala:76)
at scala.App$$anonfun$main$1.apply(App.scala:76)
at scala.collection.immutable.List.foreach(List.scala:381)
at scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:35)
at scala.App$class.main(App.scala:76)
at WordCountJobRunner$.main(WordCountJob.scala:17)
at WordCountJobRunner.main(WordCountJob.scala)
I think I've tried alle possible examples of this "wordcount on hadoop via scalding" I can google.