Tuantuan
unread,Jun 27, 2013, 12:56:05 PM6/27/13Sign in to reply to author
Sign in to forward
You do not have permission to delete messages in this group
Sign in to report message
Either email addresses are anonymous for this group or you need the view member email addresses permission to view the original message
to spark...@googlegroups.com
Hi,
I try to initialize an RDD and run Spark locally through
JavaSparkContext sc = new JavaSparkContext("local", "Simple Job", "C:/Users/tuantuan/Desktop/spark-0.7.2",
"C:/Users/tuantuan/Desktop/spark-0.7.2/core/target/spark-core-assembly-0.7.2");
JavaRDD<String> testData= sc.textFile("data.txt");
System.out.println(testData.count());
but I get following errors:
13/06/27 12:52:16 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
13/06/27 12:52:16 WARN snappy.LoadSnappy: Snappy native library not loaded
...
13/06/27 12:52:16 ERROR local.LocalScheduler: Exception in task 0
java.io.IOException: No FileSystem for scheme: C
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1383)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254)
at spark.Utils$.fetchFile(Utils.scala:172)
at spark.scheduler.local.LocalScheduler$$anonfun$updateDependencies$6.apply(LocalScheduler.scala:131)
at spark.scheduler.local.LocalScheduler$$anonfun$updateDependencies$6.apply(LocalScheduler.scala:129)
at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:743)
at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:95)
at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:95)
at scala.collection.Iterator$class.foreach(Iterator.scala:772)
at scala.collection.mutable.HashTable$$anon$1.foreach(HashTable.scala:157)
at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:190)
at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:45)
at scala.collection.mutable.HashMap.foreach(HashMap.scala:95)
at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:742)
at spark.scheduler.local.LocalScheduler.updateDependencies(LocalScheduler.scala:129)
at spark.scheduler.local.LocalScheduler.runTask$1(LocalScheduler.scala:69)
at spark.scheduler.local.LocalScheduler$$anon$1.run(LocalScheduler.scala:49)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask$Sync.innerRun(Unknown Source)
at java.util.concurrent.FutureTask.run(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.lang.Thread.run(Unknown Source)
13/06/27 12:52:16 INFO scheduler.DAGScheduler: Failed to run count at example1.java:24
Exception in thread "main" spark.SparkException: Job failed: ResultTask(0, 0) failed: ExceptionFailure(java.io.IOException: No FileSystem for scheme: C)
at spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:642)
at spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:640)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:60)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
at spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:640)
at spark.scheduler.DAGScheduler.handleTaskCompletion(DAGScheduler.scala:601)
at spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:300)
at spark.scheduler.DAGScheduler.spark$scheduler$DAGScheduler$$run(DAGScheduler.scala:364)
at spark.scheduler.DAGScheduler$$anon$1.run(DAGScheduler.scala:107)
Does anyone know how to solve it?
Thanks.