export SPARK_VERSION="1.2.0-cdh5.3.3"
I then recreated the uber jar file: ./sbt "project spark" assembly
Hi guys,OK, so I have the Mesos-leader and one Mesos-follower up on AWS. Running the example of parallelizing a list of numbers and collecting a filtered list back to the driver (in the README file of the GitHub repo) works fine. When running the attached ingestion script, the rasters fail to be ingested into Accumulo. From the command line, if I run something like: hadoop fs -ls /accumulo, I get back a directory listing. I was able to create directories and place files in HDFS manually. I believe that the issue is with the value for the CATALOG variable on L22 of the attached file. The current CATALOG value is 'hdfs://namenode.service.geotrellis-spark.internal/accumulo/data/catalog' This directory exists in HDFS and is empty.Any assistance would be appreciated.Thanks,PittBelow is the entire output from the script.
ubuntu@ip-10-0-1-42:~$ python ./scripts/raster_processing.pyInput file size is 2591, 25020...10...20...30...40...50...60...70...80...90...100 - done.Spark assembly has been built with Hive, including Datanucleus jars on classpath13:55:55 Slf4jLogger: Slf4jLogger started13:55:55 Remoting: Starting remoting
13:55:55 Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@zookeeper.service.geotrellis-spark.internal:42507]
--
You received this message because you are subscribed to the Google Groups "geotrellis-user" group.
To unsubscribe from this group and stop receiving emails from it, send an email to geotrellis-us...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
Hi guys,OK, so I have the Mesos-leader and one Mesos-follower up on AWS. Running the example of parallelizing a list of numbers and collecting a filtered list back to the driver (in the README file of the GitHub repo) works fine. When running the attached ingestion script, the rasters fail to be ingested into Accumulo. From the command line, if I run something like: hadoop fs -ls /accumulo, I get back a directory listing. I was able to create directories and place files in HDFS manually. I believe that the issue is with the value for the CATALOG variable on L22 of the attached file. The current CATALOG value is 'hdfs://namenode.service.geotrellis-spark.internal/accumulo/data/catalog' This directory exists in HDFS and is empty.Any assistance would be appreciated.Thanks,PittBelow is the entire output from the script.
ubuntu@ip-10-0-1-42:~$ python ./scripts/raster_processing.pyInput file size is 2591, 25020...10...20...30...40...50...60...70...80...90...100 - done.Spark assembly has been built with Hive, including Datanucleus jars on classpath13:55:55 Slf4jLogger: Slf4jLogger started13:55:55 Remoting: Starting remoting
13:55:55 Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@zookeeper.service.geotrellis-spark.internal:42507]
>> :[akka.tcp://sparkDriver@zookeeper.service.geotrellis-spark.internal:42507]
Hi guys,OK, so I have the Mesos-leader and one Mesos-follower up on AWS. Running the example of parallelizing a list of numbers and collecting a filtered list back to the driver (in the README file of the GitHub repo) works fine. When running the attached ingestion script, the rasters fail to be ingested into Accumulo. From the command line, if I run something like: hadoop fs -ls /accumulo, I get back a directory listing. I was able to create directories and place files in HDFS manually. I believe that the issue is with the value for the CATALOG variable on L22 of the attached file. The current CATALOG value is 'hdfs://namenode.service.geotrellis-spark.internal/accumulo/data/catalog' This directory exists in HDFS and is empty.Any assistance would be appreciated.Thanks,PittBelow is the entire output from the script.
ubuntu@ip-10-0-1-42:~$ python ./scripts/raster_processing.pyInput file size is 2591, 25020...10...20...30...40...50...60...70...80...90...100 - done.Spark assembly has been built with Hive, including Datanucleus jars on classpath13:55:55 Slf4jLogger: Slf4jLogger started13:55:55 Remoting: Starting remoting
13:55:55 Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@zookeeper.service.geotrellis-spark.internal:42507]
--
You received this message because you are subscribed to the Google Groups "geotrellis-user" group.
To unsubscribe from this group and stop receiving emails from it, send an email to geotrellis-us...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
15/04/09 16:29:56 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@zookeeper.service.geotrellis-spark.internal:38369]15/04/09 16:29:56 INFO Remoting: Remoting now listens on addresses: [akka.tcp://sparkDriver@zookeeper.service.geotrellis-spark.internal:38369]
Hi guys,OK, so I have the Mesos-leader and one Mesos-follower up on AWS. Running the example of parallelizing a list of numbers and collecting a filtered list back to the driver (in the README file of the GitHub repo) works fine. When running the attached ingestion script, the rasters fail to be ingested into Accumulo. From the command line, if I run something like: hadoop fs -ls /accumulo, I get back a directory listing. I was able to create directories and place files in HDFS manually. I believe that the issue is with the value for the CATALOG variable on L22 of the attached file. The current CATALOG value is 'hdfs://namenode.service.geotrellis-spark.internal/accumulo/data/catalog' This directory exists in HDFS and is empty.Any assistance would be appreciated.Thanks,PittBelow is the entire output from the script.
ubuntu@ip-10-0-1-42:~$ python ./scripts/raster_processing.pyInput file size is 2591, 25020...10...20...30...40...50...60...70...80...90...100 - done.Spark assembly has been built with Hive, including Datanucleus jars on classpath13:55:55 Slf4jLogger: Slf4jLogger started13:55:55 Remoting: Starting remoting
13:55:55 Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@zookeeper.service.geotrellis-spark.internal:42507]