Zeppelin - Spark - HDFS

501 views
Skip to first unread message

Gigi

unread,
Apr 30, 2015, 5:55:01 AM4/30/15
to zeppelin-...@googlegroups.com
Hi all,

I'm using Zeppelin installed on one of my CDH cluster node. So I have access to HDFS and Spark.

For instance I can do in my Zeppelin notebook:

val temp = sc.textFile("hdfs://user/temp/folder/myfile.csv")
temp.count


My question is: is it possible to install Zeppelin on a separate machine, without hadoop/spark installed on it, and link it to my CDH cluster?

I've already tried but I have to specify the NameNode address in the path:

val temp = sc.textFile("hdfs://name-node-server:port/user/temp/folder/myfile.csv")
temp.count


Is there a way to avoid this?

moon soo Lee

unread,
May 1, 2015, 8:39:00 PM5/1/15
to Gigi, zeppelin-...@googlegroups.com
Hi, 

We have moved to apache. This mailing list is deprecated.

Yes you can install Zeppelin without installing Hadoop/Spark.

In that case, to avoid full URL for path in HDFS, copy core-site.xml and hdfs-site.xml into ZEPPELIN_HOME/conf. 

Thanks,

moon

--
You received this message because you are subscribed to the Google Groups "zeppelin-developers" group.
To unsubscribe from this group and stop receiving emails from it, send an email to zeppelin-develo...@googlegroups.com.
To post to this group, send email to zeppelin-...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
Reply all
Reply to author
Forward
0 new messages