Hi guys,
I've got a configuration issue running pyspark from hue notebooks examples.
Python spark jobs sent to yarn are failing with an ImportError: No module named pyspark.shell
error.
Googling about it I just see that SPARK_HOME
and PYTHONPATH
must be set and contain the path to pyspark.
I think I have done this right, by adding the required exports to spark-env template
on ambari spark service configuration and restarting the service. The exports are now present on spark-env.sh
on every node, but the issue remains.
I don't know if this is relevant, but I am sending the task using livy from hue's notebook. Both services are running on one of the nodes of the cluster. I have also try adding the exports to /etc/profiles.d/spark-env.sh
on each node and restarting both services.
Python spark jobs sent to yarn are failing with an Looking at other answers I just see that I think I have done this right, by adding the required exports to I don't know if this is relevant, but I am sending the task using livy from hue's notebook. Both services are running on one of the nodes of the cluster. I have also try adding the exports to Any ideas on how to fix it? |
. Alberto Gael Abadin Martinez www.imatia.com . |
||||
|
||||
. Este mensaje, y en su caso, cualquier fichero anexo al mismo, puede
contener información confidencial, siendo para uso exclusivo del
destinatario. Queda prohibida su divulgación copia o distribución a
terceros sin la autorización expresa del remitente. Si usted ha recibido
este mensaje erróneamente, se ruega lo notifique al remitente y proceda
a su borrado. Gracias por su colaboración. . |
||||
--
You received this message because you are subscribed to the Google Groups "Hue-Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to hue-user+u...@cloudera.org.