I am having a 3 node cluster and getting below error when I ran a machine learning algorithm. I have also provided the spark-env.sh configurations please let me know how to fix this error.
Error
16/11/01 17:36:06 ERROR TaskSchedulerImpl: Lost executor 4 on
cassandra104-01.dev.wgu.edu: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
16/11/01 17:36:06 WARN TaskSetManager: Lost task 4.3 in stage 33.0 (TID 130,
cassandra104-01.dev.wgu.edu): ExecutorLostFailure (executor 4 exited caused by one of the running tasks) Reason: Remote RPC client disassociated. Likely due to containers exceeding thresholds, or network issues. Check driver logs for WARN messages.
spark-env.sh
export JAVA_HOME=/usr/lib/jvm/java
export SPARK_MASTER_IP=10.20.20.165
export SPARK_WORKER_CORES=2
export SPARK_WORKER_MEMORY=2g
export SPARK_WORKER_INSTANCES=2
export SPARK_LOCAL_IP=10.20.20.165
spark-default.conf
#spark.executor.extraClassPath /usr/local/spark/lib/spark-cassandra-connector-assembly-2.0.0-M1-2-g70018a6.jar
spark.executor.extraClassPath /usr/local/spark/lib/spark-cassandra-connector-1.6.0-M1-s_2.10.jar:/usr/local/spark/lib/cassandra-driver-core-3.0.0.jar:/usr/local/spark/lib/slf4j-api-1.7.5.jar:/usr/local/spark/lib/guava-16.0.1.jar:/usr/local/spark/lib/metrics-core-3.0.2.jar:/usr/local/spark/lib/netty-3.9.0.Final.jar
spark.driver.extraClassPath /usr/local/spark/lib/spark-cassandra-connector-1.6.0-M1-s_2.10.jar:/usr/local/spark/lib/cassandra-driver-core-3.0.0.jar:/usr/local/spark/lib/slf4j-api-1.7.5.jar:/usr/local/spark/lib/guava-16.0.1.jar:/usr/local/spark/lib/metrics-core-3.0.2.jar:/usr/local/spark/lib/netty-3.9.0.Final.jar
--