#!/usr/bin/env bash
export HADOOP_CONF_DIR=/opt/yarn-conf
export YARN_CONF_DIR=/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf
export CLASSPATH=$HADOOP_CONF_DIR:$YARN_CONF_DIR
export HADOOP_HOME=/opt/cloudera/parcels/CDH/lib/hadoop
export JAVA_OPTIONS="$JAVA_OPTIONS -Djava.library.path=$HADOOP_HOME/lib/native:$HADOOP_HOME/lib/native/Linux-amd64-64"
java -cp graph_analyzer-1.0-SNAPSHOT.jar:/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/*:/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/lib/* com.fosun.graph_analyzer.GraphAnalyzer
#
# Hadoop Graph Configuration
#
gremlin.graph=org.apache.tinkerpop.gremlin.hadoop.structure.HadoopGraph
gremlin.hadoop.graphInputFormat=org.janusgraph.hadoop.formats.cassandra.Cassandra3InputFormat
gremlin.hadoop.graphOutputFormat=org.apache.tinkerpop.gremlin.hadoop.structure.io.gryo.GryoOutputFormat
gremlin.hadoop.jarsInDistributedCache=true
gremlin.hadoop.inputLocation=none
gremlin.hadoop.outputLocation=output
gremlin.spark.persistContext=true
#
# JanusGraph Cassandra InputFormat configuration
#
janusgraphmr.ioformat.conf.storage.backend=cassandrathrift
janusgraphmr.ioformat.conf.storage.hostname=fonova-app-jan01,fonova-app-jan02,fonova-app-jan03
janusgraphmr.ioformat.conf.storage.port=9160
janusgraphmr.ioformat.conf.storage.cassandra.keyspace=janusgraph
janusgraphmr.ioformat.conf.storage.cassandra.frame-size-mb=128
janusgraphmr.ioformat.conf.storage.cassandra.astyanax.frame-size=128
storage.cassandra.thrift.frame-size=128
storage.cassandra.thrift.max_message_size_mb=128
storage.cassandra.frame-size-mb=128
storage.cassandra.astyanax.frame-size=128
#
# Apache Cassandra InputFormat configuration
#
cassandra.input.partitioner.class=org.apache.cassandra.dht.Murmur3Partitioner
cassandra.input.keyspace=janusgraph
cassandra.input.predicate=0c00020b0001000000000b000200000000020003000800047fffffff0000
cassandra.input.columnfamily=edgestore
cassandra.range.batch.size=2147483647
#
# SparkGraphComputer Configuration
#
spark.master=yarn
spark.submit.deployMode=client
spark.serializer=org.apache.spark.serializer.KryoSerializer
spark.security.authentication=kerberos
spark.keytab=/data/graph/graph@FONOVA_AHZ.COM.keytab
spark.principal=graph/GRAPHCOMPUTE@FONOVA_AHZ.COM
spark.queue=root.ahz_batch.dev
spark.driver.memory=1g
spark.driverEnv.HADOOP_CONF_DIR=/etc/hadoop/conf
spark.executorEnv.HADOOP_CONF_DIR=/etc/hadoop/conf
spark.files=/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/core-site.xml,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/hadoop-env.sh,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/hbase-env.sh,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/hbase-site.xml,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/hdfs-site.xml,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/hive-site.xml,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/jaas.conf,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/log4j.properties,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/mapred-site.xml,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/read-cassandra-3.properties,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/ssl-client.xml,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/topology.map,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/topology.py,/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf/yarn-site.xml
spark.executor.cores=1
spark.executor.memory=2g
spark.num.executors=25
spark.driver.class.path=/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf:/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/lib
spark.yarn.appMasterEnv.CLASSPATH=/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/lib/*:/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf
spark.yarn.dist.jars=/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/lib/*
spark.executor.extraClassPath=/opt/cloudera/parcels/CDH/lib/hadoop/native:/opt/cloudera/parcels/CDH/lib/hadoop-0.20-mapreduce/lib/native/Linux-amd64-64:/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf:/data/fosundb/ran/graph/spark-1.6.1-bin-hadoop2.6/lib/spark-assembly-1.6.1-hadoop2.6.0.jar:/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/lib
spark.dirver.extraClassPath=/opt/cloudera/parcels/CDH/lib/hadoop/native:/opt/cloudera/parcels/CDH/lib/hadoop-0.20-mapreduce/lib/native/Linux-amd64-64:/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf:/data/fosundb/ran/graph/spark-1.6.1-bin-hadoop2.6/lib/spark-assembly-1.6.1-hadoop2.6.0.jar:/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/lib
spark.history.kerberos.enabled=true
spark.history.kerberos.principal=graph/GRAPHCOMPUTE@FONOVA_AHZ.COM
spark.history.kerberos.keytab=/data/graph/graph@FONOVA_AHZ.COM.keytab/home/myuser/graph/graph_analyzer-1.0-SNAPSHOT/conf
Now the problems are:
1. It shows that no authentication is allowed:
11:10:44,907 INFO SecurityManager:58 - SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(graph); users with modify permissions: Set(graph)
2. It shows that yarn-site.xml do no help on setting yarn-host:
11:10:50,833 INFO RMProxy:56 - Connecting to ResourceManager at /0.0.0.0:8032
spark.history.kerberos.principal=graph/GRAPHC...@FONOVA_AHZ.COM
spark.history.kerberos.keytab=/data/graph/graph@FONOVA_AHZ.COM.keytab
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.yarn.conf.YarnConfiguration;
import org.apache.tinkerpop.gremlin.hadoop.structure.HadoopConfiguration;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
public class GraphAnalyzer {
private static final Logger LOGGER = LoggerFactory.getLogger(GraphAnalyzer.class);
public static void main(String[] args) {
Configuration hadoopConf = new Configuration();
HadoopConfiguration hadoopConfiguration = new HadoopConfiguration();
YarnConfiguration yarnConfiguration = new YarnConfiguration();
System.out.println("------------------------------------------------------------------");
System.out.println(hadoopConf.get("yarn.resourcemanager.address"));
System.out.println("------------------------------------------------------------------");
System.out.println(yarnConfiguration.get("yarn.resourcemanager.address"));
System.out.println("------------------------------------------------------------------");
System.out.println(hadoopConfiguration.getProperty("yarn.resourcemanager.address"));
}
}export HADOOP_CONF_DIR=/opt/yarn-conf
export YARN_CONF_DIR=/opt/spark-2.2.0/conf/yarn-conf
export SPARK_HOME=/opt/spark-2.2.0
export MASTER=yarn
export DEPLOY_MODE=client