17/03/16 10:07:49 INFO type: Alluxio client (version 1.3.0) is trying to connect with FileSystemMasterClient master @ /xxx.xxx.xxx.xxx:19998
java.io.IOException: Failed to login: No Alluxio User is found. Current login principals are [UnixPrincipal: xxx, UnixNumericUserPrincipal: 1001, UnixNumericGroupPrincipal [Primary Group]: 1001, xxx].
If you are running hadoop with Alluxio, please double check whether you have alluxio client jar included in HADOOP_CLASSPATH. Otherwise, you are expected to see this error.
at alluxio.security.LoginUser.login(LoginUser.java:98)
at alluxio.security.LoginUser.get(LoginUser.java:57)
at alluxio.security.authentication.PlainSaslTransportProvider.getClientTransport(PlainSaslTransportProvider.java:53)
at alluxio.AbstractClient.connect(AbstractClient.java:173)
at alluxio.AbstractClient.retryRPC(AbstractClient.java:346)
at alluxio.client.file.FileSystemMasterClient.getStatus(FileSystemMasterClient.java:183)
at alluxio.client.file.BaseFileSystem.getStatus(BaseFileSystem.java:176)
at alluxio.client.file.BaseFileSystem.getStatus(BaseFileSystem.java:168)
at alluxio.hadoop.AbstractFileSystem.getFileStatus(AbstractFileSystem.java:293)
at alluxio.hadoop.FileSystem.getFileStatus(FileSystem.java:25)
at org.apache.hadoop.fs.Globber.getFileStatus(Globber.java:57)
at org.apache.hadoop.fs.Globber.glob(Globber.java:252)
at org.apache.hadoop.fs.FileSystem.globStatus(FileSystem.java:1676)
at org.apache.hadoop.mapred.FileInputFormat.singleThreadedListStatus(FileInputFormat.java:259)
at org.apache.hadoop.mapred.FileInputFormat.listStatus(FileInputFormat.java:229)
at org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:315)
at org.apache.spark.rdd.HadoopRDD.getPartitions(HadoopRDD.scala:202)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:252)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:250)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:250)
at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:252)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:250)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:250)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1958)
at org.apache.spark.rdd.RDD.count(RDD.scala:1157)
... 48 elided