Fw: Help required setting up CDAP in distributed mode using HDP

388 views
Skip to first unread message

Nilkanth Patel

unread,
Aug 19, 2015, 1:10:43 AM8/19/15
to cdap...@googlegroups.com, robert geiger, Avinash Dongre


Kindly Help in starting CDAP in distributed mode with Hortonworks data platform.


 

I am trying  to set up CDAP in distributed mode with HDP 2.2.4.2. Steps that i have followed are


1.  Installed Sandbox_HDP_2.2.4.2_VirtualBox.ova into Virtualbox. Setup done successfully. IP address for this machine is 192.168.1.110

 

 

2. followed http://docs.cask.co/cdap/current/en/admin-manual/installation/installation.html 


and installed locally,


sudo apt-get install cdap-gateway cdap-kafka cdap-master cdap-security cdap-ui

Though documentation says, for configuring Hortonworks data platform you need to config following in cdap-env.sh

export OPTS="${OPTS} -Dhdp.version=<version>"

I could not find cdap-env.sh in my setup.


My /etc/cdap/conf/cdap-site.xml is as following


<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
  ~ Copyright © 2015 Cask Data, Inc.
  ~
  ~ Licensed under the Apache License, Version 2.0 (the "License"); you may not
  ~ use this file except in compliance with the License. You may obtain a copy of
  ~ the License at
  ~
  ~ http://www.apache.org/licenses/LICENSE-2.0
  ~
  ~ Unless required by applicable law or agreed to in writing, software
  ~ distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
  ~ WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
  ~ License for the specific language governing permissions and limitations under
  ~ the License.
  -->
<configuration>

  <property>
    <name>root.namespace</name>
    <value>cdap</value>
    <description>Specifies the root namespace</description>
  </property>

  <!-- Substitute the zookeeper quorum for components here -->
  <property>
    <name>zookeeper.quorum</name>
    <value>192.168.1.110:2181/${root.namespace}</value>
    <description>Specifies the zookeeper host:port</description>
  </property>

  <property>
    <name>hdfs.namespace</name>
    <value>/${root.namespace}</value>
    <description>Namespace for HDFS files</description>
  </property>

  <property>
    <name>hdfs.user</name>
    <value>yarn</value>
    <description>User name for accessing HDFS</description>
  </property>

  <!--
    Router configuration
  -->
  <!-- Substitue the IP to which Router service should bind to and listen on -->
  <property>
    <name>router.bind.address</name>
    <value>localhost</value>
    <description>Specifies the inet address on which the Router service will listen</description>
  </property>

  <!--
    App Fabric configuration
  -->
  <!-- Substitute the IP to which App-Fabric service should bind to and listen on -->
  <property>
    <name>app.bind.address</name>
    <value>localhost</value>
    <description>Specifies the inet address on which the app fabric service will listen</description>
  </property>

  <!--
    Data Fabric configuration
  -->
  <!-- Substitute the IP to which Data-Fabric tx service should bind to and listen on -->
  <property>
    <name>data.tx.bind.address</name>
    <value>localhost</value>
    <description>Specifies the inet address on which the transaction service will listen</description>
  </property>

  <!-- 
    Kafka Configuration
  -->
  <property>
    <name>kafka.log.dir</name>
    <value>/opt/cdap/kafka/logs</value>
    <description>Directory to store Kafka logs</description>
  </property>

  <!-- Must be <= the number of kafka.seed.brokers configured above.  For HA this should be at least 2. -->
  <property>
    <name>kafka.default.replication.factor</name>
    <value>1</value>
    <description>Kafka replication factor</description>
  </property>

  <!--
    Watchdog Configuration
  -->
  <!-- Substitute the IP to which metrics-query service should bind to and listen on -->
  <property>
    <name>metrics.query.bind.address</name>
    <value>localhost</value>
    <description>Specifies the inet address on which the metrics-query service will listen</description>
  </property>

  <!--
    Web-App Configuration
  -->
  <property>
    <name>dashboard.bind.port</name>
    <value>9999</value>
    <description>Specifies the port on which dashboard listens</description>
  </property>

  <!-- Substitute the IP of the Router service to which the UI should connect -->
  <property>
    <name>router.server.address</name>
    <value>localhost</value>
    <description>Specifies the destination IP where Router service is running</description>
  </property>

  <property>
    <name>router.server.port</name>
    <value>10000</value>
    <description>Specifies the destination Port where Router service is listening</description>
  </property>

  <property>
    <name>app.program.jvm.opts</name>
    <value>-XX:MaxPermSize=128M ${twill.jvm.gc.opts} -Dhdp.version=2.2.4.2</value>
    <description>Java options for all program containers</description>
  </property>

</configuration>

Not sure what should be the various find addresses for my case. kindly confirm it.


3. Starting services


root@ampool-HP-EliteBook-8570w:/opt/cdap# for i in `ls /etc/init.d/ | grep cdap` ; do sudo service $i start ; done
Wed Aug 19 09:26:48 IST 2015 Starting Java auth-server service on ampool-HP-EliteBook-8570w
Wed Aug 19 09:26:49 IST 2015 Starting Java kafka-server service on ampool-HP-EliteBook-8570w
Wed Aug 19 09:26:50 IST 2015 Starting Java master service on ampool-HP-EliteBook-8570w
Wed Aug 19 09:26:50 IST 2015 Starting Java router service on ampool-HP-EliteBook-8570w
Wed Aug 19 09:26:50 IST 2015 Starting ui service on ampool-HP-EliteBook-8570w

4. Status of cdap-services


root@ampool-HP-EliteBook-8570w:/opt/cdap# for i in `ls /etc/init.d/ | grep cdap` ; do sudo service $i status ; done
checking status
 * auth-server is running
checking status
 * kafka-server is running
checking status
 * master is running
checking status
 * router is running
checking status
 * ui is running


5. auth-server-cdap-ampool-HP-EliteBook-8570w.log has a following exception


SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [ch.qos.logback.classic.util.ContextSelectorStaticBinder]
09:26:49.535 [main] DEBUG org.apache.hadoop.util.Shell - Failed to detect a valid hadoop home directory
java.io.IOException: HADOOP_HOME or hadoop.home.dir are not set.
        at org.apache.hadoop.util.Shell.checkHadoopHome(Shell.java:225) [hadoop-common-2.2.0.jar:na]
        at org.apache.hadoop.util.Shell.<clinit>(Shell.java:250) [hadoop-common-2.2.0.jar:na]
        at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:76) [hadoop-common-2.2.0.jar:na]
        at org.apache.hadoop.yarn.conf.YarnConfiguration.<clinit>(YarnConfiguration.java:345) [hadoop-yarn-api-2.2.0.jar:na]
        at co.cask.cdap.common.guice.ConfigModule.configure(ConfigModule.java:64) [co.cask.cdap.cdap-common-3.1.1.jar:na]
        at com.google.inject.AbstractModule.configure(AbstractModule.java:59) [com.google.inject.guice-3.0.jar:na]
        at com.google.inject.spi.Elements$RecordingBinder.install(Elements.java:223) [com.google.inject.guice-3.0.jar:na]
        at com.google.inject.spi.Elements.getElements(Elements.java:101) [com.google.inject.guice-3.0.jar:na]
        at com.google.inject.internal.InjectorShell$Builder.build(InjectorShell.java:133) [com.google.inject.guice-3.0.jar:na]
        at com.google.inject.internal.InternalInjectorCreator.build(InternalInjectorCreator.java:103) [com.google.inject.guice-3.0.jar:na]
        at com.google.inject.Guice.createInjector(Guice.java:95) [com.google.inject.guice-3.0.jar:na]
        at com.google.inject.Guice.createInjector(Guice.java:72) [com.google.inject.guice-3.0.jar:na]
        at com.google.inject.Guice.createInjector(Guice.java:62) [com.google.inject.guice-3.0.jar:na]
        at co.cask.cdap.security.runtime.AuthenticationServerMain.init(AuthenticationServerMain.java:50) [co.cask.cdap.cdap-security-3.1.1.jar:na]
        at co.cask.cdap.common.runtime.DaemonMain.doMain(DaemonMain.java:36) [co.cask.cdap.cdap-common-3.1.1.jar:na]
        at co.cask.cdap.security.runtime.AuthenticationServerMain.main(AuthenticationServerMain.java:97) [co.cask.cdap.cdap-security-3.1.1.jar:na]
09:26:49.545 [main] DEBUG org.apache.hadoop.util.Shell - setsid exited with exit code 0


/opt/cdap/security/bin/common-env.sh , my changes are marked in bold red colour.


#!/usr/bin/env bash

#
# Copyright © 2014 Cask Data, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may not
# use this file except in compliance with the License. You may obtain a copy of
# the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations under
# the License.

# Set environment variables here.

# The java implementation to use.  Java 1.6 required.
 export JAVA_HOME=/opt/cdap/jdk1.7.0_79

# The maximum amount of heap to use, in MB. Default is 1000.
# export HEAPSIZE=1000

# Extra Java runtime options.
# Below are what we set by default.  May only work with SUN JVM.
# For more on why as well as other possible settings,
# see http://wiki.apache.org/hadoop/PerformanceTuning
export OPTS="-XX:+UseConcMarkSweepGC"

# Uncomment below to enable java garbage collection logging in the .out file.
# export GC_OPTS="-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps"

# Uncomment below (along with above GC logging) to put GC information in its own logfile
# export USE_GC_LOGFILE=true

# Where log files are stored.  $CDAP_HOME/logs by default.
export LOG_DIR=/var/log/cdap

# A string representing this instance of hbase. $USER by default.
export IDENT_STRING=$USER

# The scheduling priority for daemon processes.  See 'man nice'.
# export NICENESS=10

# The directory where pid files are stored. /tmp by default.
export PID_DIR=/var/cdap/run

# The directory serving as the user directory for master
export LOCAL_DIR=/var/tmp/cdap

# Specifies the JAVA_HEAPMAX
export JAVA_HEAPMAX=${JAVA_HEAPMAX:--Xmx128m}

# The options below can be set in the sourced component-specific conf/[component]-env.sh scripts

# Main class to be invoked.
#MAIN_CLASS=

# Arguments for main class.
#MAIN_CLASS_ARGS=""

# Adds Hadoop and HBase libs to the classpath on startup.
# If the "hbase" command is on the PATH, this will be done automatically.
# Or uncomment the line below to point to the HBase installation directly.
 HBASE_HOME=/opt/cdap/hbase-0.98.13-hadoop2

# Extra CLASSPATH
# EXTRA_CLASSPATH=""


6. kafka-server seems started properly. No exception found in log. part of kafka-server-cdap-ampool-HP-EliteBook-8570w.log 


2015-08-19 09:44:09,892 INFO  [EmbeddedKafkaServer STARTING] controller.KafkaController: [Controller 2130706433]: Controller startup complete
2015-08-19 09:44:09,910 INFO  [EmbeddedKafkaServer STARTING] server.KafkaServer: [Kafka Server 2130706433], Started
09:44:09.910 [main] INFO  c.c.cdap.kafka.run.KafkaServerMain - Embedded kafka server started successfully.
2015-08-19 09:44:09,912 INFO  [ZkClient-EventThread-22-192.168.1.110:2181/cdap/kafka] server.ZookeeperLeaderElector$LeaderChangeListener: New leader is 2130706433
0


changes  to /opt/cdap/kafka/bin/common-env.sh .  Added following configuration.

export JAVA_HOME=/opt/cdap/jdk1.7.0_79
HBASE_HOME=/opt/cdap/hbase-0.98.13-hadoop2



7. some of the exception found in master-cdap-ampool-HP-EliteBook-8570w.log


09:44:09.107 [main] DEBUG org.apache.hadoop.util.Shell - Failed to detect a valid hadoop home directory
java.io.IOException: HADOOP_HOME or hadoop.home.dir are not set.
        at org.apache.hadoop.util.Shell.checkHadoopHome(Shell.java:225) [hadoop-common-2.2.0.jar:na]
        at org.apache.hadoop.util.Shell.<clinit>(Shell.java:250) [hadoop-common-2.2.0.jar:na]
        at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:76) [hadoop-common-2.2.0.jar:na]
        at org.apache.hadoop.yarn.conf.YarnConfiguration.<clinit>(YarnConfiguration.java:345) [hadoop-yarn-api-2.2.0.jar:na]
        at co.cask.cdap.common.guice.ConfigModule.configure(ConfigModule.java:64) [co.cask.cdap.cdap-common-3.1.1.jar:na]
        at com.google.inject.AbstractModule.configure(AbstractModule.java:59) [com.google.inject.guice-3.0.jar:na]
        at com.google.inject.spi.Elements$RecordingBinder.install(Elements.java:223) [com.google.inject.guice-3.0.jar:na]
        at com.google.inject.spi.Elements.getElements(Elements.java:101) [com.google.inject.guice-3.0.jar:na]
        at com.google.inject.internal.InjectorShell$Builder.build(InjectorShell.java:133) [com.google.inject.guice-3.0.jar:na]


09:44:10.753 [main] INFO  c.c.c.d.d.InMemoryDatasetFramework - Adding Default module basicKVTable to system namespace
09:44:10.768 [main] INFO  c.c.c.d.r.m.TokenSecureStoreUpdater - Setting token renewal time to: 86100000 ms
09:44:10.770 [main] DEBUG c.c.c.d.r.main.MasterServiceMain - Failed to cleanup temp directory /var/tmp/cdap/data/tmp
java.io.IOException: Not a directory: /var/tmp/cdap/data/tmp
        at co.cask.cdap.common.utils.DirUtils.deleteDirectoryContents(DirUtils.java:61) ~[co.cask.cdap.cdap-common-3.1.1.jar:na]
        at co.cask.cdap.data.runtime.main.MasterServiceMain.cleanupTempDir(MasterServiceMain.java:418) [co.cask.cdap.cdap-master-3.1.1.jar:na]
        at co.cask.cdap.data.runtime.main.MasterServiceMain.init(MasterServiceMain.java:165) [co.cask.cdap.cdap-master-3.1.1.jar:na]
        at co.cask.cdap.common.runtime.DaemonMain.doMain(DaemonMain.java:36) [co.cask.cdap.cdap-common-3.1.1.jar:na]
        at co.cask.cdap.data.runtime.main.MasterServiceMain.main(MasterServiceMain.java:144) [co.cask.cdap.cdap-master-3.1.1.jar:na]
09:44:10.771 [main] INFO  c.cask.cdap.common.io.URLConnections - Turning off default caching in URLConnection


09:44:10.898 [main] DEBUG org.apache.zookeeper.ClientCnxn - zookeeper.disableAutoWatchReset is false
09:44:10.914 [main-SendThread(localhost:2181)] INFO  org.apache.zookeeper.ClientCnxn - Opening socket connection to server localhost/127.0.0.1:2181. Will not attempt to authenticate using SASL (unknown error)
09:44:10.918 [main-SendThread(localhost:2181)] WARN  org.apache.zookeeper.ClientCnxn - Session 0x0 for server null, unexpected error, closing socket connection and attempting reconnect
java.net.ConnectException: Connection refused
        at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) ~[na:1.7.0_79]
        at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:739) ~[na:1.7.0_79]
        at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:361) ~[zookeeper-3.4.6.jar:3.4.6-1569965]
        at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1081) ~[zookeeper-3.4.6.jar:3.4.6-1569965]
09:44:10.919 [main-SendThread(localhost:2181)] DEBUG o.a.zookeeper.ClientCnxnSocketNIO - Ignoring exception during shutdown input
java.nio.channels.ClosedChannelException: null

changes  to /opt/cdap/master/bin/common-env.sh .  Added following configuration.


export JAVA_HOME=/opt/cdap/jdk1.7.0_79
HBASE_HOME=/opt/cdap/hbase-0.98.13-hadoop2


8. Exceptions found in router-cdap-ampool-HP-EliteBook-8570w.log


java.io.IOException: HADOOP_HOME or hadoop.home.dir are not set.
        at org.apache.hadoop.util.Shell.checkHadoopHome(Shell.java:225) [hadoop-common-2.2.0.jar:na]
        at org.apache.hadoop.util.Shell.<clinit>(Shell.java:250) [hadoop-common-2.2.0.jar:na]
        at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:76) [hadoop-common-2.2.0.jar:na]
        at org.apache.hadoop.yarn.conf.YarnConfiguration.<clinit>(YarnConfiguration.java:345) [hadoop-yarn-api-2.2.0.jar:na]
        at co.cask.cdap.common.guice.ConfigModule.configure(ConfigModule.java:64) [co.cask.cdap.cdap-common-3.1.1.jar:na]
        at com.google.inject.AbstractModule.configure(AbstractModule.java:59) [com.google.inject.guice-3.0.jar:na]


co.cask.cdap.common.HandlerException: No endpoint strategy found for request : /ping
        at co.cask.cdap.gateway.router.handlers.HttpRequestHandler.getDiscoverable(HttpRequestHandler.java:197) ~[co.cask.cdap.cdap-gateway-3.1.1.jar:na]
        at co.cask.cdap.gateway.router.handlers.HttpRequestHandler.messageReceived(HttpRequestHandler.java:106) ~[co.cask.cdap.cdap-gateway-3.1.1.jar:na]
        at co.cask.cdap.gateway.router.handlers.HttpStatusRequestHandler.messageReceived(HttpStatusRequestHandler.java:65) ~[co.cask.cdap.cdap-gateway-3.1.1.jar:na]
        at org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:296) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.handler.codec.frame.FrameDecoder.unfoldAndFireMessageReceived(FrameDecoder.java:459) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.handler.codec.replay.ReplayingDecoder.callDecode(ReplayingDecoder.java:536) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.handler.codec.replay.ReplayingDecoder.messageReceived(ReplayingDecoder.java:435) ~[io.netty.netty-3.6.6.Final.jar:na]
        at co.cask.cdap.gateway.router.NettyRouter$1.handleUpstream(NettyRouter.java:177) ~[co.cask.cdap.cdap-gateway-3.1.1.jar:na]
        at org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:268) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:255) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.channel.socket.nio.NioWorker.read(NioWorker.java:88) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.channel.socket.nio.AbstractNioWorker.process(AbstractNioWorker.java:109) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:312) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.channel.socket.nio.AbstractNioWorker.run(AbstractNioWorker.java:90) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:178) ~[io.netty.netty-3.6.6.Final.jar:na]
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) [na:1.7.0_79]
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) [na:1.7.0_79]
        at java.lang.Thread.run(Thread.java:745) [na:1.7.0_79]
09:44:16.528 [New I/O worker #13] DEBUG c.c.c.g.router.RouterServiceLookup - Looking up service name appfabric
09:44:16.829 [New I/O worker #13] DEBUG c.c.c.g.router.RouterServiceLookup - Discoverable endpoint appfabric not found
09:44:16.829 [New I/O worker #13] ERROR c.c.c.g.router.RouterServiceLookup - No discoverable endpoints found for service CacheKey{service=appfabric, host=localhost:10000, firstPathPart=/v3}
09:44:16.831 [New I/O worker #13] ERROR c.c.c.g.r.h.HttpRequestHandler - Exception raised in Request Handler [id: 0x0d2ed568, /127.0.0.1:58068 => /127.0.0.1:10000]
co.cask.cdap.common.HandlerException: No endpoint strategy found for request : /v3/namespaces
        at co.cask.cdap.gateway.router.handlers.HttpRequestHandler.getDiscoverable(HttpRequestHandler.java:197) ~[co.cask.cdap.cdap-gateway-3.1.1.jar:na]
        at co.cask.cdap.gateway.router.handlers.HttpRequestHandler.messageReceived(HttpRequestHandler.java:106) ~[co.cask.cdap.cdap-gateway-3.1.1.jar:na]
        at co.cask.cdap.gateway.router.handlers.HttpStatusRequestHandler.messageReceived(HttpStatusRequestHandler.java:65) ~[co.cask.cdap.cdap-gateway-3.1.1.jar:na]
        at org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:296) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.handler.codec.frame.FrameDecoder.unfoldAndFireMessageReceived(FrameDecoder.java:459) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.handler.codec.replay.ReplayingDecoder.callDecode(ReplayingDecoder.java:536) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.handler.codec.replay.ReplayingDecoder.messageReceived(ReplayingDecoder.java:435) ~[io.netty.netty-3.6.6.Final.jar:na]
        at co.cask.cdap.gateway.router.NettyRouter$1.handleUpstream(NettyRouter.java:177) ~[co.cask.cdap.cdap-gateway-3.1.1.jar:na]
        at org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:268) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.channel.Channels.fireMessageReceived(Channels.java:255) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.channel.socket.nio.NioWorker.read(NioWorker.java:88) ~[io.netty.netty-3.6.6.Final.jar:na]
        at org.jboss.netty.channel.socket.nio.AbstractNioWorker.process(AbstractNioWorker.java:109) ~[io.netty.netty-3.6.6.Final.jar:na]

changes  to/opt/cdap/gateway/bin/common-env.sh .  Added following configuration.


export JAVA_HOME=/opt/cdap/jdk1.7.0_79
HBASE_HOME=/opt/cdap/hbase-0.98.13-hadoop2



9. ui-cdap-ampool-HP-EliteBook-8570w.log  does not log any exception.



10, health check of the CDAP UI http://localhost:9999/status gives me following screenshot




11. health check of the CDAP Router , http://localhost:10000/status


OK


12. health check of the CDAP Authentication Server, http://localhost:10009/status

Unable to connect


13. health check of all the services running in YARNhttp://localhost:10000/v3/system/services

No endpoint strategy found for request : /v3/system/services
14. Classpath for cdap master service
root@ampool-HP-EliteBook-8570w:/opt/cdap/security# /opt/cdap/master/bin/svc-master classpath
/opt/cdap/hbase-compat-0.98/lib/*:/opt/cdap/master/lib/*:/opt/cdap/hbase-0.98.13-hadoop2/bin/../conf:/opt/cdap/jdk1.7.0_79/lib/tools.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/..:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/activation-1.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/aopalliance-1.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/asm-3.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/avro-1.7.4.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-beanutils-1.7.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-beanutils-core-1.8.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-cli-1.2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-codec-1.7.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-collections-3.2.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-compress-1.4.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-configuration-1.6.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-daemon-1.0.13.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-digester-1.8.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-el-1.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-httpclient-3.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-io-2.4.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-lang-2.6.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-logging-1.1.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-math-2.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/commons-net-3.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/findbugs-annotations-1.3.9-1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/gmbal-api-only-3.0.0-b023.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/grizzly-framework-2.1.2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/grizzly-http-2.1.2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/grizzly-http-server-2.1.2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/grizzly-http-servlet-2.1.2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/grizzly-rcm-2.1.2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/guava-12.0.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/guice-3.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/guice-servlet-3.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-annotations-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-auth-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-client-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-common-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-hdfs-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-mapreduce-client-app-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-mapreduce-client-common-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-mapreduce-client-core-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-mapreduce-client-jobclient-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-mapreduce-client-shuffle-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-yarn-api-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-yarn-client-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-yarn-common-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-yarn-server-common-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hadoop-yarn-server-nodemanager-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hamcrest-core-1.3.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-annotations-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-checkstyle-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-client-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-common-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-common-0.98.13-hadoop2-tests.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-examples-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-hadoop2-compat-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-hadoop-compat-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-it-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-it-0.98.13-hadoop2-tests.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-prefix-tree-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-protocol-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-rest-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-server-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-server-0.98.13-hadoop2-tests.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-shell-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-testing-util-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/hbase-thrift-0.98.13-hadoop2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/high-scale-lib-1.1.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/htrace-core-2.04.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/httpclient-4.1.3.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/httpcore-4.1.3.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jackson-core-asl-1.8.8.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jackson-jaxrs-1.8.8.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jackson-mapper-asl-1.8.8.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jackson-xc-1.8.8.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jamon-runtime-2.3.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jasper-compiler-5.5.23.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jasper-runtime-5.5.23.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/javax.inject-1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/javax.servlet-3.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/javax.servlet-api-3.0.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jaxb-api-2.2.2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jaxb-impl-2.2.3-1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jcodings-1.0.8.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jersey-client-1.8.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jersey-core-1.8.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jersey-grizzly2-1.9.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jersey-guice-1.9.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jersey-json-1.8.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jersey-server-1.8.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jersey-test-framework-core-1.9.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jersey-test-framework-grizzly2-1.9.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jets3t-0.6.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jettison-1.3.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jetty-6.1.26.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jetty-sslengine-6.1.26.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jetty-util-6.1.26.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/joni-2.1.2.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jruby-complete-1.6.8.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jsch-0.1.42.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jsp-2.1-6.1.14.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jsp-api-2.1-6.1.14.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/jsr305-1.3.9.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/junit-4.11.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/libthrift-0.9.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/log4j-1.2.17.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/management-api-3.0.0-b012.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/metrics-core-2.2.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/netty-3.6.6.Final.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/paranamer-2.3.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/protobuf-java-2.5.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/servlet-api-2.5-6.1.14.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/slf4j-api-1.6.4.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/slf4j-log4j12-1.6.4.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/snappy-java-1.0.4.1.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/xmlenc-0.52.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/xz-1.0.jar:/opt/cdap/hbase-0.98.13-hadoop2/bin/../lib/zookeeper-3.4.6.jar:/usr/hdp/2.2.4.2-2/hadoop/conf:/usr/hdp/2.2.4.2-2/hadoop/lib/*:/usr/hdp/2.2.4.2-2/hadoop/.//*:/usr/hdp/2.2.4.2-2/hadoop-hdfs/./:/usr/hdp/2.2.4.2-2/hadoop-hdfs/lib/*:/usr/hdp/2.2.4.2-2/hadoop-hdfs/.//*:/usr/hdp/2.2.4.2-2/hadoop-yarn/lib/*:/usr/hdp/2.2.4.2-2/hadoop-yarn/.//*:/usr/hdp/2.2.4.2-2/hadoop-mapreduce/lib/*:/usr/hdp/2.2.4.2-2/hadoop-mapreduce/.//*:/etc/cdap/conf/:/opt/cdap/master/conf/:/etc/hbase/conf/
Looking forward to your help.

Thanks in advance

Nilkanth patel.
Ampool India Pvt. Ltd.



Derek Wood

unread,
Aug 19, 2015, 4:00:20 PM8/19/15
to Nilkanth Patel, cdap...@googlegroups.com, robert geiger, Avinash Dongre
Hi Nilkanth,

In your step 2, you need to actually create /etc/cdap/conf/cdap-env.sh and add the export OPTS="${OPTS} -Dhdp.version=<version>". I've opened https://issues.cask.co/browse/CDAP-3443 for us to ship an empty file by default. This is very likely the cause of your issue.

As for the various addresses, assuming you have installed cdap components on the same box, it's probably easiest to just set them all to the externally routable ip or hostname.

*.bind.address: the interface the services will listen on.  
*.server.address: the remote address for which a service will attempt to connect to another

For example, cdap-router will listen on router.bind.address. The UI will connect to router according to the configured router.server.address, so these two must match.

Hope this helps,
-Derek

--
You received this message because you are subscribed to the Google Groups "CDAP User" group.
To unsubscribe from this group and stop receiving emails from it, send an email to cdap-user+...@googlegroups.com.
To post to this group, send email to cdap...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/cdap-user/BN3PR10MB0196500CE9C6CD0B03C0BC62D1670%40BN3PR10MB0196.namprd10.prod.outlook.com.
For more options, visit https://groups.google.com/d/optout.

Reply all
Reply to author
Forward
0 new messages