Could you please recommend how to troubleshoot this issue?
Thank you,
Nella
spark-shell
> sqlContext.sql('use default')
> sqlContext.sql('show tables').collect()Required executor memory (4608+460 MB) is above the max threshold (3072 MB) of this cluster! Please check the values of 'yarn.scheduler.maximum-allocation-mb' and/or 'yarn.nodemanager.resource.memory-mb
although spark-submit works correctly. I have not been able to figure out how to change configuration. Could you please recommend what do to about this one?
Thank you,
Nella
spark-shell --master localThank you!! I have created a new EMR cluster and seems like spark-shell --master local works so far.
I have repointed edge node to this new cluster and created trust and tunnel. Now Register Tables box is failing with error
Cannot create PoolableConnectionFactory (Could not open client transport with JDBC Uri: jdbc:hive2://localhost:10000/default: java.net.ConnectException: Connection refused (Connection refused))
It used to work with the previous EMR. Could you please help where to look for the cause of this issue?
Thank you,
Nella
It fixed the issue with table registration.
Now I am having an error in Validate and Split
org.apache.spark.sql.catalyst.analysis.NoSuchTableException: Table or view 'init21_valid' not found in database 'website';
Although I see in hive that this table has been created.
I have created a soft link from SPARK_HOME directory and it worked.
Now I verified that hive tables have been created and would like to see tables in Data Wrangler.
When I go to Tables menu - I have error on the bottom regarding unexpected error.
In the logs I see exception:
org.springframework.jdbc.support.MetaDataAccessException: Could not get Connection for extracting meta data; nested exception is org.springframework.jdbc.CannotGetJdbcConnectionException: Could not get JDBC Connection; nested exception is java.sql.SQLSyntaxErrorException: Could not connect: Unknown database 'hive'
On EMR cluster in mySQL I cant see hive table when I login as hadoop. When I login as root - I can see it.
Thank you,
Nella
hive.metastore.datasource.driverClassName=org.mariadb.jdbc.Driverhive.metastore.datasource.url=jdbc:mysql://localhost:3306/hivehive.metastore.datasource.username=roothive.metastore.datasource.password=hadoophive.metastore.datasource.driverClassName=org.mariadb.jdbc.Driver
hive.metastore.datasource.url=jdbc:mysql://emr-internal-ip:3306/hive
hive.metastore.datasource.username=hive
hive.metastore.datasource.password=hive-password
hive.metastore.datasource.validationQuery=SELECT 1
hive.metastore.datasource.testOnBorrow=true
I have created hive user in MariaDB on EMR and granted all privileges.
I can see tables now!!!!
When I go to Visual query though and execute select * statement it is failing. I can also see in the kylo log that spark service failed since it had multiple messages
22:56:46,079 Client.logInfo: Application report for application_1503078848312_0015 (state: ACCEPTED)
ValidateAndSplit job is working though.
Is there anything I can change on SPARK parameters?
spark.driver.memory=512m
spark.executor.memory=512m
Is there anything else that can be done or tested in this case? How much resources are required on EMR cluster so that Data Wrangler could run?
Thank you,
Nella
#spark.shell.server.host = localhost#spark.shell.server.port = 8450spark.shell.master = localIt worked.
I am reconfiguring the setup with the latest kylo using document in JIRA and my notes from previous installation.
Feed is running in nifi although I cant see the progress in Kylo
ERROR DefaultMessageListenerContainer-1:DefaultMessageListenerContainer:941 - Could not refresh JMS Connection for destination 'thinkbig.provenance-event-stats' - retrying using FixedBackOff{interval=5000, currentAttempts=60, maxAttempts=unlimited}. Cause: Error while attempting to add new Connection to the pool; nested exception is javax.jms.JMSException: Could not connect to broker URL: tcp://localhost:61616. Reason: java.net.ConnectException: Connection refused (Connection refused)
Could you please help to understand what might be going wrong?
Thank you,
Nella