Akshar--
You received this message because you are subscribed to the Google Groups "Google Cloud Bigtable Discuss" group.
To unsubscribe from this group and stop receiving emails from it, send an email to google-cloud-bigtable-discuss+unsub...@googlegroups.com.
To post to this group, send email to google-cloud-bigtable-discuss@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/google-cloud-bigtable-discuss/d8958ff5-3cc7-43e5-8ac5-c2dfa41bedba%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
To view this discussion on the web visit https://groups.google.com/d/msgid/google-cloud-bigtable-discuss/CAEbjtcMH3y_qM2uf2CNAHRaWW_PYBALAAsT67g0nofuMaVuq8A%40mail.gmail.com.
Akshar
To unsubscribe from this group and stop receiving emails from it, send an email to google-cloud-bigtabl...@googlegroups.com.
To post to this group, send email to google-cloud-b...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/google-cloud-bigtable-discuss/d8958ff5-3cc7-43e5-8ac5-c2dfa41bedba%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
--
You received this message because you are subscribed to the Google Groups "Google Cloud Bigtable Discuss" group.
To unsubscribe from this group and stop receiving emails from it, send an email to google-cloud-bigtabl...@googlegroups.com.
To post to this group, send email to google-cloud-b...@googlegroups.com.
Akshar
To unsubscribe from this group and stop receiving emails from it, send an email to google-cloud-bigtable-discuss+unsub...@googlegroups.com.
To post to this group, send email to google-cloud-bigtable-discuss@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/google-cloud-bigtable-discuss/d8958ff5-3cc7-43e5-8ac5-c2dfa41bedba%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
--
You received this message because you are subscribed to the Google Groups "Google Cloud Bigtable Discuss" group.
To unsubscribe from this group and stop receiving emails from it, send an email to google-cloud-bigtable-discuss+unsub...@googlegroups.com.
To post to this group, send email to google-cloud-bigtable-discuss@googlegroups.com.
Finally got to trying out the recommended change ...inject, generate, fetch, parse have all worked. The change does get rid of the failure. I can also see the crawled html in the big table.I only did about 6 urls, but for some reason the updatedb phase is still running after 15 minutes and only 28%the jobs do always output this error a bunch of times at the start, but it doens't seem to affect the jobJan 03, 2018 9:49:56 PM com.google.bigtable.repackaged.io.grpc.internal.ManagedChannelImpl$ManagedChannelReference cleanQueueSEVERE: *~*~*~ Channel com.google.bigtable.repackaged.io.grpc.internal.ManagedChannelImpl-35 for target bigtable.googleapis.com:443 was not shutdown properly!!! ~*~*~*Make sure to call shutdown()/shutdownNow() and awaitTermination().java.lang.RuntimeException: ManagedChannel allocation siteat com.google.bigtable.repackaged.io.grpc.internal.ManagedChannelImpl$ManagedChannelReference.<init>(ManagedChannelImpl.java:991)at com.google.bigtable.repackaged.io.grpc.internal.ManagedChannelImpl.<init>(ManagedChannelImpl.java:421)at com.google.bigtable.repackaged.io.grpc.internal.AbstractManagedChannelImplBuilder.build(AbstractManagedChannelImplBuilder.java:329)at com.google.bigtable.repackaged.com.google.cloud.bigtable.grpc.BigtableSession.createNettyChannel(BigtableSession.java:483)at com.google.bigtable.repackaged.com.google.cloud.bigtable.grpc.BigtableSession$3.create(BigtableSession.java:398)at com.google.bigtable.repackaged.com.google.cloud.bigtable.grpc.io.ChannelPool.<init>(ChannelPool.java:246)at com.google.bigtable.repackaged.com.google.cloud.bigtable.grpc.BigtableSession.createChannelPool(BigtableSession.java:401)at com.google.bigtable.repackaged.com.google.cloud.bigtable.grpc.BigtableSession.createManagedPool(BigtableSession.java:413)at com.google.bigtable.repackaged.com.google.cloud.bigtable.grpc.BigtableSession.getDataChannelPool(BigtableSession.java:276)at com.google.bigtable.repackaged.com.google.cloud.bigtable.grpc.BigtableSession.<init>(BigtableSession.java:236)at org.apache.hadoop.hbase.client.AbstractBigtableConnection.<init>(AbstractBigtableConnection.java:143)at com.google.cloud.bigtable.hbase1_x.BigtableConnection.<init>(BigtableConnection.java:58)at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)at java.lang.reflect.Constructor.newInstance(Constructor.java:423)at org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:238)at org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:218)at org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:119)at org.apache.gora.hbase.store.HBaseStore.initialize(HBaseStore.java:131)at org.apache.gora.store.DataStoreFactory.initializeDataStore(DataStoreFactory.java:97)at org.apache.gora.store.DataStoreFactory.createDataStore(DataStoreFactory.java:156)at org.apache.gora.store.DataStoreFactory.createDataStore(DataStoreFactory.java:130)at org.apache.nutch.storage.StorageUtils.createWebStore(StorageUtils.java:78)at org.apache.nutch.storage.StorageUtils.initMapperJob(StorageUtils.java:133)at org.apache.nutch.storage.StorageUtils.initMapperJob(StorageUtils.java:122)at org.apache.nutch.crawl.GeneratorJob.run(GeneratorJob.java:224)at org.apache.nutch.crawl.GeneratorJob.generate(GeneratorJob.java:260)at org.apache.nutch.crawl.GeneratorJob.run(GeneratorJob.java:326)at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)at org.apache.nutch.crawl.GeneratorJob.main(GeneratorJob.java:334)at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)at java.lang.reflect.Method.invoke(Method.java:498)at org.apache.hadoop.util.RunJar.run(RunJar.java:221)at org.apache.hadoop.util.RunJar.main(RunJar.java:136)Looks like this is likely why the job is running so long ...Task attempt attempt_1507853061749_0037_r_000002_0 is done from TaskUmbilicalProtocol's point of view. However, it stays in finishing state for too long Container killed by the ApplicationMaster. Container killed on request. Exit code is 143 Container exited with a non-zero exit code 143Its in the logs for all the completed tasks for the updatedb jobits saying that all of the tasks are just getting killed by the Application Master because they aren't cleaning themselves up. is that there is zero work for all these tasks to do since i only used 6 urls, so they aren't properly reporting a finished state when they don't have anything to do?Thanks for all the help so far.AksharOn Thu, Nov 30, 2017 at 10:39 AM, Akshar Dave <aksha...@gmail.com> wrote:Great, thanks! Will look into building gora locally and try again. Is there a more recent one after pre3?On Thu, Nov 30, 2017 at 9:10 AM Solomon Duskis <sdu...@google.com> wrote:Yes. Changing one line will get you moving forward.Delete delete = new Delete(keyRaw, timeStamp - PUTS_AND_DELETES_DELETE_TS_OFFSET);should beDelete delete = new Delete(keyRaw);Then you can build gora/nutch locally with snapshots. I would also suggest upgrading the bigtable-hbase dependency to release v. 1.0.0.-SolomonOn Thu, Nov 30, 2017, 11:57 AM Akshar Dave <aksha...@gmail.com> wrote:Apologies for Doing DM ... Thanks a lot for figuring out the root cause...is there any way to use Bigtable since we are not interested in delete workflow and fine with data being in the backend till we get this resolved? Would you try commenting out those lines to get it working?