Hbase Region Server Down Due to Indexer

45 views
Skip to first unread message

ankit beohar

unread,
Jul 29, 2016, 3:53:32 AM7/29/16
to HBase Indexer Users
Hi All,

I am using Solr with Hbase(Lilly Indexer) but I am getting below error in my region server logs due to this my region server goes down every time:-


Indexer_DocumentIndexer xxxxxxxxxxxxxxxxxxxxx Got:
java.io.IOException: File from recovered queue is nowhere to be found
        at org.apache.hadoop.hbase.replication.regionserver.ReplicationSource$ReplicationSourceWorkerThread.openReader(ReplicationSource.java:804)
        at org.apache.hadoop.hbase.replication.regionserver.ReplicationSource$ReplicationSourceWorkerThread.run(ReplicationSource.java:542)
Caused by: java.io.FileNotFoundException: File does not exist: /hbase/oldWALs/xxxxxxxxxxxxxxxxxxxxxx
        at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:66)
        at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:56)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1932)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1873)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1853)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1825)
        at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:559)
        at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.getBlockLocations(AuthorizationProviderProxyClientProtocol.java:87)
        at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:363)
        at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2040)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1707)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2038)

        at sun.reflect.GeneratedConstructorAccessor16.newInstance(Unknown Source)
        at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
        at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
        at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
        at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
        at org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:1215)
        at org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:1203)
        at org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:1193)
        at org.apache.hadoop.hdfs.DFSInputStream.fetchLocatedBlocksAndGetLastBlockLength(DFSInputStream.java:299)
        at org.apache.hadoop.hdfs.DFSInputStream.openInfo(DFSInputStream.java:265)
        at org.apache.hadoop.hdfs.DFSInputStream.<init>(DFSInputStream.java:257)
        at org.apache.hadoop.hdfs.DFSClient.open(DFSClient.java:1492)
        at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:302)
        at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:298)
        at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
        at org.apache.hadoop.hdfs.DistributedFileSystem.open(DistributedFileSystem.java:298)
        at org.apache.hadoop.fs.FilterFileSystem.open(FilterFileSystem.java:161)
        at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:766)
        at org.apache.hadoop.hbase.wal.WALFactory.createReader(WALFactory.java:291)
        at org.apache.hadoop.hbase.wal.WALFactory.createReader(WALFactory.java:267)
        at org.apache.hadoop.hbase.wal.WALFactory.createReader(WALFactory.java:255)
        at org.apache.hadoop.hbase.wal.WALFactory.createReader(WALFactory.java:397)
        at org.apache.hadoop.hbase.replication.regionserver.ReplicationWALReaderManager.openReader(ReplicationWALReaderManager.java:69)
        at org.apache.hadoop.hbase.replication.regionserver.ReplicationSource$ReplicationSourceWorkerThread.openReader(ReplicationSource.java:746)
        ... 1 more
Caused by: org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException): File does not exist: /hbase/oldWALs/
        at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:66)
        at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:56)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsUpdateTimes(FSNamesystem.java:1932)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocationsInt(FSNamesystem.java:1873)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1853)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1825)
        at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:559)
        at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.getBlockLocations(AuthorizationProviderProxyClientProtocol.java:87)
        at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:363)
        at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044)

Thanks

Gabriel Reid

unread,
Jul 31, 2016, 3:38:54 AM7/31/16
to ankit beohar, HBase Indexer Users
Hi Ankit,

Which version of HBase and hbase-indexer are you using? This sounds
like it's some kind of bug or corruption in HBase, although it could
be due to incompatible versions of HBase and hbase-indexer.

- Gabriel
> --
> You received this message because you are subscribed to the Google Groups
> "HBase Indexer Users" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to hbase-indexer-u...@googlegroups.com.
> For more options, visit https://groups.google.com/d/optout.

ankit beohar

unread,
Aug 1, 2016, 1:30:58 AM8/1/16
to Gabriel Reid, HBase Indexer Users
Hi Gabriel,

We are using CDH5.4 and HBase 1.0.0-cdh5.4.10 and for indexing we are using Lilly Indexer.

Best Regards,
ANKIT BEOHAR

Gabriel Reid

unread,
Aug 3, 2016, 2:40:56 AM8/3/16
to ankit beohar, HBase Indexer Users
Hi Ankit,

Hmm, those versions look fine.

This looks to be an issue in HBase replication itself, although I've
never encountered it before myself.

Have you run into this situation in multiple environments? Or is there
any chance you could set up a clean environment somewhere to see if it
can be replicated.

I would also suggest taking a look through the HBase JIRA to see if
there is something similar that has already been reported in HBase.

- Gabriel
Reply all
Reply to author
Forward
0 new messages