We have a nexus setup actually not showing any problems at the surface, i.e. we can upload and download artifacts, search for existing components etc.
However when going through the logs after a recent uncontrolled disruption of the service, we noticed regular messages (dating from before the mentioned disruption) by elasicsearch concerning "failed shards" for some indexes.
Those messages occur quite frequently - probably whenever the index would have to be updated.
I have two questions here:
How can we determine, if and how this is an issue for us at all?
How can the indexes be rebuilt or restored?
Unfortunately documentation about the elasticsearch component of nexus is quite scarce.
2017-06-12 00:00:00,206+0000 WARN [elasticsearch[0830B166-40C32B3F-29D5B1EA-31CA71C6-25CC4CED][local_transport][T#1]] *S
YSTEM org.elasticsearch.cluster.action.shard - [0830B166-40C32B3F-29D5B1EA-31CA71C6-25CC4CED] [a52b0b100dbae53dfc54b6233c
c368c57e8dca35][0] received shard failed for [a52b0b100dbae53dfc54b6233cc368c57e8dca35][0], node[ThNZvHC4RCqRuP9g14s8HQ],
[P], v[5], s[INITIALIZING], a[id=zQeMmOLyR6ui4XcTwDotWw], unassigned_info[[reason=ALLOCATION_FAILED], at[2017-06-11T23:5
9:59.851Z], details[failed recovery, failure IndexShardRecoveryException[failed to recovery from gateway]; nested: Engine
CreationFailureException[failed to create engine]; nested: EOFException; ]], indexUUID [jl80KDHmS-ijq14OsbwxHA], message
[master {0830B166-40C32B3F-29D5B1EA-31CA71C6-25CC4CED}{ThNZvHC4RCqRuP9g14s8HQ}{local}{local[1]}{local=true, master=true}
marked shard as initializing, but shard is marked as failed, resend shard failure], failure [Unknown]
2017-06-12 00:00:00,211+0000 WARN [elasticsearch[0830B166-40C32B3F-29D5B1EA-31CA71C6-25CC4CED][generic][T#1479]] *SYSTEM
org.elasticsearch.indices.cluster - [0830B166-40C32B3F-29D5B1EA-31CA71C6-25CC4CED] [[62d189601e138867db4d6cbfee3fc80fe18
7efe6][0]] marking and sending shard failed due to [failed recovery]
org.elasticsearch.index.shard.IndexShardRecoveryException: failed to recovery from gateway
at org.elasticsearch.index.shard.StoreRecoveryService.recoverFromStore(StoreRecoveryService.java:254) [na:na]
... 5 common frames omitted
Caused by: java.io.EOFException: null
at org.apache.lucene.store.InputStreamDataInput.readByte(InputStreamDataInput.java:37) [na:na]
at org.apache.lucene.store.DataInput.readInt(DataInput.java:101) [na:na]
at org.apache.lucene.store.DataInput.readLong(DataInput.java:157) [na:na]
at org.elasticsearch.index.translog.Checkpoint.<init>(Checkpoint.java:54) [na:na]
at org.elasticsearch.index.translog.Checkpoint.read(Checkpoint.java:83) [na:na]
at org.elasticsearch.index.translog.Translog.recoverFromFiles(Translog.java:337) [na:na]
at org.elasticsearch.index.translog.Translog.<init>(Translog.java:179) [na:na]
at org.elasticsearch.index.engine.InternalEngine.openTranslog(InternalEngine.java:209) [na:na]
at org.elasticsearch.index.engine.InternalEngine.<init>(InternalEngine.java:152) [na:na]
... 11 common frames omitted