You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Similar to elastic/elasticsearch#101180, which occured on an ElasticSearch cluster we were running, we would like to understand if the same issue can occur for OpenSearch.
Set the FS for the primary node, to readonly, via:
sudo mount -o remount,ro /dev/mapper/<volume_name>
Note: We were unable to reproduce this on our end, as we were unable to execute step 2 so far, as the mount point was always being used (translog writes?), but theoretically these are the steps that should reproduce the issue.
Expected behavior
A replica should have been auto-promoted when the primary node had an issue, rather than the entire shard becoming unavailable.
Additional Details
Plugins
Please list all plugins currently enabled.
Host/Environment (please complete the following information):
OS: Linux
Version: ES v7.10.2
Additional context
Timeline:
18:40 Cluster was in yellow state due to ongoing ingestion / replication in the cluster.
18:41 Primary shard tried to write metadata but failed with Exception occurred when storing metadata. Disc probably officially turned to readonly at this time.
18:41 Cluster turned to red state, implying no copies of the shard were found within the cluster anymore.
18:41:37 - Master node logs showed that the current state of the shard was closed, with no available replicas.
org.elasticsearch.index.shard.IndexShardClosedException: CurrentState[CLOSED] Replica unavailable - replica could have left ReplicationGroup or IndexShard might have closed
18:41:37 - Master node logs showed replica node failed to perform replication for a bulk write request
[master_node_name_placeholder] failing shard [failed shard, shard [index_name_placeholder][6], node[F8KJuNpISluhjrzZImpLFw], [R], s[STARTED], a[id=FrkhgR1WTrK9C6innxeYLw], message [failed to perform indices:data/write/bulk[s] on replica [index_name_placeholder][6], node[F8KJuNpISluhjrzZImpLFw], [R], s[STARTED], a[id=FrkhgR1WTrK9C6innxeYLw]], failure [IndexShardClosedException[CurrentState[CLOSED] Replica unavailable - replica could have left ReplicationGroup or IndexShard might have closed]], markAsStale [true]]
19:10
An engineer tried gracefully manual reroute with cluster reroute API , but got failure on replica unavailable due to indexshard close exception:
"6": [ -
{ -
"state": "UNASSIGNED",
"primary": true,
"node": null,
"relocating_node": null,
"shard": 6,
"index": "index_name_placeholder",
"recovery_source": { -
"type": "EXISTING_STORE",
"bootstrap_new_history_uuid": false
},
"unassigned_info": { -
"reason": "MANUAL_ALLOCATION",
"delayed": false,
"details": "failed shard on node [4NuDZt9TSueczPGeM1ANLQ]: failed to perform indices:data/write/bulk[s] on replica [index_name_placeholder][6], node[4NuDZt9TSueczPGeM1ANLQ], [R], s[STARTED], a[id=pGGuYwM7SLWOFu1ZZXy8ig], failure IndexShardClosedException[CurrentState[CLOSED] Replica unavailable - replica could have left ReplicationGroup or IndexShard might have closed]",
"allocation_status": "no_valid_shard_copy"
}
},
19:19 Added "accept_data_loss": true to the command, to ungracefully promote a "stale" replica shard to primary:
Describe the bug
Similar to elastic/elasticsearch#101180, which occured on an ElasticSearch cluster we were running, we would like to understand if the same issue can occur for OpenSearch.
Logs collected for the primary node and master node during the incident: https://docs.google.com/spreadsheets/d/1EkZeoGpMM_fDywgx0keOV-UQwWOaXZHYUuW2hstURv4/edit?gid=1305019963#gid=1305019963 Please keep that in mind it is unfortunately unclear whether these stack traces/logs are in order.
Related component
Search:Resiliency
To Reproduce
Note: We were unable to reproduce this on our end, as we were unable to execute step 2 so far, as the mount point was always being used (translog writes?), but theoretically these are the steps that should reproduce the issue.
Expected behavior
A replica should have been auto-promoted when the primary node had an issue, rather than the entire shard becoming unavailable.
Additional Details
Plugins
Please list all plugins currently enabled.
Screenshots
If applicable, add screenshots to help explain your problem.
Logs from the primary node hosting the shard and active master node: https://docs.google.com/spreadsheets/d/1EkZeoGpMM_fDywgx0keOV-UQwWOaXZHYUuW2hstURv4/edit?gid=1305019963#gid=1305019963 Please keep that in mind it is unfortunately unclear whether these stack traces/logs are in order.
Host/Environment (please complete the following information):
Additional context
Timeline:
18:40 Cluster was in yellow state due to ongoing ingestion / replication in the cluster.
18:41 Primary shard tried to write metadata but failed with
Exception occurred when storing metadata
. Disc probably officially turned to readonly at this time.18:41 Cluster turned to red state, implying no copies of the shard were found within the cluster anymore.
18:41:37 - Master node logs showed that the current state of the shard was closed, with no available replicas.
18:41:37 - Master node logs showed replica node failed to perform replication for a bulk write request
19:10
An engineer tried gracefully manual reroute with cluster reroute API , but got failure on
replica unavailable due to indexshard close exception
:19:19 Added
"accept_data_loss": true
to the command, to ungracefully promote a "stale" replica shard to primary:19:19 Cluster 5xx failures subsided, and the cluster state went back to green.
The text was updated successfully, but these errors were encountered: