-
-
Notifications
You must be signed in to change notification settings - Fork 229
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Redis can't recover after node is down with redis-replicas + HA sentinel #1052
Labels
bug
Something isn't working
Comments
Had a very similar incident after a node autoscaling event followed by a pod rebalance. One thing worth noticing is that my Redis replication nodes (non-master ones) printed a lot of logs like this:
While the master Redis node prints:
It looks like no Redis instances are receiving the correct replication config. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
V0.18.0
redis-operator version:
Does this issue reproduce with the latest release?
Yes
What operating system and processor architecture are you using (
kubectl version
)?Ubuntu 22 + kube
kubectl version
OutputWhat did you do?
Create local cluster with k3d
Create namespace
Configure helm
Create redis replication OT operator
Create redis sentinel
Create redis replication
At the end you should have 3 agents node where each one have 1 sentinel and 1 replica. 1 replica is the master and other are slaves.
then look for agent name where master redis is deployed
Test a chaos scenario where node with master is down.
What did you expect to see?
Sentinel agent + replica instance are redeployed and sentinel process new master election
What did you see instead?
Sentinel agent + replica instance are redeployed on a node that already have sentinel + replica instances.
Thus, no master election and each sentienl are stuck.
Below log from sentinel
The text was updated successfully, but these errors were encountered: