You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I manage bare metal cluster and suffer from downtime when nodes failing
Worker node failed - status become NotReady but traffic still come to the pods that was on this node. Often this leads to downtime if there was critical components. Now I need to delete this node from cluster to reshedule pods. Nothing special for your project.
Control plane node failed - that was only one time. I have 3 control plane nodes. After one node failed i immediatly deleted it from cluster. But nothing works because kubedns stoped resolving internal names. As I understand this happend because remaing etcd daemons didn't reach an agreement who become a master. Only systemctl restart docker on one of two working control plane nodes resolves an issue. So, could kube-fencing be used on control-plane nodes as well?
The text was updated successfully, but these errors were encountered:
I manage bare metal cluster and suffer from downtime when nodes failing
systemctl restart docker
on one of two working control plane nodes resolves an issue. So, couldkube-fencing
be used on control-plane nodes as well?The text was updated successfully, but these errors were encountered: