Can I stop a "broken" etcd from communicating with "dead" nodes? #15885
-
I started with three etcd nodes, which were part of a Kubernetes cluster. I have since lost two of these nodes (let's call them "14" and "30") (i.e. data's either been lost or the machine has been erased) (let's not go into that). Now I am left with one "broken" I use the term "broken" because: -
The etcd container isn't crashing but it's repeatedly issuing the following to stdout/log relating to nodes
I am unable to issue any
I assume that one etcd node should still be able to operate?
|
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
I have found a solution for my scenario! It's inspired by the articles breaking-down-and-fixing-etcd-cluster and how-to-start-a-stopped-docker-container-with-a-different-command. The Loss of quorum section in the first tells me I probably need the Once I do this my etcd service starts as a single node and continues to operate as normal and (crucially) givens me
|
Beta Was this translation helpful? Give feedback.
I have found a solution for my scenario!
It's inspired by the articles breaking-down-and-fixing-etcd-cluster and how-to-start-a-stopped-docker-container-with-a-different-command.
The Loss of quorum section in the first tells me I probably need the
--force-new-cluster
flag in my etcd container. To do this I use the second article to illustrate how to stop the container, edit the JSON (the run-command arguments in my case), which contains the start command and then restart docker and the container.Once I do this my etcd service starts as a single node and continues to operate as normal and (crucially) givens me
etcdctl
again...