As described in this other thread, I am trying to re-add a node to my cluster after removing it.
The process I followed so far is this:
- Drain the node via Rancher UI.
- Delete the node via Rancher UI.
- Wait for cluster to stabilize.
- Reboot the node.
- Run through the manual cleanup steps documented here
- Reboot the nodes again (last step of manual cleanup process).
- Attempt to re-add the nodes (
docker run ...
)
The node shows as Unavailble in the Rancher UI with the error message “Kubelet stopped posting node status.” The Roles shown in Rancher UI match what I specified when attempting to re-add the node.
From the command line kubectl get nodes
also lists the node, but the ROLES
are listed as <none>
and the AGE
is about 15 minutes larger than it should be based on when I ran the docker run...
command.
docker ps
run on the node shows that the only container running is rancher/rancher-agent
. docker logs
shows the following message over and over again:
I have tried searching the forums but did not find anything that seemed relevant to my situation.
The unusual AGE
shown in kubectl
output makes me wonder if maybe rebooting the node before purging the Docker bits caused the node to attempt to re-register itself automatically. Is this a possibility?
Any insight into what might have caused the issue or how I can resolve it would be greatly appreciated.