After single Node failure in a HA deployment Scheduler and Controller Manager are not working properly. When checking docker logs on a Node last messages are
1 leaderelection.go:213] failed to renew lease kube-system/kube-controller-manager: timed out waiting for the condition
1 controllermanager.go:215] leaderelection lost
When trying to restart docker container kube-controller-manager nothing happens.
If restarting the container does nothing, there is something else going on. docker restart kube-controller-manager should restart the container and you should get logging from the startup.
The only problem I have caused (and why I write this reply) is that I haven’t drained the compromised node before reloading which is silly but serious mistake.