This repository has been archived by the owner on Dec 24, 2019. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 21
Node draining during instance shutdown leads to "flapping" state #7
Labels
Comments
Another scenario where a single instance (i-02e6f917c5d3bd5d8) is terminated manually:
|
hjacobs
added a commit
that referenced
this issue
Feb 11, 2017
hjacobs
added a commit
that referenced
this issue
Feb 11, 2017
Unschedulable nodes which are terminating (i.e. the There is still the problem that the readiness of nodes is not properly evaluated, i.e. |
Closing this and created a follow-up issue for defining the node "readiness" concept: #23 |
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Using autoscaling without a proper node shutdown sequence will kill all pods/containers without any grace period. Node draining such as
kube-node-drainer.service
should generally be recommended to avoid service disruptions: zalando-incubator/kubernetes-on-aws#257Problem with node draining: the autoscaler currently goes in "flapping" state as cordoned nodes (nodes marked as unschedulable) are compensated. Example:
kube-aws-autoscaler
figures out newDesiredCapacity
and scales down from 6 to 5kube-node-drainer.service
on the EC2 instance callskubectl drain
kube-aws-autoscaler
now sees 6 nodes, but one of them is cordoned, so it compensates to 7 nodeskube-aws-autoscaler
sets ASGDesiredCapacity
to 7The text was updated successfully, but these errors were encountered: