-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Prevent additional scheduling on Node - helps with system maintenance #1500
Comments
@olenm Isn't |
node-drain is excellent for removing nodes from a host, but often I am wanting to block jobs being pushed to a particular node to enforce jobs being pushed to another node. Using node-drain causes all containers on the node to migrate to other nodes, which is exactly what I want to prevent. |
I think #1523 this is what you want? |
@olenm: In this case, it really can't be put under the |
I tried to do this (prevent new jobs to be scheduled on a node) by setting reserved resources to currently allocated resources in config and reloading the agent config with SIGHUP, but it did not update the available resources in node-status. |
This would be incredibly powerful for upgrading the nomad-agent cluster, in removing the 'magic' of a Imagine a 3-node nomad-agent cluster via AWS ASG's. We want to upgrade the nomad binary by simply creating new ami's and a new ASG joined to the same nomad cluster. We have a dozen jobs, some with multiple group counts, and some are singles. The containers are currently across 2 of the 3 (older) nodes, and calling We are left with 2 (older) nodes with containers on it, calling a With a node-block status in place, blocking new jobs on the old agents would allow a deploy to occur and not interrupt service and also allow testing out new nomad-agent clusters added to an already existing cluster. A deploy would still be able to terminate the jobs on the blocked-agent after a successful deploy occurs on the new systems. |
Do you not add a new client node before calling |
yes, we use ASG's and would essentially (at least) double the size of the nomad-agent cluster |
@dadgar the node had degraded and was marked for termination by AWS, I wanted the long running jobs on it to finish and no new services/batch jobs to be scheduled there before I drain it. |
I wonder if adding the flag
possibly defaulting to |
Revamped node drain in 0.8 addresses this, closing |
I'm going to lock this issue because it has been closed for 120 days ⏳. This helps our maintainers find and focus on the active issues. |
A feature to stop new-jobs from starting on a select node would be incredibly useful.
This would allow for enhanced testing of new nomad clients and testing out infrastructure before having to deprecate old systems.
The text was updated successfully, but these errors were encountered: