Change or Add VM to worker pool - Failure after Master nodes change VM SKU #324
Replies: 3 comments 8 replies
-
@vitobotta , probably related to this one? I can connect directly via SSH, and the auth.log is:
with ssh_agent = enabled and port = 22:
my config: (again, this is an old existent cluster), 9 months ago we were able to add a new worker pool, but it was before the VM SKU change.
|
Beta Was this translation helpful? Give feedback.
-
I am planning to fix the SKU issue by removing it from the instance names for new clusters. Unfortunately for existing clusters there is not much that can be done. I mean for worker nodes you can just add new pools with the new instance type, migrate stuff and delete the old node pools if you want, but for masters it's not so easy. |
Beta Was this translation helpful? Give feedback.
-
It does not affect the cluster, on my understanding as it uses mainly the
private IP.
Yes, it is a requirement to shutdown the VM.
That is why I suggest doing it at deploy :)
Romulo II Melillo
*Mobile:* +44 07367437841
*Skype:* rmelilloii
Linkedin <https://uk.linkedin.com/in/rmelilloii>
…On Fri, 12 Apr 2024 at 17:14, Vito Botta ***@***.***> wrote:
On workaround for issue 2, if you remove the ip and assign a new one does
it affect the functionality of the node to the cluster or not at all? Does
it require rebooting the node?
—
Reply to this email directly, view it on GitHub
<#324 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AG3HX5D2XVXMQ26XBID5D73Y5AB6XAVCNFSM6AAAAABD73DP5WVHI2DSMVQWIX3LMV43SRDJONRXK43TNFXW4Q3PNVWWK3TUHM4TAOJYGA2TK>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
Hello and good morning/afternoon!
Opening a dedicated thread here as I believe it deserves dedicated attention. @vitobotta , I will delete my comment on the discussion 158. Any help or direction are very welcome.
Case: Hetzner Cloud decommissioned old VM SKU that was in use by our Master nodes. Each Master node had the VM SKU changed via Hetzner Cloud Console. After that, we can't add a new worker pool.
Details so far:
FailedToCreateRoute Node rome-test-cx21-master1 (combined from similar events): Could not create route 4c1fd58a-028d-4fc4-b706-ac9212a922ff 10.244.0.0/24 for node rome-test-cx21-master1 after 1.933111343s: hcloud/CreateRoute: hcops/AllServersCache.ByName: rome-test-cx21-master1 hcops/AllServersCache.getCache: not found
To try to "fix it" I am trying to change the name on labels and annotations.
My question is @vitobotta, is there anything to do on my case or the only solution is to deploy the new nodes manually?
I have the impression that IF the VM SKU is not added to the VM name during deploy, this would not be an issue.
additional info: I think this is a valid workaround for my case:
A: to be able to deploy I need to inform the "new" existent VM SKU on the deploy yaml for hetzner-k3s
B: to avoid that a new set of masters are deployed I need to change the VM name in hetzner
C: I can deploy the new VMs using hetzner-k3s, it will deploy only the new stuff as it will be able to find all the "pre-existent VMs"
D: I change the VM name back to fix the Kubernetes route issue
This for some reason gave me timeout on the SSH connectivity validation (worked on a test cluster)
With the timeout my next step was to comment out the other worker pools, leaving only "master" and the new worker pool. Then, I got the new VMs deployed, got the SSH validation, but it times out there.
Beta Was this translation helpful? Give feedback.
All reactions