AWS SSM Patch manager and EKS node groups

0

Hello,

I've been digging into the documentation but I cannot find and answer to my doubts. I've been trying to test it, but so far no reboots needed for my nodes.

I have an EKS cluster with two node groups. I'm using Patch manager (no quick setup). So far, I set-up my maintenance window to run weekly, to patch one node at time and to reboot if needed.

So my doubt is, when a node from the node group needs to be patched and restarted. Does the pods are moved to another node? what if my pod is busy? Is the node cordoned?

Thanks for your help.

JosRiv
已提问 3 个月前337 查看次数
1 回答
2

The effectiveness of the draining process and the rescheduling of pods also depends on your application's specific setup and configuration. If you have multiple nodes and your pods have the appropriate tolerations and affinity rules, Kubernetes will reschedule your evicted pods to other healthy nodes. Before stopping an instance, AWS Systems Manager attempts to gracefully drain the instance. This involves cordoning the node to prevent new pods from being scheduled onto it and evicting existing pods to other nodes in the cluster. Pods are generally rescheduled to other healthy nodes during the draining process. Kubernetes tries to maintain the desired number of replicas for each deploymentYou can use features like PodDisruptionBudgets in Kubernetes to control the disruption caused by voluntary disruptions, such as draining nodes during maintenance.

Hope it clarifies and if does I would appreciate answer to be accepted so that community can benefit for clarity, thanks ;)

profile picture
专家
已回答 3 个月前
profile picture
专家
Kallu
已审核 3 个月前

您未登录。 登录 发布回答。

一个好的回答可以清楚地解答问题和提供建设性反馈,并能促进提问者的职业发展。

回答问题的准则