AWS SSM Patch manager and EKS node groups

0

Hello,

I've been digging into the documentation but I cannot find and answer to my doubts. I've been trying to test it, but so far no reboots needed for my nodes.

I have an EKS cluster with two node groups. I'm using Patch manager (no quick setup). So far, I set-up my maintenance window to run weekly, to patch one node at time and to reboot if needed.

So my doubt is, when a node from the node group needs to be patched and restarted. Does the pods are moved to another node? what if my pod is busy? Is the node cordoned?

Thanks for your help.

1개 답변
2

The effectiveness of the draining process and the rescheduling of pods also depends on your application's specific setup and configuration. If you have multiple nodes and your pods have the appropriate tolerations and affinity rules, Kubernetes will reschedule your evicted pods to other healthy nodes. Before stopping an instance, AWS Systems Manager attempts to gracefully drain the instance. This involves cordoning the node to prevent new pods from being scheduled onto it and evicting existing pods to other nodes in the cluster. Pods are generally rescheduled to other healthy nodes during the draining process. Kubernetes tries to maintain the desired number of replicas for each deploymentYou can use features like PodDisruptionBudgets in Kubernetes to control the disruption caused by voluntary disruptions, such as draining nodes during maintenance.

Hope it clarifies and if does I would appreciate answer to be accepted so that community can benefit for clarity, thanks ;)

profile picture
전문가
답변함 3달 전
profile picture
전문가
Kallu
검토됨 3달 전

로그인하지 않았습니다. 로그인해야 답변을 게시할 수 있습니다.

좋은 답변은 질문에 명확하게 답하고 건설적인 피드백을 제공하며 질문자의 전문적인 성장을 장려합니다.

질문 답변하기에 대한 가이드라인

관련 콘텐츠