跳至內容

vpc-cni stuck in Updating status

0

I am trying to update the version of the vpc-cni add-on on several EKS clusters and I am getting a range of results. On one cluster, the upgrade goes smoothly and completes. On another, the status remains on "Updating" despite the daemonset in the cluster being updated to the new version images and they are running without any errors. The clusters all started from the same cluster and add-on versions.

Cluster version: 1.29 (at time of upgrade, now running 1.30) Add-on version: v1.16.x -> 1.17.1

It has been in this state for over a week. What might be the issue/how do I reset this? Thanks.

已提問 1 年前檢視次數 2075 次
3 個答案
0

Well now it is stuck in "Deleting" status after trying to delete the add-on while retaining the resources. I am also unable to get it to delete including the existing resources.

已回答 1 年前
  • Hello,

    you can resolve it by first trying to force delete the add-on using the AWS CLI command aws eks delete-addon --cluster-name <cluster_name> --addon-name <addon_name> --force. If that doesn't work, manually delete any Kubernetes resources related to the add-on using kubectl, and check for stuck finalizers on those resources. After cleaning up the resources, retry the deletion.

  • I have resolved the issue now by deleting the aws-node daemonset. Although this did create downtime on the cluster. It was a staging cluster so no issue there but this would not be a suitable option for a production cluster. How would this be resolved with zero downtime?

0

Use the AWS CLI to describe the add-on and check its status.

aws eks describe-addon --cluster-name your-cluster-name --addon-name vpc-cni

Sometimes, there might be pending updates or dependencies that need to be resolved before the add-on can fully update.

Run

kubectl get daemonset aws-node -n kube-system

Use kubectl to check for any events that might indicate why the update is stuck.

kubectl get events -n kube-system

You can try to manually re-trigger the update by reapplying the same version or trying a rollback and then applying the update again.

First, roll back to the previous version:

aws eks update-addon --cluster-name your-cluster-name --addon-name vpc-cni --resolve-conflicts=overwrite --addon-version previous-version

Then, try updating to the desired version again:

aws eks update-addon --cluster-name your-cluster-name --addon-name vpc-cni --resolve-conflicts=overwrite --addon-version 1.17.1

AWS
專家
已回答 1 年前
專家
已審閱 1 年前
0

Hello,

the stuck VPC CNI add-on using the AWS CLI, then reinstall it with the correct version. After that, verify that all aws-node pods are running smoothly in the cluster.

please look at Document link you will get more information.

https://repost.aws/knowledge-center/eks-plan-upgrade-cluster

專家
已回答 1 年前
專家
已審閱 1 年前

您尚未登入。 登入 去張貼答案。

一個好的回答可以清楚地回答問題並提供建設性的意見回饋,同時有助於提問者的專業成長。