スキップしてコンテンツを表示

vpc-cni stuck in Updating status

0

I am trying to update the version of the vpc-cni add-on on several EKS clusters and I am getting a range of results. On one cluster, the upgrade goes smoothly and completes. On another, the status remains on "Updating" despite the daemonset in the cluster being updated to the new version images and they are running without any errors. The clusters all started from the same cluster and add-on versions.

Cluster version: 1.29 (at time of upgrade, now running 1.30) Add-on version: v1.16.x -> 1.17.1

It has been in this state for over a week. What might be the issue/how do I reset this? Thanks.

質問済み 1年前2083ビュー
3回答
0

Well now it is stuck in "Deleting" status after trying to delete the add-on while retaining the resources. I am also unable to get it to delete including the existing resources.

回答済み 1年前
  • Hello,

    you can resolve it by first trying to force delete the add-on using the AWS CLI command aws eks delete-addon --cluster-name <cluster_name> --addon-name <addon_name> --force. If that doesn't work, manually delete any Kubernetes resources related to the add-on using kubectl, and check for stuck finalizers on those resources. After cleaning up the resources, retry the deletion.

  • I have resolved the issue now by deleting the aws-node daemonset. Although this did create downtime on the cluster. It was a staging cluster so no issue there but this would not be a suitable option for a production cluster. How would this be resolved with zero downtime?

0

Use the AWS CLI to describe the add-on and check its status.

aws eks describe-addon --cluster-name your-cluster-name --addon-name vpc-cni

Sometimes, there might be pending updates or dependencies that need to be resolved before the add-on can fully update.

Run

kubectl get daemonset aws-node -n kube-system

Use kubectl to check for any events that might indicate why the update is stuck.

kubectl get events -n kube-system

You can try to manually re-trigger the update by reapplying the same version or trying a rollback and then applying the update again.

First, roll back to the previous version:

aws eks update-addon --cluster-name your-cluster-name --addon-name vpc-cni --resolve-conflicts=overwrite --addon-version previous-version

Then, try updating to the desired version again:

aws eks update-addon --cluster-name your-cluster-name --addon-name vpc-cni --resolve-conflicts=overwrite --addon-version 1.17.1

AWS
エキスパート
回答済み 1年前
エキスパート
レビュー済み 1年前
0

Hello,

the stuck VPC CNI add-on using the AWS CLI, then reinstall it with the correct version. After that, verify that all aws-node pods are running smoothly in the cluster.

please look at Document link you will get more information.

https://repost.aws/knowledge-center/eks-plan-upgrade-cluster

エキスパート
回答済み 1年前
エキスパート
レビュー済み 1年前

ログインしていません。 ログイン 回答を投稿する。

優れた回答とは、質問に明確に答え、建設的なフィードバックを提供し、質問者の専門分野におけるスキルの向上を促すものです。

関連するコンテンツ