Skip to content

Node Group Upgrade Issue After AWS EKS Cluster Upgrade to v1.30

0

Problem Description:

We have upgraded our AWS EKS cluster from version 1.29 to 1.30 and also upgraded two node groups. One node group, which was created without a launch template, upgraded successfully. However, the other node group, which was created using a launch template, failed to upgrade with the following error:

Error code: NodeCreationFailure Error message: Couldn't proceed with upgrade process as new nodes are not joining node group abc-node-group

We noticed that the Node AMI version has been upgraded to 1.30, but it is not able to join the cluster.

Please help us resolve this issue. Let us know if any additional information is required.

asked 9 months ago591 views
1 Answer
1

This issue is likely related to compatibility problems between the upgraded EKS cluster and the node group configuration. Here are some steps to troubleshoot and resolve the problem:

  1. Check CNI plugin compatibility: Ensure that you have installed the correct version of the Amazon VPC CNI plugin that is compatible with Kubernetes 1.30. Verify the installed version and upgrade if necessary.

  2. Verify kube-proxy version: Make sure kube-proxy has been upgraded to the version compatible with Kubernetes 1.30 (v1.30.0-eksbuild.3 or later).

  3. Examine node logs: Check the logs on the new nodes that are failing to join the cluster. Look for specific error messages related to network plugin initialization or API server connectivity issues.

  4. Network connectivity: Verify that the security groups and network ACLs allow proper communication between the control plane and the nodes. Ensure that all necessary ports, especially 443 (HTTPS), are open for inbound and outbound traffic.

  5. DNS resolution: Confirm that DNS resolution for the API endpoint is working correctly on the new nodes.

  6. Launch template configuration: Since the node group created with a launch template is failing, review the launch template configuration. Ensure it's up to date and compatible with the new Kubernetes version.

  7. Update node group: Try updating the node group version using the AWS CLI or Management Console. This may trigger a reevaluation of the node group's status:

aws eks update-nodegroup-version --cluster-name your-cluster-name --nodegroup-name abc-node-group --kubernetes-version 1.30
  1. Delete and recreate: If updating doesn't work, consider deleting the problematic node group and recreating it with the correct configuration for Kubernetes 1.30.

  2. Check service quotas: Ensure that you haven't hit any service quotas that might prevent new nodes from being created or joining the cluster.

If these steps don't resolve the issue, you may need to contact AWS Support for further assistance, as there could be underlying problems specific to your cluster configuration.

Remember to take necessary precautions to maintain cluster stability and minimize downtime when making changes to your node groups.

Sources
Upgrade of AWS EKS Node group failed with 'CNI plugin not initialized' | AWS re:Post
Node group stuck at "Create failed" status | AWS re:Post
EKS problem after upgrade from 1.23 | AWS re:Post
Update a managed node group for your cluster - Amazon EKS

answered 9 months ago
EXPERT
reviewed 9 months ago
  • Followed above steps still same problem, FYI the Node group "Nodes" are in terminated state, even not able to connect via ssh to view the log, So is there any way to view the upgrade time generated log?

You are not logged in. Log in to post an answer.

A good answer clearly answers the question and provides constructive feedback and encourages professional growth in the question asker.