Pre-upgrade Check
The system automatically checks a cluster before its upgrade. If the cluster does not meet the pre-upgrade check conditions, the upgrade cannot continue. To avoid risks, you can perform pre-upgrade check according to the check items and solutions described in this section.
No. |
Check Item |
Description |
---|---|---|
1 |
|
|
2 |
Check whether the target cluster is under upgrade management. |
|
3 |
|
|
4 |
Check whether the current HelmRelease record contains discarded Kubernetes APIs that are not supported by the target cluster version. If yes, the Helm chart may be unavailable after the upgrade. |
|
5 |
Check whether CCE can connect to your master nodes. |
|
6 |
Check the node pool status. |
|
7 |
Check whether the Protocol & Port of the worker node security groups is set to ICMP: All and whether the security group with the source IP address set to the master node security group is deleted. |
|
8 |
|
|
9 |
Check whether nodes need to be migrated. |
|
10 |
Check whether there are discarded resources in the clusters. |
|
11 |
Read the version compatibility differences and ensure that they are not affected. The patch upgrade does not involve version compatibility differences. |
|
12 |
Check whether cce-agent on the current node is of the latest version. |
|
13 |
Check whether the CPU usage of the node exceeds 90%. |
|
14 |
|
|
15 |
|
|
16 |
|
|
17 |
Check whether the owner and owner group of the files in the /var/paas directory used by the CCE are both paas. |
|
18 |
Check whether the kubelet on the node is running properly. |
|
19 |
Check whether the memory usage of the node exceeds 90%. |
|
20 |
Check whether the clock synchronization server ntpd or chronyd of the node is running properly. |
|
21 |
Check whether the OS kernel version of the node is supported by CCE. |
|
22 |
Check whether the number of CPUs on the master node is greater than 2. |
|
23 |
Check whether the Python commands are available on a node. |
|
24 |
|
|
25 |
Check whether the nodes in the cluster are ready. |
|
26 |
Check whether journald of a node is normal. |
|
27 |
Check whether the containerd.sock file exists on the node. This file affects the startup of container runtime in the Euler OS. |
|
28 |
Before the upgrade, check whether an internal error occurs. |
|
29 |
Check whether inaccessible mount points exist on the node. |
|
30 |
Check whether the taint needed for cluster upgrade exists on the node. |
|
31 |
Check whether there are any compatibility restrictions on the current Everest add-on. |
|
32 |
Check whether the current cce-controller-hpa add-on has compatibility restrictions. |
|
33 |
Check whether the current cluster version and the target version support enhanced CPU policy. |
|
34 |
Check whether the container runtime and network components on the worker nodes are healthy. |
|
35 |
Check whether the Kubernetes, container runtime, and network components of the master nodes are healthy. |
|
36 |
Check whether the resources of Kubernetes components, such as etcd and kube-controller-manager, exceed the upper limit. |
|
37 |
The system scans the audit logs of the past day to check whether the user calls the deprecated APIs of the target Kubernetes version.
NOTE:
Due to the limited time range of audit logs, this check item is only an auxiliary method. APIs to be deprecated may have been used in the cluster, but their usage is not included in the audit logs of the past day. Check the API usage carefully. |
|
38 |
Check whether NetworkManager of a node is normal. |
|
39 |
Check the ID file format. |
|
40 |
When you upgrade a cluster to v1.19 or later, the system checks whether the following configuration files have been modified on the backend: |
|
41 |
Check whether the configuration files of key components exist on the node. |
|
42 |
Check whether the current CoreDNS key configuration Corefile is different from the Helm release record. The difference may be overwritten during the add-on upgrade, affecting domain name resolution in the cluster. |
|
43 |
Check whether the sudo commands and sudo-related files of the node are working. |
|
44 |
Whether some key commands that the node upgrade depends on are working |
|
45 |
Check whether the docker/containerd.sock file is directly mounted to the pods on a node. During an upgrade, Docker or containerd restarts and the sock file on the host changes, but the sock file mounted to pods does not change accordingly. As a result, your services cannot access Docker or containerd due to sock file inconsistency. After the pods are rebuilt, the sock file is mounted to the pods again, and the issue is resolved accordingly. |
|
46 |
Check whether the certificate used by an HTTPS load balancer has been modified on ELB. |
|
47 |
Check whether the default mount directory and soft link on the node have been manually mounted or modified. |
|
48 |
Check whether user paas is allowed to log in to a node. |
|
49 |
Check whether the load balancer associated with a Service is allocated with a private IPv4 address. |
|
50 |
Check whether the source version of the cluster is earlier than v1.11 and the target version is later than v1.23. |
|
51 |
Check whether the CIDR block of the cluster management plane is the same as that configured on the backbone network. |
|
52 |
The GPU add-on is involved in the upgrade, which may affect the GPU driver installation during the creation of a GPU node. |
|
53 |
Check whether the default system parameter settings on your nodes are modified. |
|
54 |
Check whether there are residual package version data in the current cluster. |
|
55 |
Check whether the commands required for the upgrade are available on the node. |
|
56 |
Check whether swap has been enabled on cluster nodes. |
|
57 |
Check whether there are compatibility issues that may occur during nginx-ingress upgrade. |
|
58 |
During a cluster upgrade, compatibility issues occur when the Cloud Native Cluster Monitoring add-on is upgraded from a version earlier than 3.9.0 to a version later than 3.9.0. Check whether Grafana is enabled for this add-on. |
|
59 |
Check whether the service pods running on a containerd node are restarted when containerd is upgraded. |
|
60 |
Check whether the configuration of the CCE AI Suite add-on in a cluster has been intrusively modified. If so, upgrading the cluster may fail. |
|
61 |
Check whether GPU or NPU service pods are rebuilt in a cluster when kubelet is restarted during the upgrade of the cluster. |
|
62 |
Check whether the access control of the ELB listener has been configured for the Service in the current cluster using annotations and whether the configurations are correct. |
|
63 |
Check whether the flavor of the master nodes in the cluster is the same as the actual flavor of these nodes. |
|
64 |
Check whether the number of available IP addresses in the cluster subnet supports rolling upgrade. |
|
65 |
Check whether an alarm is generated when a cluster is upgraded to v1.27 or later. Do not use Docker in clusters of versions later than 1.27 because CCE is going to stop the support for Docker. |
|
66 |
Check whether an alarm is generated when a cluster is upgraded to v1.27 or later. Do not use Docker in clusters of versions later than 1.27 because CCE is going to stop the support for Docker. |
|
67 |
Check the number of images on your node. If there are more than 1000 images, it takes a long time for Docker to start, affecting the standard Docker output and functions such as Nginx. |
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.See the reply and handling status in My Cloud VOC.
For any further questions, feel free to contact us through the chatbot.
Chatbot