Updated on 2024-11-12 GMT+08:00

Node Pool Overview

Introduction

CCE introduces node pools to help you better manage nodes in Kubernetes clusters. A node pool contains one node or a group of nodes with identical configuration in a cluster.

You can create custom node pools on the CCE console. With node pools, you can quickly create, manage, and destroy nodes without affecting the cluster. All nodes in a custom node pool have identical parameters and node type. You cannot configure a single node in a node pool; any configuration changes affect all nodes in the node pool.

You can also use node pools for auto scaling (supported only by pay-per-use node pools).

  • When a pod in a cluster cannot be scheduled due to insufficient resources, scale-out can be automatically triggered.
  • When there is an idle node or a monitoring metric threshold is met, scale-in can be automatically triggered.

This section describes how node pools work in CCE and how to create and manage node pools.

Node Pool Architecture

Figure 1 Overall architecture of a node pool

Generally, all nodes in a node pool have the following same attributes:

  • Node OS
  • Node login mode
  • Node container runtime
  • Enterprise project
  • Startup parameters of Kubernetes components on a node
  • Custom startup script of a node
  • Kubernetes labels and taints

CCE provides the following extended attributes for node pools:

  • Node pool OS
  • Maximum number of pods on each node in a node pool

Notes on a Node Pool Upgrade

For clusters of v1.21.11-r0, v1.23.9-r0, v1.25.4-r0, or later versions, new node pools support the creation of both pay-per-use and yearly/monthly nodes by default, providing you with better resource management experience.

The highlights of new node pools are as follows:

  • Both pay-per-use and yearly/monthly nodes can be created in one node pool, and the required duration of yearly/monthly nodes can be different.
  • Pay-per-use and yearly/monthly nodes can be accepted for management.
  • Auto scaling can be enabled and various scaling policies can be configured for more efficient, flexible resource management.

The following table lists the changes after an existing node pool is upgraded.

Billing Mode of the Original Node Pool

Changes After an Upgrade

Pay-per-use

All capabilities of the original node pool will be automatically inherited.

Yearly/Monthly nodes created in the new node pool cannot be manually scaled in. They can only be deleted or unsubscribed from.

Yearly/Monthly

The original node pool can be hitlessly upgraded, without affecting the existing nodes in the node pool.

After the upgrade, the following changes will occur:

  • Nodes created in the new node pools will be billed on a pay-per-use basis by default.
  • Yearly/Monthly nodes created in the new node pool cannot be manually scaled in. They can only be deleted or unsubscribed from.
  • Auto scaling can be enabled. By default, only pay-per-use nodes will be scaled by default.
  • The API for updating a node pool cannot be used to create yearly/monthly nodes. To create yearly/monthly nodes, use the API for resizing a node pool.

Description of DefaultPool

DefaultPool is not a real node pool. It only classifies nodes that are not in the custom node pools. These nodes are directly created on the console or by calling APIs. DefaultPool does not support any user-created node pool functions, including scaling and parameter configuration. DefaultPool cannot be edited, deleted, expanded, or auto scaled, and nodes in it cannot be migrated.

Applicable Scenarios

When a large-scale cluster is required, you are advised to use node pools to manage nodes.

The following table describes multiple scenarios of large-scale cluster management and the functions of node pools in each scenario.

Table 1 Using node pools for different management scenarios

Scenario

Function

Multiple heterogeneous nodes (with different models and configurations) in the cluster

Nodes can be grouped into different pools for management.

Frequent node scaling required in a cluster

Node pools support auto scaling to dynamically add or reduce nodes.

Complex application scheduling rules in a cluster

Node pool tags can be used to quickly specify service scheduling rules.

Functions and Precautions

Function

Description

Precaution

Creating a node pool

Add a node pool.

It is recommended that a cluster contains no more than 100 node pools.

Deleting a node pool

Deleting a node pool will delete nodes in the pool. Pods on these nodes will be automatically migrated to available nodes in other node pools.

If pods in the node pool have a specific node selector and none of the other nodes in the cluster satisfies the node selector, the pods will become unschedulable.

Enabling auto scaling for a node pool

After auto scaling is enabled, nodes will be automatically created or deleted in the node pool based on the cluster loads.

Do not store important data on nodes in a node pool because the nodes may be deleted after scale-in. Data on the deleted nodes cannot be restored.

Disabling auto scaling for a node pool

After auto scaling is disabled, the number of nodes in a node pool will not automatically change with the cluster loads.

None

Adjusting the size of a node pool

The number of nodes in a node pool can be directly adjusted. If the number of nodes is reduced, nodes are randomly removed from the current node pool.

After auto scaling is enabled, you are not advised to manually adjust the node pool size.

Modifying node pool configurations

You can change the node pool name and number of nodes, add or delete Kubernetes labels, resource tags, and taints, and adjust node pool configurations such as the disk, OS, and container engine of the node pool.

The deleted or added Kubernetes labels and taints (as well as their quantity) will apply to all nodes in the node pool, which may cause pod re-scheduling. Therefore, exercise caution when performing this operation.

Removing a node from a node pool

Nodes in a node pool can be migrated to the default node pool of the same cluster.

Nodes in the default node pool cannot be migrated to other node pools, and nodes in a user-created node pool cannot be migrated to other user-created node pools.

Copying a node pool

You can copy the configuration of an existing node pool to create a new node pool.

None

Setting Kubernetes parameters

You can configure core components with fine granularity.

  • This function is supported only in clusters of v1.15 and later. It is not displayed for versions earlier than v1.15.
  • The default node pool does not support this type of configuration.

Deploying a Workload in a Specified Node Pool

When creating a workload, you can constrain pods to run in a specified node pool.

For example, on the CCE console, you can set the affinity between the workload and the node on the Scheduling Policies tab page on the workload details page to forcibly deploy the workload to a specific node pool. In this way, the workload runs only on nodes in the node pool. To better control where the workload is to be scheduled, you can use affinity or anti-affinity policies between workloads and nodes described in Configuring Node Affinity Scheduling (nodeAffinity).

For example, you can use container's resource request as a nodeSelector so that workloads will run only on the nodes that meet the resource request.

If the workload definition file defines a container that requires four CPUs, the scheduler will not choose the nodes with two CPUs to run workloads.

Related Operations

You can log in to the CCE console and refer to the following sections to perform operations on node pools: