Updated on 2024-07-02 GMT+08:00

CCE Advanced HPA

cce-hpa-controller is a CCE-developed add-on, which can be used to flexibly scale in or out Deployments based on metrics such as CPU usage and memory usage.

After installing this add-on, you can create CustomedHPA policies. For details, see Creating a CustomedHPA Policy.

Main Functions

  • Scaling can be performed based on the percentage of the current number of pods.
  • The minimum scaling step can be set.
  • Different scaling operations can be performed based on the actual metric values.

Constraints

Installing the Add-on

  1. Log in to the CCE console and click the cluster name to access the cluster console. Click Add-ons in the navigation pane, locate CCE Advanced HPA on the right, and click Install.
  2. On the Install Add-on page, configure the specifications.

    Table 1 Add-on configuration

    Parameter

    Description

    Add-on Specifications

    Select Single or Custom for Add-on Specifications.

    NOTE:

    A single instance is solely for verification purposes. For commercial situations, you need to choose Custom based on the cluster specifications. The add-on specifications are influenced by the total number of containers in clusters and the number of scaling policies. For typical situations, it is recommended that you configure 500m CPU cores and 1,000 MiB of memory for every 5,000 containers. As for scaling policies, 100m CPU cores and 500 MiB of memory should be configured for every 1,000 of them.

    Pods

    Number of pods that will be created to match the selected add-on specifications.

    If you select Custom, you can adjust the number of pods as required.

    High availability is not possible with a single pod. If an error occurs on the node where the add-on instance runs, the add-on will fail.

    Containers

    CPU and memory quotas of the container allowed for the selected add-on specifications.

    If you select Custom, you can adjust the container specifications as required.

  3. Configure scheduling policies for the add-on.

    • Scheduling policies do not take effect on add-on instances of the DaemonSet type.
    • When configuring multi-AZ deployment or node affinity, ensure that there are nodes meeting the scheduling policy and that resources are sufficient in the cluster. Otherwise, the add-on cannot run.
    Table 2 Configurations for add-on scheduling

    Parameter

    Description

    Multi AZ

    • Preferred: Deployment pods of the add-on will be preferentially scheduled to nodes in different AZs. If all the nodes in the cluster are deployed in the same AZ, the pods will be scheduled to that AZ.
    • Equivalent mode: Deployment pods of the add-on are evenly scheduled to the nodes in the cluster in each AZ. If a new AZ is added, you are advised to increase add-on pods for cross-AZ HA deployment. With the Equivalent multi-AZ deployment, the difference between the number of add-on pods in different AZs will be less than or equal to 1. If resources in one of the AZs are insufficient, pods cannot be scheduled to that AZ.
    • Required: Deployment pods of the add-on will be forcibly scheduled to nodes in different AZs. If there are fewer AZs than pods, the extra pods will fail to run.

    Node Affinity

    • Not configured: Node affinity is disabled for the add-on.
    • Node Affinity: Specify the nodes where the add-on is deployed. If you do not specify the nodes, the add-on will be randomly scheduled based on the default cluster scheduling policy.
    • Specified Node Pool Scheduling: Specify the node pool where the add-on is deployed. If you do not specify the node pool, the add-on will be randomly scheduled based on the default cluster scheduling policy.
    • Custom Policies: Enter the labels of the nodes where the add-on is to be deployed for more flexible scheduling policies. If you do not specify node labels, the add-on will be randomly scheduled based on the default cluster scheduling policy.

      If multiple custom affinity policies are configured, ensure that there are nodes that meet all the affinity policies in the cluster. Otherwise, the add-on cannot run.

    Toleration

    Using both taints and tolerations allows (not forcibly) the add-on Deployment to be scheduled to a node with the matching taints, and controls the Deployment eviction policies after the node where the Deployment is located is tainted.

    The add-on adds the default tolerance policy for the node.kubernetes.io/not-ready and node.kubernetes.io/unreachable taints, respectively. The tolerance time window is 60s.

    For details, see Configuring Tolerance Policies.

  4. Click Install.

Components

Table 3 Add-on components

Component

Description

Resource Type

customedhpa-controller

CCE auto scaling component, which scales in or out Deployments based on metrics such as CPU usage and memory usage

Deployment

Change History

Table 4 Release history

Add-on Version

Supported Cluster Version

New Feature

1.4.2

v1.21

v1.23

v1.25

v1.27

v1.28

v1.29

CCE clusters 1.29 are supported.

1.3.43

v1.21

v1.23

v1.25

v1.27

v1.28

Fixed some issues.

1.3.42

v1.21

v1.23

v1.25

v1.27

v1.28

CCE clusters 1.28 are supported.

1.3.14

v1.19

v1.21

v1.23

v1.25

v1.27

CCE clusters 1.27 are supported.

1.3.10

v1.19

v1.21

v1.23

v1.25

Periodic scaling is not affected by the cooldown period.

1.3.7

v1.19

v1.21

v1.23

v1.25

Supported anti-affinity scheduling of pods on nodes in different AZs.

1.3.3

v1.19

v1.21

v1.23

v1.25

  • CCE clusters 1.25 are supported.
  • Allowed CronHPA to adjust the number of Deployments with the skip scenario supported.

1.3.1

v1.19

v1.21

v1.23

CCE clusters 1.23 are supported.

1.2.12

v1.15

v1.17

v1.19

v1.21

Optimizes the add-on performance to reduce resource consumption.

1.2.11

v1.15

v1.17

v1.19

v1.21

  • Enables the Kubernetes metrics API to obtain resource metrics.
  • Takes not-ready pods into consideration when calculating resource usage.

1.2.10

v1.15

v1.17

v1.19

v1.21

CCE clusters 1.21 are supported.

1.2.4

v1.15

v1.17

v1.19

  • Regular upgrade of add-on dependencies
  • Allows custom add-on resource specifications.

1.2.3

v1.15

v1.17

v1.19

Supports ARM64 nodes.

1.2.2

v1.15

v1.17

v1.19

Enhances the health check function.

1.2.1

v1.15

v1.17

v1.19

  • CCE clusters 1.19 are supported.
  • Updates the add-on to a stable version.

1.1.3

v1.15

v1.17

Supports periodic scaling rules.