Help Center/ Distributed Cache Service/ FAQs/ Instance Scaling and Upgrade/ Are Services Interrupted During Specification Modification?
Updated on 2024-10-31 GMT+08:00

Are Services Interrupted During Specification Modification?

Modify instance specifications during off-peak hours.

If the modification failed in peak hours (for example, when memory or CPU usage is over 90% or write traffic surges), try again during off-peak hours.

The following table describes the impact of specification modification.

Change of the Instance Type

Table 1 Instance type change options supported by different DCS instances

Version

Supported Type Change

Precautions

Redis 3.0

From single-node to master/standby

The instance cannot be connected for several seconds and remains read-only for about one minute.

From master/standby to Proxy Cluster

  1. If the data of a master/standby DCS Redis 3.0 instance is stored in multiple databases, or in non-DB0 databases, the instance cannot be changed to the Proxy Cluster type. A master/standby instance can be changed to the Proxy Cluster type only if its data is stored only on DB0.
  2. The instance cannot be connected and remains read-only for 5 to 30 minutes.

Memcached

From single-node to master/standby

Services are interrupted for several seconds and remain read-only for about 1 minute.

Redis 4.0/5.0/6.0

From master/standby or read/write splitting to Proxy Cluster

  1. Before changing the instance type to Proxy Cluster, evaluate the impact on services. For details, see What Are the Constraints on Implementing Multi-DB on a Proxy Cluster Instance? and Command Restrictions.
  2. Memory usage must be less than 70% of the maximum memory of the new flavor.
  3. Some keys may be evicted if the current memory usage exceeds 90% of the total.
  4. After the change, create alarm rules again for the instance.
  5. For instances that are currently master/standby, ensure that their read-only IP address or domain name is not used by your application.
  6. If your application cannot reconnect to Redis or handle exceptions, you may need to restart the application after the change.
  7. Modify instance specifications during off-peak hours. An instance is temporarily interrupted and remains read-only for about 1 minute during the specification change.

From Proxy Cluster to master/standby or read/write splitting

Redis 4.0/5.0/6.0

From master/standby to read/write splitting

NOTE:

Currently, a read/write splitting instance cannot be directly changed to a master/standby one.

  1. The instance memory must be greater than or equal to 4 GB, and will remain the same after the change.
  2. Some keys may be evicted if the current memory usage exceeds 90% of the total.
  3. After the change, create alarm rules again for the instance.
  4. Ensure that read-only IP addresses or domain names are not directly referred in the applications using the master/standby instance.
  5. If your application cannot reconnect to Redis or handle exceptions, you may need to restart the application after the change.
  6. Services may temporarily stutter during the change. Perform the change during off-peak hours.
  7. Unavailable for master/standby instances with ACL users.
  8. Unavailable for master/standby DCS Redis 6.0 instances with SSL enabled.

Any instance type changes not listed in the preceding table are not supported. To modify specifications while changing the instance type, see Online Migration Between Instances.

For details about the commands supported by different types of instances, see Command Compatibility.

Scaling

  • Scaling options
    Table 2 Scaling options supported by different instances

    Cache Engine

    Single-Node

    Master/Standby

    Redis Cluster

    Proxy Cluster

    Read/Write Splitting

    Redis 3.0

    Scaling up/down

    Scaling up/down

    -

    Scaling out

    -

    Redis 4.0

    Scaling up/down

    Scaling up/down and replica quantity change

    Scaling up/down, out/in, and replica quantity change

    Scaling up/down, out/in

    Scaling up/down and replica quantity change

    Redis 5.0

    Scaling up/down

    Scaling up/down and replica quantity change

    Scaling up/down, out/in, and replica quantity change

    Scaling up/down, out/in

    Scaling up/down and replica quantity change

    Redis 6.0 basic edition

    Scaling up/down

    Scaling up/down and replica quantity change

    Scaling up/down, out/in, and replica quantity change

    Scaling up/down, out/in

    Scaling up/down and replica quantity change

    Redis 6.0 professional editions

    -

    Scaling up/down

    -

    -

    -

    Memcached

    Scaling up/down

    Scaling up/down

    -

    -

    -

    • If the reserved memory of a DCS Redis 3.0 or Memcached instance is insufficient, the modification may fail when the memory is used up. For details, see Reserved Memory.
    • Change the replica quantity and capacity separately.
    • Only one replica can be deleted per operation.
  • Impact of scaling
    Table 3 Impact of scaling

    Instance Type

    Scaling Type

    Impact

    Single-node , read/write splitting, and master/standby

    Scaling up/down

    • During scaling up, a basic edition DCS Redis 4.0 or later instance will be disconnected for several seconds and remain read-only for about 1 minute. During scaling down, connections will not be interrupted.
    • A DCS Redis 3.0 instance will be disconnected for several seconds and remain read-only for 5 to 30 minutes.
    • A DCS Redis professional edition instance will be disconnected for several seconds and remain read-only for about 1 minute.
    • For scaling up, only the memory of the instance is expanded. The CPU processing capability is not improved.
    • Single-node DCS instances do not support data persistence. Scaling may compromise data reliability. After scaling, check whether the data is complete and import data if required. If there is important data, use a migration tool to migrate the data to other instances for backup.
    • For master/standby and read/write splitting instances, backup records created before scale-down cannot be used after scale-down. If necessary, download the backup file in advance or back up the data again after scale-down.

    Proxy Cluster and Redis Cluster

    Scaling up/down

    • Scaling out by adding shards:
      • Scaling out does not interrupt connections but will occupy CPU resources, decreasing performance by up to 20%.
      • If the shard quantity increases, new Redis Server nodes are added, and data is automatically balanced to the new nodes, increasing the access latency.
    • Scaling in by reducing shards:
      • If the shard quantity decreases, nodes will be deleted. Before scaling in a Redis Cluster instance, ensure that the deleted nodes are not directly referenced in your application, to prevent service access exceptions.
      • Nodes will be deleted, and connections will be interrupted. If your application cannot reconnect to Redis or handle exceptions, you may need to restart the application after scaling.
    • Scaling up by shard size without changing the shard quantity:
      • Insufficient memory of the node's VM will cause the node to migrate. Service connections may stutter and the instance may become read-only during the migration.
      • Increasing the node capacity when the VM memory is sufficient does not affect services.
      NOTE:

      Cluster DCS Redis 3.0 instances cannot be vertically scaled.

    • Scaling down by reducing the shard size without changing the shard quantity has no impact.
    • To scale down an instance, ensure that the used memory of each node is less than 70% of the maximum memory per node of the new flavor.
    • The flavor changing operation may involve data migration, and the latency may increase. For a Redis Cluster instance, ensure that the client can process the MOVED and ASK commands. Otherwise, the request will fail.
    • If the memory becomes full during scaling due to a large amount of data being written, scaling will fail.
    • Before scaling, check for big keys through Cache Analysis. Redis has a limit on key migration. If the instance has any single key greater than 512 MB, scaling will fail when big key migration between nodes times out. The bigger the key, the more likely the migration will fail.
    • Before scaling a Redis Cluster instance, ensure that automated cluster topology refresh is enabled. If it is disabled, you will need to restart the client after scaling. For details about how to enable automated refresh if you use Lettuce, see an example of using Lettuce to connect to a Redis Cluster instance.
    • Backup records created before scaling cannot be used. If necessary, download the backup file in advance or back up the data again after scaling.

    Master/Standby, read/write splitting, and Redis Cluster instances

    Scaling out/in (replica quantity change)

    • Before adding or removing replicas for a Redis Cluster instance, ensure that automated cluster topology refresh is enabled. If it is disabled, you will need to restart the client after scaling. For details about how to enable automated refresh if you use Lettuce, see an example of using Lettuce to connect to a Redis Cluster instance.
    • Deleting replicas interrupts connections. If your application cannot reconnect to Redis or handle exceptions, you may need to restart the application after scaling. Adding replicas does not interrupt connections.
    • If the number of replicas is already the minimum supported by the instance, you can no longer delete replicas.