更新时间:2025-08-12 GMT+08:00
分享

查询Kafka集群元数据信息

功能介绍

查询Kafka集群元数据信息。

调用方法

请参见如何调用API

URI

GET /v2/{project_id}/instances/{instance_id}/management/cluster

表1 路径参数

参数

是否必选

参数类型

描述

project_id

String

参数解释

项目ID,获取方式请参见获取项目ID

约束限制

不涉及。

取值范围

不涉及。

默认取值

不涉及。

instance_id

String

参数解释

实例ID。获取方法如下:调用“查询所有实例列表”接口,从响应体中获取实例ID。

约束限制

不涉及。

取值范围

不涉及。

默认取值

不涉及。

请求参数

响应参数

状态码:200

表2 响应Body参数

参数

参数类型

描述

cluster

cluster object

参数解释

集群基本信息。

表3 cluster

参数

参数类型

描述

controller

String

参数解释

控制器ID。

取值范围

不涉及。

brokers

Array of brokers objects

参数解释

节点列表。

topics_count

Integer

参数解释

主题数量。

取值范围

不涉及。

partitions_count

Integer

参数解释

分区数量。

取值范围

不涉及。

online_partitions_count

Integer

参数解释

在线分区数量。

取值范围

不涉及。

replicas_count

Integer

参数解释

副本数量。

取值范围

不涉及。

isr_replicas_count

Integer

参数解释

ISR(In-Sync Replicas) 副本总数。

取值范围

不涉及。

consumers_count

Integer

参数解释

消费组数量。

取值范围

不涉及。

表4 brokers

参数

参数类型

描述

host

String

参数解释

节点IP。

取值范围

不涉及。

port

Integer

参数解释

端口号。

取值范围

不涉及。

broker_id

String

参数解释

节点ID。

取值范围

不涉及。

is_controller

Boolean

参数解释

是否为controller节点。

取值范围

  • true:是controller节点。

  • false:不是controller节点。

version

String

参数解释

服务端版本。

取值范围

  • 1.1.0

  • 2.7

  • 3.x

register_time

Long

参数解释

broker注册时间,为unix时间戳格式。

取值范围

不涉及。

is_health

Boolean

参数解释

Kafka实例节点的连通性是否正常。

取值范围

  • true:正常。

  • false:不正常。

请求示例

GET https://{endpoint}/v2/{project_id}/instances/{instance_id}/management/cluster

响应示例

状态码:200

查询Kafka集群元数据信息成功。

{
  "cluster" : {
    "controller" : "2",
    "brokers" : [ {
      "host" : "192.168.0.159",
      "port" : 9093,
      "broker_id" : "0",
      "is_controller" : false,
      "version" : "1.1.0",
      "register_time" : 1588754647872,
      "is_health" : true
    }, {
      "host" : "192.168.0.48",
      "port" : 9093,
      "broker_id" : "1",
      "is_controller" : false,
      "version" : "1.1.0",
      "register_time" : 1588754647653,
      "is_health" : true
    }, {
      "host" : "192.168.0.212",
      "port" : 9093,
      "broker_id" : "2",
      "is_controller" : true,
      "version" : "1.1.0",
      "register_time" : 1588754647284,
      "is_health" : true
    } ],
    "topics_count" : 3,
    "partitions_count" : 9,
    "online_partitions_count" : 9,
    "replicas_count" : 27,
    "isr_replicas_count" : 27,
    "consumers_count" : 0
  }
}

SDK代码示例

SDK代码示例如下。

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
package com.huaweicloud.sdk.test;

import com.huaweicloud.sdk.core.auth.ICredential;
import com.huaweicloud.sdk.core.auth.BasicCredentials;
import com.huaweicloud.sdk.core.exception.ConnectionException;
import com.huaweicloud.sdk.core.exception.RequestTimeoutException;
import com.huaweicloud.sdk.core.exception.ServiceResponseException;
import com.huaweicloud.sdk.kafka.v2.region.KafkaRegion;
import com.huaweicloud.sdk.kafka.v2.*;
import com.huaweicloud.sdk.kafka.v2.model.*;


public class ShowClusterSolution {

    public static void main(String[] args) {
        // The AK and SK used for authentication are hard-coded or stored in plaintext, which has great security risks. It is recommended that the AK and SK be stored in ciphertext in configuration files or environment variables and decrypted during use to ensure security.
        // In this example, AK and SK are stored in environment variables for authentication. Before running this example, set environment variables CLOUD_SDK_AK and CLOUD_SDK_SK in the local environment
        String ak = System.getenv("CLOUD_SDK_AK");
        String sk = System.getenv("CLOUD_SDK_SK");
        String projectId = "{project_id}";

        ICredential auth = new BasicCredentials()
                .withProjectId(projectId)
                .withAk(ak)
                .withSk(sk);

        KafkaClient client = KafkaClient.newBuilder()
                .withCredential(auth)
                .withRegion(KafkaRegion.valueOf("<YOUR REGION>"))
                .build();
        ShowClusterRequest request = new ShowClusterRequest();
        request.withInstanceId("{instance_id}");
        try {
            ShowClusterResponse response = client.showCluster(request);
            System.out.println(response.toString());
        } catch (ConnectionException e) {
            e.printStackTrace();
        } catch (RequestTimeoutException e) {
            e.printStackTrace();
        } catch (ServiceResponseException e) {
            e.printStackTrace();
            System.out.println(e.getHttpStatusCode());
            System.out.println(e.getRequestId());
            System.out.println(e.getErrorCode());
            System.out.println(e.getErrorMsg());
        }
    }
}
 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
# coding: utf-8

import os
from huaweicloudsdkcore.auth.credentials import BasicCredentials
from huaweicloudsdkkafka.v2.region.kafka_region import KafkaRegion
from huaweicloudsdkcore.exceptions import exceptions
from huaweicloudsdkkafka.v2 import *

if __name__ == "__main__":
    # The AK and SK used for authentication are hard-coded or stored in plaintext, which has great security risks. It is recommended that the AK and SK be stored in ciphertext in configuration files or environment variables and decrypted during use to ensure security.
    # In this example, AK and SK are stored in environment variables for authentication. Before running this example, set environment variables CLOUD_SDK_AK and CLOUD_SDK_SK in the local environment
    ak = os.environ["CLOUD_SDK_AK"]
    sk = os.environ["CLOUD_SDK_SK"]
    projectId = "{project_id}"

    credentials = BasicCredentials(ak, sk, projectId)

    client = KafkaClient.new_builder() \
        .with_credentials(credentials) \
        .with_region(KafkaRegion.value_of("<YOUR REGION>")) \
        .build()

    try:
        request = ShowClusterRequest()
        request.instance_id = "{instance_id}"
        response = client.show_cluster(request)
        print(response)
    except exceptions.ClientRequestException as e:
        print(e.status_code)
        print(e.request_id)
        print(e.error_code)
        print(e.error_msg)
 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
package main

import (
	"fmt"
	"github.com/huaweicloud/huaweicloud-sdk-go-v3/core/auth/basic"
    kafka "github.com/huaweicloud/huaweicloud-sdk-go-v3/services/kafka/v2"
	"github.com/huaweicloud/huaweicloud-sdk-go-v3/services/kafka/v2/model"
    region "github.com/huaweicloud/huaweicloud-sdk-go-v3/services/kafka/v2/region"
)

func main() {
    // The AK and SK used for authentication are hard-coded or stored in plaintext, which has great security risks. It is recommended that the AK and SK be stored in ciphertext in configuration files or environment variables and decrypted during use to ensure security.
    // In this example, AK and SK are stored in environment variables for authentication. Before running this example, set environment variables CLOUD_SDK_AK and CLOUD_SDK_SK in the local environment
    ak := os.Getenv("CLOUD_SDK_AK")
    sk := os.Getenv("CLOUD_SDK_SK")
    projectId := "{project_id}"

    auth := basic.NewCredentialsBuilder().
        WithAk(ak).
        WithSk(sk).
        WithProjectId(projectId).
        Build()

    client := kafka.NewKafkaClient(
        kafka.KafkaClientBuilder().
            WithRegion(region.ValueOf("<YOUR REGION>")).
            WithCredential(auth).
            Build())

    request := &model.ShowClusterRequest{}
	request.InstanceId = "{instance_id}"
	response, err := client.ShowCluster(request)
	if err == nil {
        fmt.Printf("%+v\n", response)
    } else {
        fmt.Println(err)
    }
}

更多编程语言的SDK代码示例,请参见API Explorer的代码示例页签,可生成自动对应的SDK代码示例。

状态码

状态码

描述

200

查询Kafka集群元数据信息成功。

错误码

请参见错误码

相关文档