Help Center> Cloud Stream Service> Eco Development> Cloud Ecosystem: MRS> Interaction with HBase Using User-Defined Jobs

Interaction with HBase Using User-Defined Jobs

Overview

You can perform secondary development based on Flink and Spark APIs to build your own JAR packages and submit them to the CS cluster to implement interactions with CS and MRS HBase clusters.

Apache HBase is a column-oriented distributed cloud storage system that features enhanced reliability, excellent performance, and elastic scalability. It applies to the storage of massive amounts of data and distributed computing. You can use HBase to build a storage system capable of storing TB- or even PB-level data. With HBase, you can filter and analyze data with ease and get responses in milliseconds, rapidly mining data value. HBase clusters are deployed and hosted on MRS that is powered on Apache HBase.

Prerequisites

  • To use Kafka in an MRS cluster, you need to use the VPC peering connection to interconnect CS with the MRS cluster.

    For details about how to set up the VPC peering connection, see VPC Peering Connection in the Cloud Stream Service User Guide.

  • If the Kafka server listens on the port using hostname, you need to add the mapping between the hostname and IP address of the Kafka Broker node to the CS cluster. Contact the Kafka service deployment personnel to obtain the hostname and IP address of the Kafka Broker node. For details about how to add an IP-domain mapping, see the description of Adding an IP-Domain Mapping in Cluster Management in the Cloud Stream Service User Guide.

Procedure

Create and submit a user-defined Flink job. For details, see Creating a User-Defined Flink Job in the Cloud Stream Service User Guide.

Create and submit a user-defined Spark job. For details, see Creating a User-Defined Spark Job in the Cloud Stream Service User Guide.