Interaction with HBase Using User-Defined Jobs
Overview
You can perform secondary development based on Flink and Spark APIs to build your own JAR packages and submit them to the CS cluster to implement interactions with CS and MRS HBase clusters.
Apache HBase is a column-oriented distributed cloud storage system that features enhanced reliability, excellent performance, and elastic scalability. It applies to the storage of massive amounts of data and distributed computing. You can use HBase to build a storage system capable of storing TB- or even PB-level data. With HBase, you can filter and analyze data with ease and get responses in milliseconds, rapidly mining data value. HBase clusters are deployed and hosted on MRS that is powered on Apache HBase.
Prerequisites
- To use Kafka in an MRS cluster, you need to use the VPC peering connection to interconnect CS with the MRS cluster.
For details about how to set up the VPC peering connection, see VPC Peering Connection in the Cloud Stream Service User Guide.
- If the Kafka server listens on the port using hostname, you need to add the mapping between the hostname and IP address of the Kafka Broker node to the CS cluster. Contact the Kafka service deployment personnel to obtain the hostname and IP address of the Kafka Broker node. For details about how to add an IP-domain mapping, see the description of Adding an IP-Domain Mapping in Cluster Management in the Cloud Stream Service User Guide.
Procedure
Create and submit a user-defined Flink job. For details, see Creating a User-Defined Flink Job in the Cloud Stream Service User Guide.
Create and submit a user-defined Spark job. For details, see Creating a User-Defined Spark Job in the Cloud Stream Service User Guide.
Last Article: Interaction with Kafka Using User-Defined Jobs
Next Article: Cloud Ecosystem: APIG
Did this article solve your problem?
Thank you for your score!Your feedback would help us improve the website.