Creating a Custom Topology Cluster
The analysis cluster, streaming cluster, and hybrid cluster provided by MRS use fixed templates to deploy cluster processes. Therefore, you cannot customize service processes on management nodes and control nodes. If you want to customize the cluster deployment, set Cluster Type to Custom when creating a cluster. In this way, you can customize the deployment mode of process instances on the management nodes and control nodes in the cluster. Only MRS 3.x and later versions support the creation of clusters in a custom topology.
A custom cluster provides the following functions:
- Separated deployment of the management and control roles: The management role and control role are deployed on different Master nodes.
- Co-deployment of the management and control roles: The management and control roles are co-deployed on the Master node.
- ZooKeeper is deployed on an independent node to improve reliability.
- Components are deployed separately to avoid resource contention.
Roles in an MRS cluster:
- Management Node (MN): is the node to install Manager (the management system of the MRS cluster). It provides a unified access entry. Manager centrally manages nodes and services deployed in the cluster.
- Control Node (CN): controls and monitors how data nodes store and receive data, and send process status, and provides other public functions. Control nodes of MRS include HMaster, HiveServer, ResourceManager, NameNode, JournalNode, and SlapdServer.
- Data Node (DN): A data node executes the instructions sent by the management node, reports task status, stores data, and provides other public functions. Data nodes of MRS include DataNode, RegionServer, and NodeManager.
Customizing a Cluster
- Log in to the MRS console.
- Click Create Cluster. The page for creating a cluster is displayed.
- Click the Custom Config tab.
- Configure basic cluster information. For details about the parameters, see Software Configurations.
- Region: Retain the default value.
- Cluster Name: You can use the default name. However, you are advised to include a project name abbreviation or date for consolidated memory and easy distinguishing, for example, mrs_20180321.
- Cluster Version: Currently, only MRS 3.x are supported.
- Cluster Type: Select Custom and select components as required.
- Click Next. Configure hardware information.
- AZ: Retain the default value.
- VPC: Retain the default value. If there is no available VPC, click View VPC to access the VPC console and create a new VPC.
- Subnet: Retain the default value.
- Security Group: Select Auto create.
- EIP: Select Bind later.
- Common Node: For details, see Custom Cluster Template Description.
- Instance Specifications: Click to configure the instance specifications, system disk and data disk storage types, and storage space.
- Instance Count: Adjust the number of cluster instances based on the service volume. For details, see Table 2.
- Topology Adjustment: If the deployment mode in the Common Node does not meet the requirements, you need to manually install some instances that are not deployed by default, or you need to manually install some instances, set Topology Adjustment to Enable and adjust the instance deployment mode based on service requirements. For details, see Topology Adjustment for a Custom Cluster.
- Click Next and set advanced options.
For details about the parameters, see (Optional) Advanced Configuration.
- Click Create Now.
If Kerberos authentication is enabled for a cluster, check whether Kerberos authentication is required. If yes, click Continue. If no, click Back to disable Kerberos authentication and then create a cluster.
- Click Back to Cluster List to view the cluster status.
It takes some time to create a cluster. The initial status of the cluster is Starting. After the cluster has been created successfully, the cluster status becomes Running.
Custom Cluster Template Description
Common Node |
Description |
Node Range |
---|---|---|
Compact |
The management role and control role are deployed on the Master node, and data instances are deployed in the same node group. This deployment mode applies to scenarios where the number of control nodes is less than 100, reducing costs. |
|
OMS-separate |
The management role and control role are deployed on different Master nodes, and data instances are deployed in the same node group. This deployment mode is applicable to a cluster with 100 to 500 nodes and delivers better performance in high-concurrency load scenarios. |
|
Full-size |
The management role and control role are deployed on different Master nodes, and data instances are deployed in different node groups. This deployment mode is applicable to a cluster with more than 500 nodes. Components can be deployed separately, which can be used for a larger cluster scale. |
|
Node Deployment Principle |
Applicable Scenario |
Networking Rule |
|
---|---|---|---|
Management nodes, control nodes, and data nodes are deployed separately. (This scheme requires at least eight nodes.) |
MN × 2 + CN × 9 + DN × n |
(Recommended) This scheme is used when the number of data nodes is 500–2000. |
|
MN × 2 + CN × 5 + DN × n |
(Recommended) This scheme is used when the number of data nodes is 100–500. |
||
MN × 2 + CN × 3 + DN × n |
(Recommended) This scheme is used when the number of data nodes is 30–100. |
||
The management nodes and control nodes are deployed together, and the data nodes are deployed separately. |
(MN+CN) × 3 + DN × n |
(Recommended) This scheme is used when the number of data nodes is 3–30. |
Nodes in the cluster are deployed in the same subnet and are interconnected with each other at Layer 2 through aggregation switches. |
The management nodes, control nodes, and data nodes are deployed together. |
NOTE:
This template is not recommended in the production environment or commercial environment.
|
Nodes in the cluster are deployed in the same subnet and are interconnected with each other at Layer 2 through aggregation switches. |
Topology Adjustment for a Custom Cluster
Service |
Dependency |
Role |
Role Deployment Suggestions |
Description |
---|---|---|---|---|
OMSServer |
- |
OMSServer |
This role can be deployed it on the Master node and cannot be modified. |
- |
ClickHouse |
Depends on ZooKeeper. |
CHS (ClickHouseServer) |
This role can be deployed on all nodes. Number of role instances to be deployed: an even number ranging from 2 to 256 |
A non-Master node group with this role assigned is considered as a Core node. |
CLB (ClickHouseBalancer) |
This role can be deployed on all nodes. Number of role instances to be deployed: 2 to 256 |
- |
||
ZooKeeper |
- |
QP(quorumpeer) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 3 to 9, with the step size of 2 |
- |
Hadoop |
Depends on ZooKeeper. |
NN(NameNode) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 2 |
The NameNode and ZKFC processes are deployed on the same server for cluster HA. |
HFS (HttpFS) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 0 to 10 |
- |
||
JN(JournalNode) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 3 to 60, with the step size of 2 |
- |
||
DN(DataNode) |
This role can be deployed on all nodes. Number of role instances to be deployed: 3 to 10,000 |
A non-Master node group with this role assigned is considered as a Core node. |
||
RM(ResourceManager) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 2 |
- |
||
NM(NodeManager) |
This role can be deployed on all nodes. Number of role instances to be deployed: 3 to 10,000 |
- |
||
JHS(JobHistoryServer) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 1 to 2 |
- |
||
TLS(TimelineServer) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 0 to 1 |
- |
||
Presto |
Depends on Hive. |
PCD(Coordinator) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 2 |
- |
PWK(Worker) |
This role can be deployed on all nodes. Number of role instances to be deployed: 1 to 10,000 |
- |
||
Spark2x |
|
JS2X(JDBCServer2x) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 2 to 10 |
- |
JH2X(JobHistory2x) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 2 |
- |
||
SR2X(SparkResource2x) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 2 to 50 |
- |
||
IS2X(IndexServer2x) |
(Optional) This role can be deployed on the Master node only. Number of role instances to be deployed: 0 to 2, with the step size of 2 |
- |
||
HBase |
Depends on Hadoop. |
HM(HMaster) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 2 |
- |
TS(ThriftServer) |
This role can be deployed on all nodes. Number of role instances to be deployed: 0 to 10,000 |
- |
||
RT(RESTServer) |
This role can be deployed on all nodes. Number of role instances to be deployed: 0 to 10,000 |
- |
||
RS(RegionServer) |
This role can be deployed on all nodes. Number of role instances to be deployed: 3 to 10,000 |
- |
||
TS1(Thrift1Server) |
This role can be deployed on all nodes. Number of role instances to be deployed: 0 to 10,000 |
If the Hue service is installed in a cluster and HBase needs to be used on the Hue web UI, install this instance for the HBase service. |
||
Hive |
|
MS(MetaStore) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 2 to 10 |
- |
WH (WebHCat) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 1 to 10 |
- |
||
HS(HiveServer) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 2 to 80 |
- |
||
Hue |
Depends on DBService. |
H(Hue) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 2 |
- |
Sqoop |
Depends on Hadoop. |
SC(SqoopClient) |
This role can be deployed on all nodes. Number of role instances to be deployed: 1 to 10,000 |
- |
Kafka |
Depends on ZooKeeper. |
B(Broker) |
This role can be deployed on all nodes. Number of role instances to be deployed: 3 to 10,000 |
- |
Flume |
- |
MS(MonitorServer) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 1 to 2 |
- |
F(Flume) |
This role can be deployed on all nodes. Number of role instances to be deployed: 1 to 10,000 |
A non-Master node group with this role assigned is considered as a Core node. |
||
Tez |
|
TUI(TezUI) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 1 to 2 |
- |
Flink |
|
FR(FlinkResource) |
This role can be deployed on all nodes. Number of role instances to be deployed: 1 to 10,000 |
- |
FS(FlinkServer) |
This role can be deployed on all nodes. Number of role instances to be deployed: 0 to 2 |
- |
||
Oozie |
|
O(oozie) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 2 |
- |
Impala |
|
StateStore |
This role can be deployed on the Master node only. Number of role instances to be deployed: 1 |
- |
Catalog |
This role can be deployed on the Master node only. Number of role instances to be deployed: 1 |
- |
||
Impalad |
This role can be deployed on all nodes. Number of role instances to be deployed: 1 to 10,000 |
- |
||
Kudu |
- |
KuduMaster |
This role can be deployed on the Master node only. Number of role instances to be deployed: 3 or 5 |
- |
KuduTserver |
This role can be deployed on all nodes. Number of role instances to be deployed: 3 to 10,000 |
- |
||
Ranger |
Depends on DBService. |
RA(RangerAdmin) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 1 to 2 |
- |
USC(UserSync) |
This role can be deployed on the Master node only. Number of role instances to be deployed: 1 |
- |
||
TSC (TagSync) |
This role can be deployed on all nodes. Number of role instances to be deployed: 0 to 1 |
- |
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.See the reply and handling status in My Cloud VOC.
For any further questions, feel free to contact us through the chatbot.
Chatbot