Practices
You can better use DLI for big data analytics and processing by following the scenario-specific instructions and best practices provided in this section.
Scenario |
Instructions |
Description |
---|---|---|
Connecting a queue to an external data source |
Configuring the Connection Between a DLI Queue and a Data Source in a Private Network |
When creating and running a job on a DLI queue, you need to connect the DLI queue to external data sources. This section describes how to connect DLI queues to external data sources. For example, to connect a DLI queue to MRS, RDS, CSS, Kafka, or GaussDB(DWS), you need to configure the connection between the queue and the external data source. |
Configuring the Connection Between a DLI Queue and a Data Source in the Internet |
Connect a DLI queue to a data source on the Internet. You can configure SNAT rules and add routes to the public network to enable communications between a queue and the Internet. |
|
Spark SQL job development |
Use a Spark SQL job to create OBS tables, and import, insert, and query OBS table data. |
|
Flink OpenSource SQL job development |
Use a Flink OpenSource SQL job to read data from Kafka and write the data to RDS. |
|
Use a Flink OpenSource SQL job to read data from Kafka and write the data to GaussDB(DWS). |
||
Use a Flink OpenSource SQL job to read data from Kafka and write the data to Elasticsearch. |
||
Reading Data from MySQL CDC and Writing Data to GaussDB(DWS) |
Use a Flink OpenSource SQL job to read data from MySQL CDC and write the data to GaussDB(DWS). |
|
Reading Data from PostgreSQL CDC and Writing Data to GaussDB(DWS) |
Use a Flink OpenSource SQL job to read data from PostgreSQL CDC and write the data to GaussDB(DWS). |
|
Flink Jar job development |
Create a custom Flink Jar job to interact with MRS. |
|
Write Kafka data to OBS. |
||
Using Flink Jar to Connect to Kafka with SASL_SSL Authentication Enabled |
Use Flink OpenSource SQL to connect to Kafka with SASL_SSL authentication enabled. |
|
Use a Flink Jar job to read and write data from and to DIS. |
||
Spark Jar job development |
Write a Spark program to read and query OBS data, compile and package your code, and submit a Spark Jar job. |
|
Data migration |
Migrate data from MRS Hive to DLI using the CDM data synchronization function. |
|
Migrate data from MRS Kafka to DLI using the CDM data synchronization function. |
||
Migrate data from a CSS Elasticsearch cluster to DLI using the CDM data synchronization function. |
||
Migrate data from an RDS database to DLI using the CDM data synchronization function. |
||
Migrate data from GaussDB(DWS) to DLI using the CDM data synchronization function. |
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.See the reply and handling status in My Cloud VOC.
For any further questions, feel free to contact us through the chatbot.
Chatbot