Practices
You can better use DLI for big data analytics and processing by following the scenario-specific instructions and best practices provided in this section.
Scenario |
Instructions |
Description |
---|---|---|
Connecting a queue to an external data source |
Configuring the Connection Between a DLI Queue and a Data Source in a Private Network |
When creating and running a job on a DLI queue, you need to connect the DLI queue to external data sources. This section describes how to connect DLI queues to external data sources. For example, to connect a DLI queue to MRS, RDS, CSS, Kafka, or GaussDB(DWS), you need to configure the connection between the queue and the external data source. |
Configuring the Connection Between a DLI Queue and a Data Source in the Internet |
Connect a DLI queue to a data source on the Internet. You can configure SNAT rules and add routes to the public network to enable communications between a queue and the Internet. |
|
Spark SQL job development |
Use a Spark SQL job to create OBS tables, and import, insert, and query OBS table data. |
|
Flink OpenSource SQL job development |
Use a Flink OpenSource SQL job to read data from Kafka and write the data to RDS. |
|
Use a Flink OpenSource SQL job to read data from Kafka and write the data to GaussDB(DWS). |
||
Use a Flink OpenSource SQL job to read data from Kafka and write the data to Elasticsearch. |
||
Reading Data from MySQL CDC and Writing Data to GaussDB(DWS) |
Use a Flink OpenSource SQL job to read data from MySQL CDC and write the data to GaussDB(DWS). |
|
Reading Data from PostgreSQL CDC and Writing Data to GaussDB(DWS) |
Use a Flink OpenSource SQL job to read data from PostgreSQL CDC and write the data to GaussDB(DWS). |
|
Flink Jar job development |
Create a custom Flink Jar job to interact with MRS. |
|
Write Kafka data to OBS. |
||
Using Flink Jar to Connect to Kafka with SASL_SSL Authentication Enabled |
Use Flink OpenSource SQL to connect to Kafka with SASL_SSL authentication enabled. |
|
Spark Jar job development |
Write a Spark program to read and query OBS data, compile and package your code, and submit a Spark Jar job. |
|
Data migration |
Migrate data from MRS Hive to DLI using the CDM data synchronization function. |
|
Migrate data from MRS Kafka to DLI using the CDM data synchronization function. |
||
Migrate data from a CSS Elasticsearch cluster to DLI using the CDM data synchronization function. |
||
Migrate data from an RDS database to DLI using the CDM data synchronization function. |
||
Migrate data from GaussDB(DWS) to DLI using the CDM data synchronization function. |
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.