Developing a Real-Time Processing Single-Task DLI Spark Job
Prerequisites
A single-task real-time processing DLI Spark job has been created. For details, see Creating a Job.
Configuring a DLI Spark job
Parameter |
Mandatory |
Description |
---|---|---|
Job Name |
Yes |
Enter the DLI Spark job name. The job name can contain 1 to 64 characters. Only letters, digits, hyphens (-), and underscores (_) are allowed. |
DLI Queue |
Yes |
Select a DLI queue. |
Spark Version |
No |
|
Job Type |
No |
Type of the Spark image used by the job. The following options are available:
|
Job Running Resource |
No |
|
Major Job Class |
No |
Java/Scala main class of the job |
Spark program resource package |
Yes |
Resource package on which the Spark program depends |
Resource Type |
Yes |
DLI program package: The resource package file will not be uploaded to the DLI resource management system before the job is executed. OBS path: The resource package file will not be uploaded to DLI resource management system before the job is executed. The OBS path where the file is located is part of the message body for starting the job. This type is recommended. |
Group |
No |
This parameter is required when Resource Type is set to DLI program package. A Spark program resource package is uploaded to a specified group. The main JAR package and dependency package are uploaded to the same group.
|
Major-Class Entry Parameters |
No |
Press Enter to separate parameters. |
Spark program resource package |
No |
Enter parameters in key=value format and separate parameters by pressing Enter. |
Module Name |
No |
Select one or more module names. |
Metadata Access |
No |
Whether metadata can be accessed To access the OBS table created by the DLI SQL job in the DLI Spark job, enable metadata access. |
Parameter |
Mandatory |
Description |
---|---|---|
Job Status Polling Interval (s) |
Yes |
Set the interval at which the system checks whether the job is complete. The interval can range from 30s to 60s, or 120s, 180s, 240s, or 300s. During job execution, the system checks the job status at the configured interval. |
Maximum Wait Time |
Yes |
Set the timeout interval for the job. If the job is not complete within the timeout interval and retry is enabled, the job will be executed again.
NOTE:
If the job is in starting state and fails to start, it will fail upon timeout. |
Retry upon Failure |
No |
Whether to re-execute a node if it fails to be executed.
|
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.See the reply and handling status in My Cloud VOC.
For any further questions, feel free to contact us through the chatbot.
Chatbot