Common Configuration Items of Batch SQL Jobs
This section describes the common configuration items of the SQL syntax for DLI batch jobs.
|
Item |
Default Value |
Description |
|---|---|---|
|
spark.sql.files.maxRecordsPerFile |
0 |
Maximum number of records to be written into a single file. If the value is zero or negative, there is no limit. |
|
spark.sql.shuffle.partitions |
200 |
Default number of partitions used to filter data for join or aggregation. |
|
spark.sql.dynamicPartitionOverwrite.enabled |
false |
Whether DLI overwrites the partitions where data will be written into during runtime. If you set this parameter to false, all partitions that meet the specified condition will be deleted before data overwrite starts. For example, if you set false and use INSERT OVERWRITE to write partition 2021-02 to a partitioned table that has the 2021-01 partition, this partition will be deleted. If you set this parameter to true, DLI does not delete partitions before overwrite starts. |
|
spark.sql.files.maxPartitionBytes |
134217728 |
Maximum number of bytes to be packed into a single partition when a file is read. |
|
spark.sql.badRecordsPath |
- |
Path of bad records. |
|
dli.sql.sqlasync.enabled |
true |
Whether DDL and DCL statements are executed asynchronously. The value true indicates that asynchronous execution is enabled. |
|
dli.sql.job.timeout |
- |
Job running timeout interval, in seconds. If the job times out, it will be canceled. |
|
spark.sql.keep.distinct.expandThreshold |
- |
|
|
spark.sql.distinct.aggregator.enabled |
false |
|
|
spark.sql.dli.job.shareLevel |
Queue |
This configuration item is used to set the isolation level of SQL statements. Different isolation levels (job, user, project, queue) determine whether SQL jobs are executed by independent Spark Drivers and Executors or share existing ones.
NOTE:
The maximum number of Spark Drivers that can be started (maximum Spark Driver instances) and the maximum number of concurrent SQL queries that can be executed by each Spark Driver (maximum concurrency per Spark Driver instance) can be configured in the queue properties. |
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.See the reply and handling status in My Cloud VOC.
For any further questions, feel free to contact us through the chatbot.
Chatbot