Common Configuration Items of Batch SQL Jobs
This section describes the common configuration items of the SQL syntax for DLI batch jobs.
| Item | Default Value | Description |
|---|---|---|
| spark.sql.files.maxRecordsPerFile | 0 | Maximum number of records to be written into a single file. If the value is zero or negative, there is no limit. |
| spark.sql.autoBroadcastJoinThreshold | 209715200 | Maximum size of the table that displays all working nodes when a connection is executed. You can set this parameter to -1 to disable the display. NOTE: Currently, only the configuration unit metastore table that runs the ANALYZE TABLE COMPUTE statistics noscan command and the file-based data source table that directly calculates statistics based on data files are supported. |
| spark.sql.shuffle.partitions | 4096 | Default number of partitions used to filter data for join or aggregation. |
| spark.sql.dynamicPartitionOverwrite.enabled | false | In dynamic mode, Spark does not delete the previous partitions and only overwrites the partitions without data during execution. |
| spark.sql.files.maxPartitionBytes | 134217728 | Maximum number of bytes to be packed into a single partition when a file is read. |
| spark.sql.badRecordsPath | - | Path of bad records. |
Last Article: SQL Syntax of Batch Jobs
Next Article: SQL Syntax Overview of Batch Jobs
Did this article solve your problem?
Thank you for your score!Your feedback would help us improve the website.