Help Center> Data Lake Insight> FAQs> Problems Related to Spark Jobs> Job O&M Errors> Why Is a Job Running Timeout Reported When a Spark Job Runs a Large Amount of Data?
Updated on 2023-05-19 GMT+08:00

Why Is a Job Running Timeout Reported When a Spark Job Runs a Large Amount of Data?

When a Spark job accesses a large amount of data, for example, accessing data in a GaussDB(DWS) database, you are advised to set the number of concurrent tasks and enable multi-task processing.

For details about how to set the concurrency, see the partitionColumn and numPartitions fields and Scala Example Code for connecting to GaussDB(DWS).

Job O&M Errors FAQs

more