Help Center> Data Lake Insight> FAQs> Problems Related to Spark Jobs> Job O&M Errors> Why Is a Job Running Timeout Reported When a Spark Job Runs a Large Amount of Data?
Updated on 2023-05-19 GMT+08:00

Why Is a Job Running Timeout Reported When a Spark Job Runs a Large Amount of Data?

When a Spark job accesses a large amount of data, for example, accessing data in a GaussDB(DWS) database, you are advised to set the number of concurrent tasks and enable multi-task processing.