Help Center/ Data Lake Insight/ FAQs/ Problems Related to Spark Jobs/ Job O&M Errors/ Why Is a Job Running Timeout Reported When a Spark Job Runs a Large Amount of Data?
Updated on 2023-05-19 GMT+08:00

Why Is a Job Running Timeout Reported When a Spark Job Runs a Large Amount of Data?

When a Spark job accesses a large amount of data, for example, accessing data in a GaussDB(DWS) database, you are advised to set the number of concurrent tasks and enable multi-task processing.