Help Center>
Data Lake Insight>
FAQs>
Flink Jobs>
O&M Guide>
How Do I Locate a Flink Job Running Error?
Updated on 2022-09-20 GMT+08:00
How Do I Locate a Flink Job Running Error?
- On the Flink job management, click Edit in the Operation column of the target job. On the displayed page, check whether Save Job Log in the Running Parameters tab is enabled.
Figure 1 Enabling Save Job Logs
- On the job running page, select Save Job Log and specify an OBS bucket for storing the logs. Click Start to run the job again. After the executed is complete, perform 3 and subsequent steps.
- In the Flink job list, click the job name. On the displayed job details page, click the Run Log tab.
- Click view OBS Bucket to obtain the complete run logs of the job.
Figure 2 Viewing run logs
- Download the latest jobmanager.log file, search for the keyword RUNNING to FAILED, and determine the failure cause based on the errors in the context.
- If the information in the jobmanager.log file is insufficient for locating the fault, find the corresponding taskmanager.log file in the run logs and search for the keyword RUNNING to FAILED to confirm the failure cause.
Parent topic: O&M Guide
O&M Guide FAQs
- How Do I Locate a Flink Job Submission Error?
- How Do I Locate a Flink Job Running Error?
- How Do I Know Whether a Flink Job Can Be Restored from a Checkpoint After Being Restarted?
- Why Does DIS Stream Not Exist During Job Semantic Check?
- Why Is the OBS Bucket Selected for Job Not Authorized?
- Why Are Logs Not Written to the OBS Bucket After a DLI Flink Job Fails to Be Submitted for Running?
- How Do I Configure Connection Retries for Kafka Sink If it is Disconnected?
- Why Is Information Displayed on the FlinkUI/Spark UI Page Incomplete?
- Why Is the Flink Job Abnormal Due to Heartbeat Timeout Between JobManager and TaskManager?
- Why Is Error "Timeout expired while fetching topic metadata" Repeatedly Reported in Flink JobManager Logs?
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.
The system is busy. Please try again later.
more