Help Center/ MapReduce Service/ User Guide/ MRS Cluster O&M/ MRS Cluster Alarm Handling Reference/ ALM-16053 Average HQL Submission Time of Hive in the Last 5 Minutes Exceeds the Threshold
Updated on 2024-11-13 GMT+08:00

ALM-16053 Average HQL Submission Time of Hive in the Last 5 Minutes Exceeds the Threshold

Alarm Description

The system periodically checks the average HQL submission time, which is the time for calling the MapReduce/Spark/Tez APIs to submit Yarn jobs, including the time for uploading dependent temporary JAR packages and splitting files. This alarm is generated when the average HQL submission time exceeds the threshold.

This alarm is cleared when the HQL submission time falls below the threshold.

This alarm applies to MRS 3.5.0 or later.

Alarm Attributes

Alarm ID

Alarm Severity

Auto Cleared

16053

Critical (default threshold: 240 seconds)

Major (default threshold: 120 seconds)

Yes

Alarm Parameters

Type

Parameter

Description

Location Information

Source

Specifies the cluster for which the alarm was generated.

ServiceName

Specifies the service for which the alarm was generated.

RoleName

Specifies the role for which the alarm was generated.

HostName

Specifies the host for which the alarm was generated.

Additional Information

Trigger Condition

Specifies the alarm triggering condition.

Impact on the System

If this alarm is generated, the average HQL submission time in the last 5 minutes exceeds the threshold. As a result, the HQL running time is prolonged. Errors may occur in Hive On Spark jobs.

Possible Causes

The HiveServer GC time is too long or the HDFS NameNode/Router RPC latency is too long.

Handling Procedure

Check whether the GC time of HiveServer is too long.

  1. Log in to FusionInsight Manager, choose O&M > Alarm > Alarms, and check whether alarm Heap Memory Usage of the Hive Process Exceeds the Threshold exists in the alarm list.

    • If yes, go to 2.
    • If no, go to 4.

  2. Rectify the fault by following the handling procedure of ALM-16005 Heap Memory Usage of the Hive Process Exceeds the Threshold.
  3. Check whether the alarm is cleared in the alarm list.

    • If yes, no further action is required.
    • If no, go to 4.

Check whether the HDFS RPC latency is too long.

  1. Check whether alarm Average NameNode RPC Processing Time Exceeds the Threshold exists in the alarm list.

    • If yes, go to 5.
    • If no, go to 7.

  2. Rectify the fault by following the handling procedure of ALM-14021 Average NameNode RPC Processing Time Exceeds the Threshold.
  3. Check whether the alarm is cleared in the alarm list.

    • If yes, no further action is required.
    • If no, go to 7.

Collect fault information.

  1. On FusionInsight Manager, choose O&M. In the navigation pane on the left, choose Log > Download.
  2. Expand the Service drop-down list, and select Hive for the target cluster.
  3. Click the edit icon in the upper right corner, and set Start Date and End Date for log collection to 10 minutes ahead of and after the alarm generation time, respectively. Then, click Download.
  4. On FusionInsight Manager, choose Cluster > Services > Hive. On the displayed Dashboard page, click More and select Collect Stack Information. On the displayed page, set the following parameters:

    • Select HiveServer for the role where you want to collect data.
    • Select jstack and Enable continuous collection of jstack and jmap -histo information.
    • Set the collection interval to 10 seconds and the duration to 2 minutes.

  5. Click OK. After the collection is complete, click Download.
  6. Contact O&M engineers and provide the collected logs and stack information.

Alarm Clearance

This alarm is automatically cleared after the fault is rectified.

Related Information

None.