ALM-38005 GC Duration of the Broker Process Exceeds the Threshold
Description
The system checks the garbage collection (GC) duration of the Broker process every 60 seconds. This alarm is generated when the GC duration exceeds the threshold (12 seconds by default) for 3 consecutive times.
When the Trigger Count is 1, this alarm is cleared when the GC duration is less than or equal to the threshold. When the Trigger Count is greater than 1, this alarm is cleared when the GC duration is less than or equal to 90% of the threshold.
Attribute
Alarm ID |
Alarm Severity |
Automatically Cleared |
---|---|---|
38005 |
Major |
Yes |
Parameters
Name |
Meaning |
---|---|
Source |
Specifies the cluster for which the alarm is generated. |
ServiceName |
Specifies the service for which the alarm is generated. |
RoleName |
Specifies the role for which the alarm is generated. |
HostName |
Specifies the host for which the alarm is generated. |
Trigger Condition |
Specifies the threshold triggering the alarm. If the current indicator value exceeds this threshold, the alarm is generated. |
Impact on the System
A long GC duration of the Broker process may interrupt the services.
Possible Causes
The Kafka GC duration of the node is too long or the heap memory is inappropriately allocated. As a result, GCs occur frequently.
Procedure
Check the GC duration.
- On the FusionInsight Manager portal, choose O&M > Alarm > Alarms > GC Duration of the Broker Process Exceeds the Threshold > Location. Check the host name of the instance involved in this alarm.
- On the FusionInsight Manager portal, choose Cluster > Name of the desired cluster > Services > Kafka > Instance. Click the instance for which the alarm is generated to go to the page for the instance. Click the drop-down list in the upper right corner of the chart area, choose Customize > Process > Broker GC Duration per Minute, and click OK.
Figure 1 Broker GC Duration per Minute
- Check whether the GC duration of the Broker process collected every minute exceeds the threshold (12 seconds by default).
Check the direct memory size configured for the Kafka.
- On the FusionInsight Manager portal, choose Cluster > Name of the desired cluster > Services > Kafka > Configurations > All Configurations > Broker(Role) > Environment to increase the value of -Xmx configured in the KAFKA_HEAP_OPTS parameter by referring to the Note.
- It is recommended that -Xmx and -Xms be set to the same value.
- You are advised to set the value of KAFKA_HEAP_OPTS to twice the value of Direct Memory Used by Kafka.
On the FusionInsight Manager portal, choose Cluster > Name of the desired cluster > Services > Kafka > Instance. Click the instance for which the alarm is generated to go to the page for the instance. Click the drop-down list in the upper right corner of the chart area and choose Customize > Process > Kafka Direct Memory Resource Status to check the value of Direct Memory Used by Kafka.
- Save the configuration and restart the Kafka service.
- Check whether the alarm is cleared.
- If yes, no further action is required.
- If no, go to 7.
Collect fault information.
- On the FusionInsight Manager portal, choose O&M > Log > Download.
- Select Kafka in the required cluster from the Service drop-down list.
- Click in the upper right corner, and set Start Date and End Date for log collection to 10 minutes ahead of and after the alarm generation time, respectively. Then, click Download.
- Contact the O&M personnel and send the collected logs.
Alarm Clearing
After the fault is rectified, the system automatically clears this alarm.
Related Information
None
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.