Connecting Elasticsearch Exporter
Application Scenario
When using Elasticsearch, you need to monitor Elasticsearch running, such as the cluster and index status. The Prometheus monitoring function monitors Elasticsearch running using Exporter in the CCE container scenario. This section describes how to deploy Elasticsearch Exporter and implement alarm access.
You are advised to use CCE for unified Exporter management.
Prerequisites
- A CCE cluster has been created and Elasticsearch has been installed.
- Your service has been connected for Prometheus monitoring and a CCE cluster has also been connected. For details, see Prometheus Instance for CCE.
- You have uploaded the elasticsearch_exporter image to SoftWare Repository for Container (SWR). For details, see Uploading an Image Through a Container Engine Client.
Deploying Elasticsearch Exporter
- Log in to the CCE console.
- Click the connected cluster. The cluster management page is displayed.
- Perform the following operations to deploy Exporter:
- Configure a secret.
In the navigation pane, choose ConfigMaps and Secrets. Then click Create from YAML in the upper right corner of the page. The following shows a YAML configuration example:
apiVersion: v1 kind: Secret metadata: name: es-secret-test namespace: default type: Opaque stringData: esURI: http://124.70.14.51:30920 # URI of Elasticsearch. Use the IP address of the cluster or any node in the cluster.
- Format of the Elasticsearch connection string: <proto>://<user>:<password>@<host>:<port>, for example, http://admin:pass@localhost:9200. You can also leave the password blank, for example, http://10.247.43.50:9200.
- The password has been encrypted based on Opaque requirements.
- For details about how to configure a secret, see Creating a Secret.
- Deploy Elasticsearch Exporter.
In the navigation pane, choose Workloads. In the upper right corner, click Create Workload. Then select the Deployment workload and a desired namespace to deploy Elasticsearch Exporter. YAML configuration example for deploying Exporter:
apiVersion: apps/v1 kind: Deployment metadata: labels: k8s-app: es-exporter # Change the value based on service requirements. name: es-exporter # Change the value based on service requirements. namespace: default #Select a proper namespace to deploy Exporter. If no namespace is available, create one. spec: replicas: 1 selector: matchLabels: k8s-app: es-exporter # Change the value based on service requirements. template: metadata: labels: k8s-app: es-exporter # Change the value based on service requirements. spec: containers: - env: - name: ES_URI valueFrom: secretKeyRef: name: es-secret-test # Secret name specified in the previous step. key: esURI # Secret key specified in the previous step. - name: ES_ALL value: "true" image: swr.cn-north-4.myhuaweicloud.com/mall-swarm-demo/es-exporter:1.1.0 imagePullPolicy: IfNotPresent name: es-exporter ports: - containerPort: 9114 name: metric-port securityContext: privileged: false terminationMessagePath: /dev/termination-log terminationMessagePolicy: File dnsPolicy: ClusterFirst imagePullSecrets: - name: default-secret restartPolicy: Always schedulerName: default-scheduler securityContext: {} terminationGracePeriodSeconds: 30 --- apiVersion: v1 kind: Service metadata: name: es-exporter name-space: default # Must be the same as the namespace where Exporter is deployed. spec: type: NodePort selector: k8s-app: es-exporter ports: - protocol: TCP nodePort: 30921 port: 9114 targetPort: 9114
In the preceding example, ES_ALL is used to collect all Elasticsearch monitoring items. You can change parameters if needed. For more details about Exporter parameters, see elasticsearch_exporter.
- Check whether Elasticsearch Exporter is successfully deployed.
- On the Deployments tab page, click the Deployment created in 3.b. In the pod list, choose More > View Logs in the Operation column. The Exporter is successfully started and its access address is exposed.
- Perform verification using one of the following methods:
- Log in to a cluster node and run either of the following commands:
curl http://{Cluster IP address}:9114/metrics
curl http://{Private IP address of any node in the cluster}:30921/metrics
- Access http://{Public IP address of any node in the cluster}:30921/metrics.
Figure 1 Accessing a cluster node
- In the instance list, choose More > Remote Login in the Operation column and run the following command:
curl http://localhost:9114/metric
- Log in to a cluster node and run either of the following commands:
- Configure a secret.
Collecting Service Data of the CCE Cluster
Add PodMonitor to configure a collection rule for monitoring the service data of applications deployed in the CCE cluster.
In the following example, metrics are collected every 30s. Therefore, you can check the reported metrics on the AOM page about 30s later.
apiVersion: monitoring.coreos.com/v1 kind: PodMonitor metadata: name: elasticSearch-exporter namespace: default spec: namespaceSelector: # Select the namespace where Exporter is deployed. matchNames: - default # Namespace where Exporter is located. podMetricsEndpoints: - interval: 30s # Set the metric collection period. path: /metrics # Enter the path corresponding to Prometheus Exporter. Default: /metrics. port: metric-port # Enter the name of "ports" in the YAML file corresponding to Prometheus Exporter. selector: # Enter the label of the target Exporter pod. matchLabels: k8s-app: elasticSearch-exporter
Verifying that Metrics Can Be Reported to AOM
- Log in to the AOM 2.0 console.
- In the navigation pane on the left, choose Prometheus Monitoring > Instances.
- Click the Prometheus instance connected to the CCE cluster. The instance details page is displayed.
- On the Metrics tab page of the Metric Management page, select your target cluster.
- Select job {namespace}/elasticsearch-exporter to query custom metrics starting with elasticsearch.
Setting a Dashboard and Alarm Rule on AOM
By setting a dashboard, you can monitor CCE cluster data on the same screen. By setting an alarm rule, you can detect cluster faults and implement warning in a timely manner.
- Setting a dashboard
- Log in to the AOM 2.0 console.
- In the navigation pane, choose Dashboard. On the displayed page, click Add Dashboard to add a dashboard. For details, see Creating a Dashboard.
- On the Dashboard page, select a Prometheus instance for CCE and click Add Graph. For details, see Adding a Graph to a Dashboard.
- Setting an alarm rule
- Log in to the AOM 2.0 console.
- In the navigation pane, choose Alarm Management > Alarm Rules.
- On the Metric/Event Alarm Rules tab page, click Create to create an alarm rule. For details, see Creating a Metric Alarm Rule.
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.See the reply and handling status in My Cloud VOC.
For any further questions, feel free to contact us through the chatbot.
Chatbot