System Management Health Checks and Alerts

A health check corresponds to a Prometheus query against metrics aggregated to the Prometheus instance. Core platform components like Kubernetes and Istio collect service-related metrics by default, which enables the System Management Health service to implement generic service health checks without custom instrumentation. Health checks are intended to be coarse-grained and comprehensive, as opposed to fine-grained and exhaustive. Health checks related to infrastructure adhere to the Utilization Saturation Errors (USE) method whereas services follow the Rate Errors Duration (RED) method.

Prometheus alerting rules periodically evaluate health checks and trigger alerts to Alertmanager, which manages silencing, inhibition, aggregation, and sending out notifications. Alertmanager supports a number of notification options, but the most relevant ones are listed below:

  • Email - Sends notification emails periodically regarding alerts
  • Slack - Publishes notifications to a Slack channel
  • Web hook- Send an HTTP request to a configurable URL (requires custom integration)

Similar to Prometheus metrics, alerts use labels to identify a particular dimensional instantiation, and the Alertmanager dashboard enables operators to preemptively silence alerts based on them.

Check Active Alerts from NCNs

Prometheus includes the /api/v1/alerts endpoint, which returns a JSON object containing the active alerts. From a non-compute node (NCN), can connect to sysmgmt-health/cray-sysmgmt-health-promet-prometheus directly and bypass service authentication and authorization.

Obtain the cluster IP address:

kubectl -n sysmgmt-health get svc cray-sysmgmt-health-promet-prometheus

Example output:

NAME                                    TYPE        CLUSTER-IP     EXTERNAL-IP   PORT(S)    AGE
cray-sysmgmt-health-promet-prometheus   ClusterIP   10.16.201.80   <none>        9090/TCP   2d6h

Get active alerts, which includes KubeletTooManyPods if it is going off:

curl -s http://CLUSTER-IP:PORT/api/v1/alerts | jq . | grep -B 10 -A 20 KubeletTooManyPods

Example output:

{
  "status": "success",
  "data": {
    "alerts": [
      {
        "labels": {
          "alertname": "KubeletTooManyPods",
          "endpoint": "https-metrics",
          "instance": "10.252.1.6:10250",
          "job": "kubelet",
          "namespace": "kube-system",
          "node": "ncn-w003",
          "prometheus": "kube-monitoring/cray-prometheus-operator-prometheus",
          "prometheus_replica": "prometheus-cray-prometheus-operator-prometheus-0",
          "service": "cray-prometheus-operator-kubelet",
          "severity": "warning"
        },
        "annotations": {
          "message": "Kubelet 10.252.1.6:10250 is running 107 Pods, close to the limit of 110.",
          "runbook_url": "https://github.com/kubernetes-monitoring/kubernetes-mixin/tree/master/runbook.md#alert-name-kubelettoomanypods"
        },
        "state": "firing",
        "activeAt": "2020-01-11T18:13:35.086499854Z",
        "value": 107
      },
      {
        "labels": {

In the example above, the alert actually indicates it is getting close to the limit, but the value included in the alert is the actual number of pods on ncn-w003.

Troubleshooting: If an alert titled KubeCronJobRunning is encountered, this could be an indication that a Kubernetes cronjob is misbehaving. The Labels section under the firing alert will indicate the name of the cronjob that is taking longer than expected to complete. Refer to the “CHECK CRON JOBS” header in the Power On and Start the Management Kubernetes Cluster procedure for instructions on how to troubleshoot the cronjob, as well as how to restart (export and reapply) the cronjob.