Monitor Kubernetes metrics

Discover and scrape Kubernetes resources

Configure the Collector using one of the following methods to monitor Kubernetes resources and ensure that your clusters are healthy.

Monitor kubelet or cAdvisor metrics

If running in Kubernetes, you can configure the Collector to scrape kubelet or cAdvisor metrics by setting the kubeletMetricsEnabled or cadvisorMetricsEnabled flag to true under the kubeletMonitoring YAML collection.

For example:

discovery:
  kubernetes:
    ...
    kubeletMonitoring:
      port: 10250
      bearerTokenFile: "/var/run/secrets/kubernetes.io/serviceaccount/token"
      kubeletMetricsEnabled: true
      cadvisorMetricsEnabled: true
      probesMetricsEnabled: true
      labelsToAugment: []
      annotationsToAugment: []
  • port: Port the kubelet is running on. Defaults to 10250.
  • bearerTokenFile: Path to file containing Collector service account token. Defaults to "/var/run/secrets/kubernetes.io/serviceaccount/token"
  • kubeletMetricsEnabled: Enables scraping kubelet metrics.
  • cadvisorMetricsEnabled: Enables scraping cAdvisor metrics.
  • probesMetricsEnabled: Enables collecting metrics on the status of liveness, readiness, and startup kubelet probes for Kubernetes containers.
  • labelsToAugment: Lists the metadata labels from pod labels the Collector adds to metrics.
  • annotationsToAugment: Lists the metadata labels from pod annotations the Collector adds to metrics.

Add metadata labels from pod labels

By default, container-level metrics don't include metadata labels like service or app, which searches can include when querying for these metrics. To automatically add these labels from pod labels, use the labelsToAugment flag to list the labels the Collector adds to the metrics.

For example, to add the app label to the container level metrics for a node-exporter DaemonSet deployment, use the following configuration under the kubeletMonitoring key:

discovery:
  kubernetes:
    ...
    kubeletMonitoring:
      ...
      labelsToAugment: ["app", ...]

This adds app="node-exporter" to these metrics, based on the following example node-exporter manifest:

apiVersion: apps/v1
kind: DaemonSet
metadata:
  labels:
    app: node-exporter
  name: node-exporter
  namespace: monitoring
...

Add metadata labels from pod annotations

By default, container-level metrics don't include metadata labels, which searches can include when querying for these metrics. To automatically add these labels from pod annotations, use the annotationsToAugment flag to list the labels the Collector adds to the metrics.

For example, to add the app_kubernetes_io_component label to the container-level metrics for a node-exporter DaemonSet deployment, use the following configuration under the kubeletMonitoring key:

discovery:
  kubernetes:
    ...
    kubeletMonitoring:
      ...
      annotationsToAugment: ["app.kubernetes.io/component", ...]

This adds app_kubernetes_io_component="infrastructure" to these metrics, assuming the following example node-exporter manifest:

apiVersion: apps/v1
kind: DaemonSet
metadata:
  annotations:
    app.kubernetes.io/component: infrastructure
  name: node-exporter
  namespace: monitoring
...

Map Kubernetes labels to Prometheus labels

Requires Chronosphere Collector version 0.93.0 or later.

The Collector lets you specify pod labels and annotations you want to keep as a Prometheus label. This feature applies to pods only.

The following configuration example converts all pod labels called my_label and all pod annotations called my.pod.annotation into Prometheus labels for the metrics scraped from discovered pods. This is equivalent to a Prometheus labelmap rule, but sanitizes the label names and values:

discovery:
  kubernetes:
    enabled: true
    metadataConfig:
        - resource: "pod"
          annotationsToKeep:
            - my.pod.annotation
          labelsToKeep:
            - my_label

Discover kube-system endpoints

Requires Chronosphere Collector version 0.85.0 or later. Requires Kubernetes version 1.19.0 or later for your Kubernetes cluster.

To discover endpoints in the kube-system namespace, set the kubeSystemEndpointsDiscoveryEnabled flag to true. Because kube-system has many constantly changing endpoints that might cause unnecessary load on the Collector, the endpoint is disabled by default.

Using EndpointSlices significantly reduces the amount of load on the Kubernetes API server.

If you modify a Collector manifest, you must update it in the cluster and restart the Collector.

Discover and scrape kube-state-metrics

You can use ServiceMonitors to scrape kube-state-metrics, which generate metrics that track the health of deployments, nodes, and pods in a Kubernetes cluster. Monitoring these metrics can help to ensure the health of your cluster because the Collector expects to continually receive kube-state-metrics. If the Collector can't scrape these metrics, it's likely your Kubernetes cluster is experiencing issues you need to resolve.

Monitoring kube-state-metrics with a DaemonSet Collector is manageable for smaller clusters, but can lead to out of memory (OOM) errors as the cluster scales. Chronosphere recommends running the Collector as a sidecar to take advantage of staleness markers. The following steps assume that:

  • You're running a separate Collector as a Deployment to monitor kube-state-metrics.
  • You've already defined a Kubernetes Service and ServiceMonitor for kube-state-metrics.

If you're already running the Collector as a DaemonSet, you must update the manifest for both Collector instances.

After installing the ServiceMonitors CRD, complete the following steps to discover kube-state-metrics:

  1. Download this manifest (opens in a new tab).

  2. In the data section, replace the values for address and api-token with your Base64-encoded API token:

    ---
    apiVersion: v1
    data:
      address: <add-base64-encoded-token-here>
      api-token: <add-base64-encoded-token-here>
  3. Apply the manifest:

    kubectl apply -f path/to/ksm-chronocollector.yaml
  4. Confirm the Deployment is started and running, and view the logs of the pod:

    kubectl get pods
     
    NAME                    READY   STATUS    RESTARTS   AGE
    chronocollector-jtgfw   1/1     Running   0          1m
     
    kubectl logs chronocollector-jtgfw
    ...

Ingest Kubernetes API server metrics

The Kubernetes API Server provides REST operations and a frontend to a cluster's shared state through which all other components interact. Unlike most other metrics emitted from a cluster, Kubernetes doesn't expose API Server metrics by using a pod, but instead exposes metrics directly from an endpoint in the API Server.

To ingest these metrics through traditional service discovery methods, you must discover and scrape the endpoints directly. The Collector supports using ServiceMonitors or job service discovery.

Discover API Server metrics with ServiceMonitors

Discover metrics for a managed Kubernetes cluster

To use ServiceMonitors to discover and scrape API server metrics from a managed Kubernetes cluster, such as Amazon Elastic Kubernetes Service (EKS) or Google Kubernetes Engine (GKE), enable both the allowSkipPodInfo flag under the top level serviceMonitor key and the endpointsDiscoveryEnabled flag under the discovery > kubernetes in the Collector configuration.

serviceMonitor:
  allowSkipPodInfo: true

discovery:
  kubernetes:
    endpointsDiscoveryEnabled: true
⚠️

If using this method, deploy the Collector as an individual Kubernetes Deployment rather than a DaemonSet. This prevents multiple Collectors from scraping the same API Server metrics. Additionally, set the appropriate ServiceMonitor matching to prevent other Collectors from discovering the API Server ServiceMonitor.

Discover metrics for a self-managed Kubernetes cluster

To use ServiceMonitors to discover and scrape API server metrics from a self-managed Kubernetes cluster, such as k0ps, enable the endpointsDiscoveryEnabled flag under the discovery > kubernetes YAML collection in the Collector configuration:

discovery:
  kubernetes:
    endpointsDiscoveryEnabled: true

In this configuration, you can deploy the Collector as a DaemonSet, which installs the Collector on the master nodes that run the Kubernetes API Server so the Collector can scrape it.

Discover API Server metrics with the jobs service

To discover and scrape API Server metrics without using ServiceMonitors, you can use the jobs section of the Collector configuration for service discovery.

⚠️

If using this method, deploy the Collector as an individual Kubernetes Deployment instead of as a DaemonSet. This prevents multiple Collectors from scraping the same API Server metrics.

The following is an example that discovers the API Server based on the value of the __meta_kubernetes_pod_label_k8s_app label equal to kube-apiserver (found in the API Server Service object).

jobs:
  - name: kube-apiserver
    selector:
      kubernetes:
        matchLabels:
          __meta_kubernetes_pod_label_k8s_app: kube-apiserver
    options:
      scheme: https
      http:
        bearerTokenFile: /var/run/secrets/kubernetes.io/serviceaccount/token
        tls:
          caFile: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt
          insecureSkipVerify: true
      relabels:
        - sourceLabels: [__name__]
          targetLabel: service
          replacement: kube-apiserver
        - sourceLabels: [__meta_kubernetes_pod_node_name]
          targetLabel: node