Get metrics from kubernetes service in real time to:
- Visualize and monitor kubernetes states
- Be notified about kubernetes failovers and events.
Agent6 uses a new set of integrations, see the update instructions and the new dedicated kubernetes documentation page for more information.
The Kubernetes check is included in the Datadog Agent package, so you don't need to install anything else on your Kubernetes servers.
Edit the kubernetes.yaml file to point to your server and port, set the masters to monitor. See the sample kubernetes.yaml for all available configuration options.
As the 5.17.0 release, Datadog Agent now supports built in leader election option for the Kubernetes event collector. Agents coordinate by performing a leader election among members of the Datadog DaemonSet through kubernetes to ensure only one leader agent instance is gathering events at a given time. If the leader agent instance fails, a re-election occurs and another cluster agent will take over collection.
This functionality is disabled by default.
To enable leader election you need to set the variable leader_candidate to true in your kubernetes.yaml file.
This feature relies on ConfigMaps , so you will need to grant Datadog Agent get, list, delete and create access to the ConfigMap resource.
Use these Kubernetes RBAC entities for your Datadog agent to properly configure the previous permissions by applying this datadog service account to your pods.
kind: ClusterRole
apiVersion: rbac.authorization.k8s.io/v1beta1
metadata:
name: datadog
rules:
- nonResourceURLs:
- "/version" # Used to get apiserver version metadata
- "/healthz" # Healthcheck
verbs: ["get"]
- apiGroups: [""]
resources:
- "nodes"
- "namespaces" #
- "events" # Cluster events + kube_service cache invalidation
- "services" # kube_service tag
verbs: ["get", "list"]
- apiGroups: [""]
resources:
- "configmaps"
resourceNames: ["datadog-leader-elector"]
verbs: ["get", "delete", "update"]
- apiGroups: [""]
resources:
- "configmaps"
verbs: ["create"]
---
# You need to use that account for your dd-agent DaemonSet
apiVersion: v1
kind: ServiceAccount
metadata:
name: datadog
automountServiceAccountToken: true
---
# Your admin user needs the same permissions to be able to grant them
# Easiest way is to bind your user to the cluster-admin role
# See https://cloud.google.com/container-engine/docs/role-based-access-control#setting_up_role-based_access_control
kind: ClusterRoleBinding
apiVersion: rbac.authorization.k8s.io/v1beta1
metadata:
name: datadog
subjects:
- kind: ServiceAccount
name: datadog
namespace: default
roleRef:
kind: ClusterRole
name: datadog
apiGroup: rbac.authorization.k8s.ioIn your kubernetes.d/conf.yaml file you will see the leader_lease_duration parameter. It's the duration for which a leader stays elected. It should be > 30 seconds.
The longer it is, the less hard your agent hits the apiserver with requests, but it also means that if the leader dies (and under certain conditions) there can be an event blackout until the lease expires and a new leader takes over.
Run the Agent's status subcommand and look for kubernetes under the Checks section.
See metadata.csv for a list of metrics provided by this integration.
As the 5.17.0 release, Datadog Agent now supports built in leader election option for the Kubernetes event collector. Once enabled, you no longer need to deploy an additional event collection container to your cluster. Instead agents will coordinate to ensure only one agent instance is gathering events at a given time, events below will be available:
- Backoff
- Conflict
- Delete
- DeletingAllPods
- Didn't have enough resource
- Error
- Failed
- FailedCreate
- FailedDelete
- FailedMount
- FailedSync
- Failedvalidation
- FreeDiskSpaceFailed
- HostPortConflict
- InsufficientFreeCPU
- InsufficientFreeMemory
- InvalidDiskCapacity
- Killing
- KubeletsetupFailed
- NodeNotReady
- NodeoutofDisk
- OutofDisk
- Rebooted
- TerminatedAllPods
- Unable
- Unhealthy
The Kubernetes check does not include any service checks.
Yes, since Kubernetes 1.6, the concept of Taints and tolerations was introduced. Now rather than the master being off limits, it's simply tainted. Add the required toleration to the pod to run it:
Add the following lines to your Deployment (or Daemonset if you are running a multi-master setup):
spec:
tolerations:
- key: node-role.kubernetes.io/master
effect: NoSchedule
The agent assumes that the kubelet API is available at the default gateway of the container. If that's not the case because you are using a software defined networks like Calico or Flannel, the agent needs to be specified using an environment variable:
- name: KUBERNETES_KUBELET_HOST
valueFrom:
fieldRef:
fieldPath: spec.nodeName
See this PR
These are pause containers (docker_image:gcr.io/google_containers/pause.*) that K8s injects into every pod to keep it populated even if the "real" container is restarting/stopped.
The docker_daemon check ignores them through a default exclusion list, but they will show up for K8s metrics like kubernetes.cpu.usage.total and kubernetes.filesystem.usage.
To get a better idea of how (or why) to integrate your Kubernetes service, check out our series of blog posts about it.
