-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
"unable to get disk metrics" when deployed to kubernetes #1961
Comments
I was also encountering the same issue with Downgrading to Is it possible that your other nodes are running a different version of the dd-agent? Hope this helps! |
I'm seeing this with |
I had this issue with I then updated to EnvironmentKube 1.9.9 deployed with kops to AWS |
I'm having this issue as well with agent 6.3.2 on k8s 1.9.8 deployed on AWS via kops. I have only three nodes in this cluster, and the host nodes are running the |
Not sure if it matters, but this is not the standard AMI, but one which supports hvm, and |
Tuning the Linux OSes using |
For the record: To fix this in a Kubernetes deployment I followed this guide: https://docs.datadoghq.com/agent/kubernetes/integrations/#configmap Leading to these changes in the DaemonSet: volumeMounts:
- name: datadog-agent-config
mountPath: /conf.d
[...]
volumes:
- name: datadog-agent-config
configMap:
name: datadog-agent
items:
- key: disk-config
path: disk_check.yaml Along with this new ConfigMap: kind: ConfigMap
apiVersion: v1
metadata:
name: datadog-agent
namespace: monitoring
data:
disk-config: |-
init_config:
instances:
- use_mount: false
excluded_filesystems:
- autofs
- /proc/sys/fs/binfmt_misc This seems to get rid of the warnings... |
Superseded by DataDog/integrations-core#2492 |
for anyone using the helm chart seeing this issue, I use a values.yaml like this:
Turns out it's called |
This fix works, but it is a workaround. |
I recently found this issue in our Agents as well. Using the container agent: 7.17.0 and still encountering the same issue. Will be putting in the workaround |
Please comment here instead: DataDog/integrations-core#2492 |
For anyone else who finds this issue and is looking to disable these locations from the disk check, the option names have been renamed in this PR.
This is the configuration I used in my Helm values file:
|
In my case (Agent disk.yaml: |-
init_config:
instances:
## Instruct the check to collect using mount points instead of volumes.
- use_mount: true
## Collect data from root mountpoint (regex)
mount_point_include:
- ^/$ I took the idea of here. |
Ran into this as well on EKS. Deploying a new cluster with the datadog agent on the hosts in the EKS node group and the helmchart resulted in my logs from the host agent being spammed constantly with the following error:
Incorporating the following changes into my helmchart removed the automount requests on the hosts in the EKS node group: #1961 (comment) There have been attempts to ignore this already in the agent: DataDog/integrations-core#7650 but they dont seem to be working in this case. Confighelmchart confighelmchart pre fix:
helmchart post fix:
Version infoversion detailskube version:
helmchart version:
datadog-agent version in pod:
datadog-agent version on EKS host in node group:
|
Describe what happened:
I have the same issue as #1730, with the container deployed to kubernetes. The "solution" to issue #1730 does not apply to a Kubernetes deployment. The issue appears to happen on a subset of my Kubernetes nodes.
[ AGENT ] 2018-07-09 00:39:31 UTC | WARN | (datadog_agent.go:135 in LogMessage) | (disk.py:105) | Unable to get disk metrics for /host/proc/sys/fs/binfmt_misc: [Errno 40] Too many levels of symbolic links: '/host/proc/sys/fs/binfmt_misc'
Describe what you expected:
No errors.
Steps to reproduce the issue:
Currently its happening on 2 out of 7 nodes. So direct reproduction steps are uncertain.
Additional environment details (Operating System, Cloud provider, etc):
Kops deployed Kubernetes on AWS. Running latest DataDog container deployed via a DaemonSet
The text was updated successfully, but these errors were encountered: