Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial unix /host/podresources/kubelet.sock: connect: permission denied" #222

Open
dgabrysch opened this issue Sep 25, 2024 · 2 comments
Labels
lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.

Comments

@dgabrysch
Copy link

dgabrysch commented Sep 25, 2024

Hello,

we are using the must-gather operator on OpenShift 4.15 and getting this log:

Collecting performance related data for node NODENAME perf-node-gather-daemonset-crgb6
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial unix /host/podresources/kubelet.sock: connect: permission denied"
command terminated with exit code 1

The requested sock does not exist on RHOCS - can we change the behaviour or something?

Here is the MustGather yaml:

apiVersion: managed.openshift.io/v1alpha1
kind: MustGather
metadata:
  generateName: mustgather-
  resourceVersion: '786104'
  name: mustgather-q7bpx
  uid: 1f03bf2f-5970-42d4-9680-ce8708352194
  creationTimestamp: '2024-09-25T09:20:07Z'
  generation: 2
  managedFields:
    - apiVersion: managed.openshift.io/v1alpha1
      fieldsType: FieldsV1
      fieldsV1:
        'f:metadata':
          'f:generateName': {}
        'f:spec':
          .: {}
          'f:audit': {}
          'f:caseID': {}
          'f:caseManagementAccountSecretRef': {}
          'f:internalUser': {}
          'f:proxyConfig':
            .: {}
            'f:httpProxy': {}
            'f:httpsProxy': {}
            'f:noProxy': {}
          'f:serviceAccountRef': {}
      manager: OpenAPI-Generator
      operation: Update
      time: '2024-09-25T09:20:07Z'
    - apiVersion: managed.openshift.io/v1alpha1
      fieldsType: FieldsV1
      fieldsV1:
        'f:metadata':
          'f:finalizers':
            .: {}
            'v:"finalizer.mustgathers.managed.openshift.io"': {}
        'f:spec':
          'f:mustGatherTimeout': {}
      manager: must-gather-operator
      operation: Update
      time: '2024-09-25T09:20:08Z'
    - apiVersion: managed.openshift.io/v1alpha1
      fieldsType: FieldsV1
      fieldsV1:
        'f:status':
          .: {}
          'f:completed': {}
          'f:conditions': {}
      manager: must-gather-operator
      operation: Update
      subresource: status
      time: '2024-09-25T09:23:01Z'
  namespace: openshift-must-gather-operator
  finalizers:
    - finalizer.mustgathers.managed.openshift.io
spec:
  audit: false
  caseID: randomid-redacted
  caseManagementAccountSecretRef:
    name: case-management-creds
  internalUser: true
  mustGatherTimeout: 0s
  proxyConfig:
redacted
  serviceAccountRef:
    name: must-gather-admin
status:
  completed: false
  conditions:
    - lastTransitionTime: '2024-09-25T09:23:01Z'
      message: ''
      observedGeneration: 2
      reason: LastReconcileCycleSucceded
      status: 'True'
      type: ReconcileSuccess

These are the image versions we have used for testing:
quay.io/app-sre/must-gather-operator:v4.15.281-g21414eb
quay.io/openshift/origin-must-gather:4.15.0

OpenShift version: 4.15.28

The serviceaccounts default, must-gather-operator and must-gather-admin have cluster-admin privileges and are able to use the privileged scc (for debugging purposes)

Any help is appreciated!

@openshift-bot
Copy link

Issues go stale after 90d of inactivity.

Mark the issue as fresh by commenting /remove-lifecycle stale.
Stale issues rot after an additional 30d of inactivity and eventually close.
Exclude this issue from closing by commenting /lifecycle frozen.

If this issue is safe to close now please do so with /close.

/lifecycle stale

@openshift-ci openshift-ci bot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Dec 25, 2024
@openshift-bot
Copy link

Stale issues rot after 30d of inactivity.

Mark the issue as fresh by commenting /remove-lifecycle rotten.
Rotten issues close after an additional 30d of inactivity.
Exclude this issue from closing by commenting /lifecycle frozen.

If this issue is safe to close now please do so with /close.

/lifecycle rotten
/remove-lifecycle stale

@openshift-ci openshift-ci bot added lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. and removed lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. labels Jan 24, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.
Projects
None yet
Development

No branches or pull requests

2 participants