<?xml version="1.0" encoding="UTF-8"?>
  <testsuites tests="34" disabled="0" errors="0" failures="16" time="34.865022123">
      <testsuite name="Kubernetes e2e suite" package="/usr/local/bin" tests="34" disabled="0" skipped="0" errors="0" failures="16" time="34.865022123" timestamp="2026-04-02T15:02:32">
          <properties>
              <property name="SuiteSucceeded" value="false"></property>
              <property name="SuiteHasProgrammaticFocus" value="false"></property>
              <property name="SpecialSuiteFailureReason" value=""></property>
              <property name="SuiteLabels" value="[]"></property>
              <property name="RandomSeed" value="1775142147"></property>
              <property name="RandomizeAllSpecs" value="true"></property>
              <property name="LabelFilter" value=""></property>
              <property name="FocusStrings" value="\[Conformance\]"></property>
              <property name="SkipStrings" value=""></property>
              <property name="FocusFiles" value=""></property>
              <property name="SkipFiles" value=""></property>
              <property name="FailOnPending" value="false"></property>
              <property name="FailOnEmpty" value="false"></property>
              <property name="FailFast" value="false"></property>
              <property name="FlakeAttempts" value="0"></property>
              <property name="DryRun" value="false"></property>
              <property name="ParallelTotal" value="16"></property>
              <property name="OutputInterceptorMode" value=""></property>
          </properties>
          <testcase name="[ReportBeforeSuite]" classname="Kubernetes e2e suite" status="passed" time="0.000537533"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.30488409">
              <failure message="" type="failed">[FAILED] Error waiting for all pods to be running and ready: Told to stop trying after 18.147s.&#xA;Unexpected final error while getting *pod.state: listing replication controllers in namespace kube-system: etcdserver: request timed out&#xA;At one point, however, the function did return successfully.&#xA;Yet, Eventually failed because the matcher was not satisfied:&#xA;Expected all pods (need at least 0) in namespace &#34;kube-system&#34; to be running and ready (except for 0).&#xA;25 / 27 pods were running and ready.&#xA;Expected 5 pod replicas, 4 are Running and Ready.&#xA;Pods that were neither completed nor running:&#xA;    &lt;[]v1.Pod | len:2, cap:2&gt;: &#xA;        - metadata:&#xA;            annotations:&#xA;              cni.projectcalico.org/containerID: 18f9c5d3bd5023ece4b6d3fe23b14c208d2d7c525b8a31ba893326fcb3cc63a0&#xA;              cni.projectcalico.org/podIP: 10.100.78.131/32&#xA;              cni.projectcalico.org/podIPs: 10.100.78.131/32&#xA;            creationTimestamp: &#34;2026-04-02T14:56:28Z&#34;&#xA;            generateName: manila-csi-openstack-manila-csi-controllerplugin-&#xA;            generation: 1&#xA;            labels:&#xA;              app: openstack-manila-csi&#xA;              apps.kubernetes.io/pod-index: &#34;0&#34;&#xA;              chart: openstack-manila-csi-2.32.0&#xA;              component: controllerplugin&#xA;              controller-revision-hash: manila-csi-openstack-manila-csi-controllerplugin-6f488f9967&#xA;              heritage: Helm&#xA;              release: manila-csi&#xA;              statefulset.kubernetes.io/pod-name: manila-csi-openstack-manila-csi-controllerplugin-0&#xA;            managedFields:&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:metadata:&#xA;                  f:generateName: {}&#xA;                  f:labels:&#xA;                    .: {}&#xA;                    f:app: {}&#xA;                    f:apps.kubernetes.io/pod-index: {}&#xA;                    f:chart: {}&#xA;                    f:component: {}&#xA;                    f:controller-revision-hash: {}&#xA;                    f:heritage: {}&#xA;                    f:release: {}&#xA;                    f:statefulset.kubernetes.io/pod-name: {}&#xA;                  f:ownerReferences:&#xA;                    .: {}&#xA;                    k:{&#34;uid&#34;:&#34;1e454624-fa6c-46e9-8a02-a1769df05b94&#34;}: {}&#xA;                f:spec:&#xA;                  f:containers:&#xA;                    k:{&#34;name&#34;:&#34;nfs-nodeplugin&#34;}:&#xA;                      .: {}&#xA;                      f:command: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;CSI_ENDPOINT&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                        k:{&#34;name&#34;:&#34;DRIVER_NAME&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                        k:{&#34;name&#34;:&#34;FWD_CSI_ENDPOINT&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                        k:{&#34;name&#34;:&#34;MANILA_SHARE_PROTO&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext:&#xA;                        .: {}&#xA;                        f:allowPrivilegeEscalation: {}&#xA;                        f:capabilities:&#xA;                          .: {}&#xA;                          f:add: {}&#xA;                        f:privileged: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/etc/config&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                          f:readOnly: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/plugins/csi-nfsplugin&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/pods&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:mountPropagation: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;nfs-provisioner&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;nfs-resizer&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;nfs-snapshotter&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                  f:dnsPolicy: {}&#xA;                  f:enableServiceLinks: {}&#xA;                  f:hostname: {}&#xA;                  f:restartPolicy: {}&#xA;                  f:schedulerName: {}&#xA;                  f:securityContext: {}&#xA;                  f:serviceAccount: {}&#xA;                  f:serviceAccountName: {}&#xA;                  f:subdomain: {}&#xA;                  f:terminationGracePeriodSeconds: {}&#xA;                  f:tolerations: {}&#xA;                  f:volumes:&#xA;                    .: {}&#xA;                    k:{&#34;name&#34;:&#34;cloud-config&#34;}:&#xA;                      .: {}&#xA;                      f:name: {}&#xA;                      f:secret:&#xA;                        .: {}&#xA;                        f:defaultMode: {}&#xA;                        f:items: {}&#xA;                        f:secretName: {}&#xA;                    k:{&#34;name&#34;:&#34;nfs-fwd-plugin-dir&#34;}:&#xA;                      .: {}&#xA;                      f:hostPath:&#xA;                        .: {}&#xA;                        f:path: {}&#xA;                        f:type: {}&#xA;                      f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;nfs-plugin-dir&#34;}:&#xA;                      .: {}&#xA;                      f:hostPath:&#xA;                        .: {}&#xA;                        f:path: {}&#xA;                        f:type: {}&#xA;                      f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;pod-mounts&#34;}:&#xA;                      .: {}&#xA;                      f:hostPath:&#xA;                        .: {}&#xA;                        f:path: {}&#xA;                        f:type: {}&#xA;                      f:name: {}&#xA;              manager: kube-controller-manager&#xA;              operation: Update&#xA;              time: &#34;2026-04-02T14:56:28Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:status:&#xA;                  f:conditions:&#xA;                    .: {}&#xA;                    k:{&#34;type&#34;:&#34;PodScheduled&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;              manager: kube-scheduler&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:03Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:status:&#xA;                  f:conditions:&#xA;                    k:{&#34;type&#34;:&#34;ContainersReady&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;Initialized&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;PodReadyToStartContainers&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;Ready&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                  f:containerStatuses: {}&#xA;                  f:startTime: {}&#xA;              manager: kubelet&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:53Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:metadata:&#xA;                  f:annotations:&#xA;                    .: {}&#xA;                    f:cni.projectcalico.org/containerID: {}&#xA;                    f:cni.projectcalico.org/podIP: {}&#xA;                    f:cni.projectcalico.org/podIPs: {}&#xA;              manager: calico&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:55Z&#34;&#xA;            name: manila-csi-openstack-manila-csi-controllerplugin-0&#xA;            namespace: kube-system&#xA;            ownerReferences:&#xA;            - apiVersion: apps/v1&#xA;              blockOwnerDeletion: true&#xA;              controller: true&#xA;              kind: StatefulSet&#xA;              name: manila-csi-openstack-manila-csi-controllerplugin&#xA;              uid: 1e454624-fa6c-46e9-8a02-a1769df05b94&#xA;            resourceVersion: &#34;1549&#34;&#xA;            uid: f3c4b392-db35-43b3-8e28-3e3d429adcf7&#xA;          spec:&#xA;            containers:&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: unix:///var/lib/kubelet/plugins/nfs.manila.csi.openstack.org/csi-controllerplugin.sock&#xA;              image: registry.k8s.io/sig-storage/csi-provisioner:v5.1.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: nfs-provisioner&#xA;              resources: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: unix:///var/lib/kubelet/plugins/nfs.manila.csi.openstack.org/csi-controllerplugin.sock&#xA;              image: registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: nfs-snapshotter&#xA;              resources: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              - --handle-volume-inuse-error=false&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: unix:///var/lib/kubelet/plugins/nfs.manila.csi.openstack.org/csi-controllerplugin.sock&#xA;              image: registry.k8s.io/sig-storage/csi-resizer:v1.12.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: nfs-resizer&#xA;              resources: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;            - command:&#xA;              - /bin/sh&#xA;              - -c&#xA;              - /bin/manila-csi-plugin -v=2 --endpoint=$(CSI_ENDPOINT) --drivername=$(DRIVER_NAME)&#xA;                --share-protocol-selector=$(MANILA_SHARE_PROTO) --fwdendpoint=$(FWD_CSI_ENDPOINT)&#xA;                --cluster-id=&#34;&#34;&#xA;              env:&#xA;              - name: DRIVER_NAME&#xA;                value: nfs.manila.csi.openstack.org&#xA;              - name: CSI_ENDPOINT&#xA;                value: unix:///var/lib/kubelet/plugins/nfs.manila.csi.openstack.org/csi-controllerplugin.sock&#xA;              - name: FWD_CSI_ENDPOINT&#xA;                value: unix:///var/lib/kubelet/plugins/csi-nfsplugin/csi.sock&#xA;              - name: MANILA_SHARE_PROTO&#xA;                value: NFS&#xA;              image: registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: nfs-nodeplugin&#xA;              resources: {}&#xA;              securityContext:&#xA;                allowPrivilegeEscalation: true&#xA;                capabilities:&#xA;                  add:&#xA;                  - SYS_ADMIN&#xA;                privileged: true&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/lib/kubelet/plugins/csi-nfsplugin&#xA;                name: nfs-fwd-plugin-dir&#xA;              - mountPath: /var/lib/kubelet/pods&#xA;                mountPropagation: Bidirectional&#xA;                name: pod-mounts&#xA;              - mountPath: /etc/config&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;            dnsPolicy: ClusterFirst&#xA;            enableServiceLinks: true&#xA;            hostname: manila-csi-openstack-manila-csi-controllerplugin-0&#xA;            nodeName: kube-uhjhf-67cxd-gq627&#xA;            preemptionPolicy: PreemptLowerPriority&#xA;            priority: 0&#xA;            restartPolicy: Always&#xA;            schedulerName: default-scheduler&#xA;            securityContext: {}&#xA;            serviceAccount: manila-csi-openstack-manila-csi-controllerplugin&#xA;            serviceAccountName: manila-csi-openstack-manila-csi-controllerplugin&#xA;            subdomain: manila-csi-openstack-manila-csi-controllerplugin&#xA;            terminationGracePeriodSeconds: 30&#xA;            tolerations:&#xA;            - effect: NoSchedule&#xA;              key: node-role.kubernetes.io/master&#xA;            - effect: NoSchedule&#xA;              key: node-role.kubernetes.io/control-plane&#xA;            - effect: NoExecute&#xA;              key: node.kubernetes.io/not-ready&#xA;              operator: Exists&#xA;              tolerationSeconds: 300&#xA;            - effect: NoExecute&#xA;              key: node.kubernetes.io/unreachable&#xA;              operator: Exists&#xA;              tolerationSeconds: 300&#xA;            volumes:&#xA;            - hostPath:&#xA;                path: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                type: DirectoryOrCreate&#xA;              name: nfs-plugin-dir&#xA;            - hostPath:&#xA;                path: /var/lib/kubelet/plugins/csi-nfsplugin&#xA;                type: Directory&#xA;              name: nfs-fwd-plugin-dir&#xA;            - hostPath:&#xA;                path: /var/lib/kubelet/pods&#xA;                type: Directory&#xA;              name: pod-mounts&#xA;            - name: cloud-config&#xA;              secret:&#xA;                defaultMode: 420&#xA;                items:&#xA;                - key: ca.crt&#xA;                  path: ca.crt&#xA;                secretName: cloud-config&#xA;            - name: kube-api-access-tq4nx&#xA;              projected:&#xA;                defaultMode: 420&#xA;                sources:&#xA;                - serviceAccountToken:&#xA;                    expirationSeconds: 3607&#xA;                    path: token&#xA;                - configMap:&#xA;                    items:&#xA;                    - key: ca.crt&#xA;                      path: ca.crt&#xA;                    name: kube-root-ca.crt&#xA;                - downwardAPI:&#xA;                    items:&#xA;                    - fieldRef:&#xA;                        apiVersion: v1&#xA;                        fieldPath: metadata.namespace&#xA;                      path: namespace&#xA;          status:&#xA;            conditions:&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              status: &#34;False&#34;&#xA;              type: PodReadyToStartContainers&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              status: &#34;True&#34;&#xA;              type: Initialized&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              message: &#39;containers with unready status: [nfs-provisioner nfs-snapshotter nfs-resizer&#xA;                nfs-nodeplugin]&#39;&#xA;              reason: ContainersNotReady&#xA;              status: &#34;False&#34;&#xA;              type: Ready&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              message: &#39;containers with unready status: [nfs-provisioner nfs-snapshotter nfs-resizer&#xA;                nfs-nodeplugin]&#39;&#xA;              reason: ContainersNotReady&#xA;              status: &#34;False&#34;&#xA;              type: ContainersReady&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:52Z&#34;&#xA;              status: &#34;True&#34;&#xA;              type: PodScheduled&#xA;            containerStatuses:&#xA;            - image: registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: nfs-nodeplugin&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/lib/kubelet/plugins/csi-nfsplugin&#xA;                name: nfs-fwd-plugin-dir&#xA;              - mountPath: /var/lib/kubelet/pods&#xA;                name: pod-mounts&#xA;              - mountPath: /etc/config&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-provisioner:v5.1.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: nfs-provisioner&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-resizer:v1.12.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: nfs-resizer&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: nfs-snapshotter&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            phase: Pending&#xA;            qosClass: BestEffort&#xA;            startTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;        - metadata:&#xA;            annotations:&#xA;              cni.projectcalico.org/containerID: cf0f08707afe3c7479ac743e82356119aae0fadac55f61f29ca480473dd08907&#xA;              cni.projectcalico.org/podIP: 10.100.78.132/32&#xA;              cni.projectcalico.org/podIPs: 10.100.78.132/32&#xA;            creationTimestamp: &#34;2026-04-02T14:56:27Z&#34;&#xA;            generateName: openstack-cinder-csi-controllerplugin-78f966889b-&#xA;            generation: 1&#xA;            labels:&#xA;              app: openstack-cinder-csi&#xA;              chart: openstack-cinder-csi-2.32.0&#xA;              component: controllerplugin&#xA;              heritage: Helm&#xA;              pod-template-hash: 78f966889b&#xA;              release: cinder-csi&#xA;            managedFields:&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:metadata:&#xA;                  f:generateName: {}&#xA;                  f:labels:&#xA;                    .: {}&#xA;                    f:app: {}&#xA;                    f:chart: {}&#xA;                    f:component: {}&#xA;                    f:heritage: {}&#xA;                    f:pod-template-hash: {}&#xA;                    f:release: {}&#xA;                  f:ownerReferences:&#xA;                    .: {}&#xA;                    k:{&#34;uid&#34;:&#34;3bd9f6ed-6669-41c3-9246-73d536880313&#34;}: {}&#xA;                f:spec:&#xA;                  f:affinity: {}&#xA;                  f:containers:&#xA;                    k:{&#34;name&#34;:&#34;cinder-csi-plugin&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;CLOUD_CONFIG&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                        k:{&#34;name&#34;:&#34;CLUSTER_NAME&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                        k:{&#34;name&#34;:&#34;CSI_ENDPOINT&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:livenessProbe:&#xA;                        .: {}&#xA;                        f:failureThreshold: {}&#xA;                        f:httpGet:&#xA;                          .: {}&#xA;                          f:path: {}&#xA;                          f:port: {}&#xA;                          f:scheme: {}&#xA;                        f:initialDelaySeconds: {}&#xA;                        f:periodSeconds: {}&#xA;                        f:successThreshold: {}&#xA;                        f:timeoutSeconds: {}&#xA;                      f:name: {}&#xA;                      f:ports:&#xA;                        .: {}&#xA;                        k:{&#34;containerPort&#34;:9808,&#34;protocol&#34;:&#34;TCP&#34;}:&#xA;                          .: {}&#xA;                          f:containerPort: {}&#xA;                          f:name: {}&#xA;                          f:protocol: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/csi&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/etc/config&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                          f:readOnly: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/etc/kubernetes&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                          f:readOnly: {}&#xA;                    k:{&#34;name&#34;:&#34;csi-attacher&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/csi/sockets/pluginproxy/&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;csi-provisioner&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/csi/sockets/pluginproxy/&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;csi-resizer&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/csi/sockets/pluginproxy/&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;csi-snapshotter&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/csi/sockets/pluginproxy/&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;liveness-probe&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/csi/sockets/pluginproxy/&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                  f:dnsPolicy: {}&#xA;                  f:enableServiceLinks: {}&#xA;                  f:restartPolicy: {}&#xA;                  f:schedulerName: {}&#xA;                  f:securityContext: {}&#xA;                  f:serviceAccount: {}&#xA;                  f:serviceAccountName: {}&#xA;                  f:terminationGracePeriodSeconds: {}&#xA;                  f:tolerations: {}&#xA;                  f:volumes:&#xA;                    .: {}&#xA;                    k:{&#34;name&#34;:&#34;cloud-config&#34;}:&#xA;                      .: {}&#xA;                      f:name: {}&#xA;                      f:secret:&#xA;                        .: {}&#xA;                        f:defaultMode: {}&#xA;                        f:secretName: {}&#xA;                    k:{&#34;name&#34;:&#34;socket-dir&#34;}:&#xA;                      .: {}&#xA;                      f:emptyDir: {}&#xA;                      f:name: {}&#xA;              manager: kube-controller-manager&#xA;              operation: Update&#xA;              time: &#34;2026-04-02T14:56:27Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:status:&#xA;                  f:conditions:&#xA;                    .: {}&#xA;                    k:{&#34;type&#34;:&#34;PodScheduled&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;              manager: kube-scheduler&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:03Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:status:&#xA;                  f:conditions:&#xA;                    k:{&#34;type&#34;:&#34;ContainersReady&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;Initialized&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;PodReadyToStartContainers&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;Ready&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                  f:containerStatuses: {}&#xA;                  f:hostIP: {}&#xA;                  f:hostIPs: {}&#xA;                  f:startTime: {}&#xA;              manager: kubelet&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:53Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:metadata:&#xA;                  f:annotations:&#xA;                    .: {}&#xA;                    f:cni.projectcalico.org/containerID: {}&#xA;                    f:cni.projectcalico.org/podIP: {}&#xA;                    f:cni.projectcalico.org/podIPs: {}&#xA;              manager: calico&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:58Z&#34;&#xA;            name: openstack-cinder-csi-controllerplugin-78f966889b-plrvm&#xA;            namespace: kube-system&#xA;            ownerReferences:&#xA;            - apiVersion: apps/v1&#xA;              blockOwnerDeletion: true&#xA;              controller: true&#xA;              kind: ReplicaSet&#xA;              name: openstack-cinder-csi-controllerplugin-78f966889b&#xA;              uid: 3bd9f6ed-6669-41c3-9246-73d536880313&#xA;            resourceVersion: &#34;1562&#34;&#xA;            uid: 4f0b1844-d34b-44cc-a923-06bca632e956&#xA;          spec:&#xA;            affinity: {}&#xA;            containers:&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              - --timeout=3m&#xA;              - --leader-election=true&#xA;              - --default-fstype=ext4&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: /var/lib/csi/sockets/pluginproxy/csi.sock&#xA;              image: registry.k8s.io/sig-storage/csi-attacher:v4.7.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: csi-attacher&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              - --timeout=3m&#xA;              - --leader-election=true&#xA;              - --default-fstype=ext4&#xA;              - --feature-gates=Topology=true&#xA;              - --extra-create-metadata&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: /var/lib/csi/sockets/pluginproxy/csi.sock&#xA;              image: registry.k8s.io/sig-storage/csi-provisioner:v5.1.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: csi-provisioner&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              - --timeout=3m&#xA;              - --leader-election=true&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: /var/lib/csi/sockets/pluginproxy/csi.sock&#xA;              image: registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: csi-snapshotter&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              - --timeout=3m&#xA;              - --handle-volume-inuse-error=false&#xA;              - --leader-election=true&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: /var/lib/csi/sockets/pluginproxy/csi.sock&#xA;              image: registry.k8s.io/sig-storage/csi-resizer:v1.12.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: csi-resizer&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: /var/lib/csi/sockets/pluginproxy/csi.sock&#xA;              image: registry.k8s.io/sig-storage/livenessprobe:v2.14.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: liveness-probe&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            - args:&#xA;              - /bin/cinder-csi-plugin&#xA;              - -v=2&#xA;              - --endpoint=$(CSI_ENDPOINT)&#xA;              - --cloud-config=$(CLOUD_CONFIG)&#xA;              - --cluster=$(CLUSTER_NAME)&#xA;              - --provide-node-service=false&#xA;              env:&#xA;              - name: CSI_ENDPOINT&#xA;                value: unix://csi/csi.sock&#xA;              - name: CLOUD_CONFIG&#xA;                value: /etc/kubernetes/cloud.conf&#xA;              - name: CLUSTER_NAME&#xA;                value: a5757e38-87f8-44aa-b5e9-2a9e6bc4c1c3&#xA;              image: registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              livenessProbe:&#xA;                failureThreshold: 5&#xA;                httpGet:&#xA;                  path: /healthz&#xA;                  port: healthz&#xA;                  scheme: HTTP&#xA;                initialDelaySeconds: 10&#xA;                periodSeconds: 60&#xA;                successThreshold: 1&#xA;                timeoutSeconds: 10&#xA;              name: cinder-csi-plugin&#xA;              ports:&#xA;              - containerPort: 9808&#xA;                name: healthz&#xA;                protocol: TCP&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /csi&#xA;                name: socket-dir&#xA;              - mountPath: /etc/kubernetes&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;              - mountPath: /etc/config&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            dnsPolicy: ClusterFirst&#xA;            enableServiceLinks: true&#xA;            nodeName: kube-uhjhf-67cxd-gq627&#xA;            preemptionPolicy: PreemptLowerPriority&#xA;            priority: 0&#xA;            restartPolicy: Always&#xA;            schedulerName: default-scheduler&#xA;            securityContext: {}&#xA;            serviceAccount: csi-cinder-controller-sa&#xA;            serviceAccountName: csi-cinder-controller-sa&#xA;            terminationGracePeriodSeconds: 30&#xA;            tolerations:&#xA;            - effect: NoSchedule&#xA;              key: node-role.kubernetes.io/master&#xA;            - effect: NoSchedule&#xA;              key: node-role.kubernetes.io/control-plane&#xA;            - effect: NoExecute&#xA;              key: node.kubernetes.io/not-ready&#xA;              operator: Exists&#xA;              tolerationSeconds: 300&#xA;            - effect: NoExecute&#xA;              key: node.kubernetes.io/unreachable&#xA;              operator: Exists&#xA;              tolerationSeconds: 300&#xA;            volumes:&#xA;            - emptyDir: {}&#xA;              name: socket-dir&#xA;            - name: cloud-config&#xA;              secret:&#xA;                defaultMode: 420&#xA;                secretName: cloud-config&#xA;            - name: kube-api-access-6tzzg&#xA;              projected:&#xA;                defaultMode: 420&#xA;                sources:&#xA;                - serviceAccountToken:&#xA;                    expirationSeconds: 3607&#xA;                    path: token&#xA;                - configMap:&#xA;                    items:&#xA;                    - key: ca.crt&#xA;                      path: ca.crt&#xA;                    name: kube-root-ca.crt&#xA;                - downwardAPI:&#xA;                    items:&#xA;                    - fieldRef:&#xA;                        apiVersion: v1&#xA;                        fieldPath: metadata.namespace&#xA;                      path: namespace&#xA;          status:&#xA;            conditions:&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              status: &#34;False&#34;&#xA;              type: PodReadyToStartContainers&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              status: &#34;True&#34;&#xA;              type: Initialized&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              message: &#39;containers with unready status: [csi-attacher csi-provisioner csi-snapshotter&#xA;                csi-resizer liveness-probe cinder-csi-plugin]&#39;&#xA;              reason: ContainersNotReady&#xA;              status: &#34;False&#34;&#xA;              type: Ready&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              message: &#39;containers with unready status: [csi-attacher csi-provisioner csi-snapshotter&#xA;                csi-resizer liveness-probe cinder-csi-plugin]&#39;&#xA;              reason: ContainersNotReady&#xA;              status: &#34;False&#34;&#xA;              type: ContainersReady&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:52Z&#34;&#xA;              status: &#34;True&#34;&#xA;              type: PodScheduled&#xA;            containerStatuses:&#xA;            - image: registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: cinder-csi-plugin&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /csi&#xA;                name: socket-dir&#xA;              - mountPath: /etc/kubernetes&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;              - mountPath: /etc/config&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-attacher:v4.7.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: csi-attacher&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-provisioner:v5.1.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: csi-provisioner&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-resizer:v1.12.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: csi-resizer&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: csi-snapshotter&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/livenessprobe:v2.14.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: liveness-probe&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            hostIP: 192.168.24.167&#xA;            hostIPs:&#xA;            - ip: 192.168.24.167&#xA;            phase: Pending&#xA;            qosClass: BestEffort&#xA;            startTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:232 @ 04/02/26 15:03:07.057&#xA;</failure>
              <system-err>&gt; Enter [SynchronizedBeforeSuite] TOP-LEVEL - k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:02:33.754&#xA;I0402 15:02:33.754779 24 util.go:453] &gt;&gt;&gt; kubeConfig: /tmp/kubeconfig-1277325560&#xA;I0402 15:02:33.755747 24 helper.go:51] Waiting up to 30m0s for all (but 0) nodes to be schedulable&#xA;STEP: Collecting events from namespace &#34;kube-system&#34;. - k8s.io/kubernetes/test/e2e/framework/debug/dump.go:42 @ 04/02/26 15:02:51.944&#xA;STEP: Found 331 events. - k8s.io/kubernetes/test/e2e/framework/debug/dump.go:46 @ 04/02/26 15:02:58.848&#xA;I0402 15:02:58.848279 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for calico-kube-controllers-9b54b4c6c-kzbgv: { } FailedScheduling: 0/3 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}, 2 node(s) had untolerated taint {node.cluster.x-k8s.io/uninitialized: }. preemption: 0/3 nodes are available: 3 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.848304 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for calico-kube-controllers-9b54b4c6c-kzbgv: { } Scheduled: Successfully assigned kube-system/calico-kube-controllers-9b54b4c6c-kzbgv to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.848320 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for coredns-674b8bbfcf-hdqcm: { } Scheduled: Successfully assigned kube-system/coredns-674b8bbfcf-hdqcm to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.848331 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for coredns-674b8bbfcf-hdqcm: { } FailedScheduling: 0/3 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}, 2 node(s) had untolerated taint {node.cluster.x-k8s.io/uninitialized: }. preemption: 0/3 nodes are available: 3 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.848345 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for coredns-674b8bbfcf-pkmgg: { } Scheduled: Successfully assigned kube-system/coredns-674b8bbfcf-pkmgg to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.848360 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for coredns-674b8bbfcf-pkmgg: { } FailedScheduling: 0/3 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}, 2 node(s) had untolerated taint {node.cluster.x-k8s.io/uninitialized: }. preemption: 0/3 nodes are available: 3 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.848371 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: { } FailedScheduling: 0/3 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}, 2 node(s) had untolerated taint {node.cluster.x-k8s.io/uninitialized: }. preemption: 0/3 nodes are available: 3 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.848383 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: { } Scheduled: Successfully assigned kube-system/csi-nfs-controller-7b79548448-npbhx to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.848394 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for k8s-keystone-auth-6dtgn: { } Scheduled: Successfully assigned kube-system/k8s-keystone-auth-6dtgn to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.848410 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: { } FailedScheduling: 0/3 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}, 2 node(s) had untolerated taint {node.cluster.x-k8s.io/uninitialized: }. preemption: 0/3 nodes are available: 3 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.848423 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: { } Scheduled: Successfully assigned kube-system/manila-csi-openstack-manila-csi-controllerplugin-0 to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.848434 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: { } FailedScheduling: 0/3 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}, 2 node(s) had untolerated taint {node.cluster.x-k8s.io/uninitialized: }. preemption: 0/3 nodes are available: 3 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.848447 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: { } Scheduled: Successfully assigned kube-system/openstack-cinder-csi-controllerplugin-78f966889b-plrvm to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.848457 24 dump.go:53] At 0001-01-01 00:00:00 +0000 UTC - event for openstack-cloud-controller-manager-chxrv: { } Scheduled: Successfully assigned kube-system/openstack-cloud-controller-manager-chxrv to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.848468 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for etcd-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container etcd&#xA;I0402 15:02:58.848478 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for etcd-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: etcd&#xA;I0402 15:02:58.848489 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for etcd-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/etcd:3.5.24-0&#34; already present on machine&#xA;I0402 15:02:58.848502 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for kube-apiserver-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container kube-apiserver&#xA;I0402 15:02:58.848512 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for kube-apiserver-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/kube-apiserver:v1.33.10&#34; already present on machine&#xA;I0402 15:02:58.848538 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for kube-apiserver-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: kube-apiserver&#xA;I0402 15:02:58.848548 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for kube-controller-manager-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container kube-controller-manager&#xA;I0402 15:02:58.848560 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for kube-controller-manager-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: kube-controller-manager&#xA;I0402 15:02:58.848569 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for kube-controller-manager-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/kube-controller-manager:v1.33.10&#34; already present on machine&#xA;I0402 15:02:58.848580 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for kube-scheduler-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: kube-scheduler&#xA;I0402 15:02:58.848737 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for kube-scheduler-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container kube-scheduler&#xA;I0402 15:02:58.848749 24 dump.go:53] At 2026-04-02 14:55:33 +0000 UTC - event for kube-scheduler-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/kube-scheduler:v1.33.10&#34; already present on machine&#xA;I0402 15:02:58.848757 24 dump.go:53] At 2026-04-02 14:55:40 +0000 UTC - event for kube-controller-manager: {kube-controller-manager } LeaderElection: kube-uhjhf-67cxd-gq627_0cdb0956-ef4e-4b3e-b41c-d84a983e23b5 became leader&#xA;I0402 15:02:58.848765 24 dump.go:53] At 2026-04-02 14:55:41 +0000 UTC - event for kube-scheduler: {default-scheduler } LeaderElection: kube-uhjhf-67cxd-gq627_1675f6dd-b445-4fe0-bfab-f65cfbc38932 became leader&#xA;I0402 15:02:58.848863 24 dump.go:53] At 2026-04-02 14:55:45 +0000 UTC - event for kube-controller-manager-kube-uhjhf-67cxd-gq627: {node-controller } NodeNotReady: Node is not ready&#xA;I0402 15:02:58.848872 24 dump.go:53] At 2026-04-02 14:56:26 +0000 UTC - event for openstack-cinder-csi-controllerplugin: {deployment-controller } ScalingReplicaSet: Scaled up replica set openstack-cinder-csi-controllerplugin-78f966889b from 0 to 1&#xA;I0402 15:02:58.848885 24 dump.go:53] At 2026-04-02 14:56:26 +0000 UTC - event for openstack-cinder-csi-nodeplugin: {daemonset-controller } SuccessfulCreate: Created pod: openstack-cinder-csi-nodeplugin-h74dd&#xA;I0402 15:02:58.848893 24 dump.go:53] At 2026-04-02 14:56:26 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {default-scheduler } Scheduled: Successfully assigned kube-system/openstack-cinder-csi-nodeplugin-h74dd to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.848957 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for calico-kube-controllers: {deployment-controller } ScalingReplicaSet: Scaled up replica set calico-kube-controllers-9b54b4c6c from 0 to 1&#xA;I0402 15:02:58.848968 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for calico-kube-controllers: {controllermanager } NoPods: No matching pods found&#xA;I0402 15:02:58.848975 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for calico-kube-controllers-9b54b4c6c: {replicaset-controller } SuccessfulCreate: Created pod: calico-kube-controllers-9b54b4c6c-kzbgv&#xA;I0402 15:02:58.848986 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for calico-kube-controllers-9b54b4c6c-kzbgv: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.848993 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for calico-node: {daemonset-controller } SuccessfulCreate: Created pod: calico-node-b7l94&#xA;I0402 15:02:58.849001 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for calico-node-b7l94: {default-scheduler } Scheduled: Successfully assigned kube-system/calico-node-b7l94 to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.849011 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for coredns: {deployment-controller } ScalingReplicaSet: Scaled up replica set coredns-674b8bbfcf from 0 to 2&#xA;I0402 15:02:58.849018 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for coredns-674b8bbfcf: {replicaset-controller } SuccessfulCreate: Created pod: coredns-674b8bbfcf-pkmgg&#xA;I0402 15:02:58.849026 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for coredns-674b8bbfcf: {replicaset-controller } SuccessfulCreate: Created pod: coredns-674b8bbfcf-hdqcm&#xA;I0402 15:02:58.849033 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for coredns-674b8bbfcf-hdqcm: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.849041 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for coredns-674b8bbfcf-pkmgg: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.849053 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for csi-nfs-controller: {deployment-controller } ScalingReplicaSet: Scaled up replica set csi-nfs-controller-7b79548448 from 0 to 1&#xA;I0402 15:02:58.849061 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for csi-nfs-controller-7b79548448: {replicaset-controller } FailedCreate: Error creating: pods &#34;csi-nfs-controller-7b79548448-&#34; is forbidden: error looking up service account kube-system/csi-nfs-controller-sa: serviceaccount &#34;csi-nfs-controller-sa&#34; not found&#xA;I0402 15:02:58.849069 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for csi-nfs-node: {daemonset-controller } FailedCreate: Error creating: pods &#34;csi-nfs-node-&#34; is forbidden: error looking up service account kube-system/csi-nfs-node-sa: serviceaccount &#34;csi-nfs-node-sa&#34; not found&#xA;I0402 15:02:58.849076 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for kube-proxy: {daemonset-controller } SuccessfulCreate: Created pod: kube-proxy-mx8ps&#xA;I0402 15:02:58.849085 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for kube-proxy-mx8ps: {default-scheduler } Scheduled: Successfully assigned kube-system/kube-proxy-mx8ps to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.849093 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b: {replicaset-controller } SuccessfulCreate: Created pod: openstack-cinder-csi-controllerplugin-78f966889b-plrvm&#xA;I0402 15:02:58.849102 24 dump.go:53] At 2026-04-02 14:56:27 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.849167 24 dump.go:53] At 2026-04-02 14:56:28 +0000 UTC - event for csi-nfs-controller-7b79548448: {replicaset-controller } SuccessfulCreate: Created pod: csi-nfs-controller-7b79548448-npbhx&#xA;I0402 15:02:58.849178 24 dump.go:53] At 2026-04-02 14:56:28 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.849186 24 dump.go:53] At 2026-04-02 14:56:28 +0000 UTC - event for csi-nfs-node: {daemonset-controller } SuccessfulCreate: Created pod: csi-nfs-node-vn2br&#xA;I0402 15:02:58.849194 24 dump.go:53] At 2026-04-02 14:56:28 +0000 UTC - event for csi-nfs-node-vn2br: {default-scheduler } Scheduled: Successfully assigned kube-system/csi-nfs-node-vn2br to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.849202 24 dump.go:53] At 2026-04-02 14:56:28 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin: {statefulset-controller } SuccessfulCreate: create Pod manila-csi-openstack-manila-csi-controllerplugin-0 in StatefulSet manila-csi-openstack-manila-csi-controllerplugin successful&#xA;I0402 15:02:58.849211 24 dump.go:53] At 2026-04-02 14:56:28 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.849219 24 dump.go:53] At 2026-04-02 14:56:28 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin: {daemonset-controller } SuccessfulCreate: Created pod: manila-csi-openstack-manila-csi-nodeplugin-pnrtc&#xA;I0402 15:02:58.849227 24 dump.go:53] At 2026-04-02 14:56:28 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-pnrtc: {default-scheduler } Scheduled: Successfully assigned kube-system/manila-csi-openstack-manila-csi-nodeplugin-pnrtc to kube-uhjhf-67cxd-gq627&#xA;I0402 15:02:58.849235 24 dump.go:53] At 2026-04-02 14:56:29 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;quay.io/calico/cni:v3.31.3&#34;&#xA;I0402 15:02:58.849245 24 dump.go:53] At 2026-04-02 14:56:29 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.8.0&#34;&#xA;I0402 15:02:58.849252 24 dump.go:53] At 2026-04-02 14:56:29 +0000 UTC - event for kube-proxy-mx8ps: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: kube-proxy&#xA;I0402 15:02:58.849260 24 dump.go:53] At 2026-04-02 14:56:29 +0000 UTC - event for kube-proxy-mx8ps: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container kube-proxy&#xA;I0402 15:02:58.849268 24 dump.go:53] At 2026-04-02 14:56:29 +0000 UTC - event for kube-proxy-mx8ps: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/kube-proxy:v1.33.10&#34; already present on machine&#xA;I0402 15:02:58.849276 24 dump.go:53] At 2026-04-02 14:56:29 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-pnrtc: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34;&#xA;I0402 15:02:58.849284 24 dump.go:53] At 2026-04-02 14:56:29 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34;&#xA;I0402 15:02:58.849291 24 dump.go:53] At 2026-04-02 14:56:30 +0000 UTC - event for kube-apiserver-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Killing: Stopping container kube-apiserver&#xA;I0402 15:02:58.849354 24 dump.go:53] At 2026-04-02 14:56:34 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;quay.io/calico/cni:v3.31.3&#34; in 4.865s (4.866s including waiting). Image size: 72141081 bytes.&#xA;I0402 15:02:58.849367 24 dump.go:53] At 2026-04-02 14:56:34 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: upgrade-ipam&#xA;I0402 15:02:58.849377 24 dump.go:53] At 2026-04-02 14:56:34 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container upgrade-ipam&#xA;I0402 15:02:58.849384 24 dump.go:53] At 2026-04-02 14:56:36 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;quay.io/calico/cni:v3.31.3&#34; already present on machine&#xA;I0402 15:02:58.849392 24 dump.go:53] At 2026-04-02 14:56:37 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container install-cni&#xA;I0402 15:02:58.849471 24 dump.go:53] At 2026-04-02 14:56:37 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: install-cni&#xA;I0402 15:02:58.849480 24 dump.go:53] At 2026-04-02 14:56:39 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container node-driver-registrar&#xA;I0402 15:02:58.849488 24 dump.go:53] At 2026-04-02 14:56:39 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: node-driver-registrar&#xA;I0402 15:02:58.849496 24 dump.go:53] At 2026-04-02 14:56:39 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34; in 5.648s (10.507s including waiting). Image size: 14038309 bytes.&#xA;I0402 15:02:58.849505 24 dump.go:53] At 2026-04-02 14:56:39 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.14.0&#34;&#xA;I0402 15:02:58.849566 24 dump.go:53] At 2026-04-02 14:56:41 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.2&#34;&#xA;I0402 15:02:58.849576 24 dump.go:53] At 2026-04-02 14:56:41 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: liveness-probe&#xA;I0402 15:02:58.849584 24 dump.go:53] At 2026-04-02 14:56:41 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.8.0&#34; in 1.698s (12.096s including waiting). Image size: 8892463 bytes.&#xA;I0402 15:02:58.849592 24 dump.go:53] At 2026-04-02 14:56:41 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container liveness-probe&#xA;I0402 15:02:58.849601 24 dump.go:53] At 2026-04-02 14:56:41 +0000 UTC - event for kube-scheduler-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Liveness probe failed: Get &#34;https://:10259/livez&#34;: dial tcp :10259: connect: connection refused&#xA;I0402 15:02:58.849610 24 dump.go:53] At 2026-04-02 14:56:42 +0000 UTC - event for kube-scheduler-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Readiness probe failed: Get &#34;https://:10259/readyz&#34;: dial tcp :10259: connect: connection refused&#xA;I0402 15:02:58.849618 24 dump.go:53] At 2026-04-02 14:56:42 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-pnrtc: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34; in 1.141s (13.213s including waiting). Image size: 14038309 bytes.&#xA;I0402 15:02:58.849626 24 dump.go:53] At 2026-04-02 14:56:44 +0000 UTC - event for kube-controller-manager-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/kube-controller-manager:v1.33.10&#34; already present on machine&#xA;I0402 15:02:58.849633 24 dump.go:53] At 2026-04-02 14:56:44 +0000 UTC - event for kube-scheduler-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/kube-scheduler:v1.33.10&#34; already present on machine&#xA;I0402 15:02:58.849641 24 dump.go:53] At 2026-04-02 14:56:45 +0000 UTC - event for kube-controller-manager-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container kube-controller-manager&#xA;I0402 15:02:58.849649 24 dump.go:53] At 2026-04-02 14:56:45 +0000 UTC - event for kube-controller-manager-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: kube-controller-manager&#xA;I0402 15:02:58.849657 24 dump.go:53] At 2026-04-02 14:56:45 +0000 UTC - event for kube-scheduler-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: kube-scheduler&#xA;I0402 15:02:58.849664 24 dump.go:53] At 2026-04-02 14:56:45 +0000 UTC - event for kube-scheduler-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container kube-scheduler&#xA;I0402 15:02:58.849672 24 dump.go:53] At 2026-04-02 14:56:45 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-pnrtc: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#34;&#xA;I0402 15:02:58.849688 24 dump.go:53] At 2026-04-02 14:56:45 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-pnrtc: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: nfs-registrar&#xA;I0402 15:02:58.849696 24 dump.go:53] At 2026-04-02 14:56:45 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-pnrtc: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container nfs-registrar&#xA;I0402 15:02:58.849704 24 dump.go:53] At 2026-04-02 14:56:46 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.14.0&#34; in 4.307s (6.979s including waiting). Image size: 14311007 bytes.&#xA;I0402 15:02:58.849712 24 dump.go:53] At 2026-04-02 14:56:46 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: liveness-probe&#xA;I0402 15:02:58.849722 24 dump.go:53] At 2026-04-02 14:56:47 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0&#34;&#xA;I0402 15:02:58.849729 24 dump.go:53] At 2026-04-02 14:56:47 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container liveness-probe&#xA;I0402 15:02:58.849737 24 dump.go:53] At 2026-04-02 14:56:48 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/nfsplugin:v4.2.0&#34;&#xA;I0402 15:02:58.849745 24 dump.go:53] At 2026-04-02 14:56:48 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.2&#34; in 1.646s (6.838s including waiting). Image size: 10074312 bytes.&#xA;I0402 15:02:58.849753 24 dump.go:53] At 2026-04-02 14:56:48 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container node-driver-registrar&#xA;I0402 15:02:58.849761 24 dump.go:53] At 2026-04-02 14:56:48 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: node-driver-registrar&#xA;I0402 15:02:58.849801 24 dump.go:53] At 2026-04-02 14:56:50 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-pnrtc: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: nfs-nodeplugin&#xA;I0402 15:02:58.849812 24 dump.go:53] At 2026-04-02 14:56:50 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-pnrtc: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#34; in 2.421s (5.604s including waiting). Image size: 21014294 bytes.&#xA;I0402 15:02:58.849820 24 dump.go:53] At 2026-04-02 14:56:51 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-pnrtc: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container nfs-nodeplugin&#xA;I0402 15:02:58.849828 24 dump.go:53] At 2026-04-02 14:56:52 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-pnrtc: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#34; already present on machine&#xA;I0402 15:02:58.849836 24 dump.go:53] At 2026-04-02 14:56:53 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-pnrtc: {kubelet kube-uhjhf-67cxd-gq627} BackOff: Back-off restarting failed container nfs-nodeplugin in pod manila-csi-openstack-manila-csi-nodeplugin-pnrtc_kube-system(522358fc-5b98-45f7-9888-54e2c821b8d8)&#xA;I0402 15:02:58.849844 24 dump.go:53] At 2026-04-02 14:56:55 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0&#34; in 4.147s (8.044s including waiting). Image size: 28718193 bytes.&#xA;I0402 15:02:58.849853 24 dump.go:53] At 2026-04-02 14:56:55 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container cinder-csi-plugin&#xA;I0402 15:02:58.849862 24 dump.go:53] At 2026-04-02 14:56:55 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: cinder-csi-plugin&#xA;I0402 15:02:58.849871 24 dump.go:53] At 2026-04-02 14:56:59 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34; already present on machine&#xA;I0402 15:02:58.849879 24 dump.go:53] At 2026-04-02 14:57:01 +0000 UTC - event for kube-scheduler-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Readiness probe failed: HTTP probe failed with statuscode: 500&#xA;I0402 15:02:58.849888 24 dump.go:53] At 2026-04-02 14:57:04 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container nfs&#xA;I0402 15:02:58.849896 24 dump.go:53] At 2026-04-02 14:57:04 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/nfsplugin:v4.2.0&#34; in 9.455s (15.875s including waiting). Image size: 49783016 bytes.&#xA;I0402 15:02:58.849903 24 dump.go:53] At 2026-04-02 14:57:04 +0000 UTC - event for csi-nfs-node-vn2br: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: nfs&#xA;I0402 15:02:58.849911 24 dump.go:53] At 2026-04-02 14:57:04 +0000 UTC - event for openstack-cinder-csi-nodeplugin-h74dd: {kubelet kube-uhjhf-67cxd-gq627} BackOff: Back-off restarting failed container node-driver-registrar in pod openstack-cinder-csi-nodeplugin-h74dd_kube-system(0988c0c0-8dae-4f7f-a1f4-f4c1614e0431)&#xA;I0402 15:02:58.849919 24 dump.go:53] At 2026-04-02 14:57:05 +0000 UTC - event for kube-apiserver-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container kube-apiserver&#xA;I0402 15:02:58.849927 24 dump.go:53] At 2026-04-02 14:57:05 +0000 UTC - event for kube-apiserver-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/kube-apiserver:v1.33.10&#34; already present on machine&#xA;I0402 15:02:58.849938 24 dump.go:53] At 2026-04-02 14:57:05 +0000 UTC - event for kube-apiserver-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: kube-apiserver&#xA;I0402 15:02:58.849945 24 dump.go:53] At 2026-04-02 14:57:26 +0000 UTC - event for kube-controller-manager: {kube-controller-manager } LeaderElection: kube-uhjhf-67cxd-gq627_c45b01b3-4174-44ef-9474-12078fd913b8 became leader&#xA;I0402 15:02:58.849955 24 dump.go:53] At 2026-04-02 14:57:42 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} BackOff: Back-off restarting failed container install-cni in pod calico-node-b7l94_kube-system(4bceced1-259d-4db5-a5ed-74084e0c715b)&#xA;I0402 15:02:58.849962 24 dump.go:53] At 2026-04-02 14:57:50 +0000 UTC - event for calico-kube-controllers-9b54b4c6c-kzbgv: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.849970 24 dump.go:53] At 2026-04-02 14:57:50 +0000 UTC - event for coredns-674b8bbfcf-hdqcm: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.849978 24 dump.go:53] At 2026-04-02 14:57:50 +0000 UTC - event for coredns-674b8bbfcf-pkmgg: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.849986 24 dump.go:53] At 2026-04-02 14:57:50 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.849994 24 dump.go:53] At 2026-04-02 14:57:50 +0000 UTC - event for kube-scheduler: {default-scheduler } LeaderElection: kube-uhjhf-67cxd-gq627_d8be3d1f-0899-4585-a3c1-5904735b2497 became leader&#xA;I0402 15:02:58.850002 24 dump.go:53] At 2026-04-02 14:57:50 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.850009 24 dump.go:53] At 2026-04-02 14:57:50 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 node(s) had untolerated taint {node.cloudprovider.kubernetes.io/uninitialized: true}. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.&#xA;I0402 15:02:58.850018 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for calico-node: {daemonset-controller } SuccessfulCreate: Created pod: calico-node-6wvr4&#xA;I0402 15:02:58.850026 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for calico-node: {daemonset-controller } SuccessfulCreate: Created pod: calico-node-7pj2s&#xA;I0402 15:02:58.850035 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for calico-node-6wvr4: {default-scheduler } Scheduled: Successfully assigned kube-system/calico-node-6wvr4 to kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb&#xA;I0402 15:02:58.850043 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for calico-node-7pj2s: {default-scheduler } Scheduled: Successfully assigned kube-system/calico-node-7pj2s to kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67&#xA;I0402 15:02:58.850051 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for csi-nfs-node: {daemonset-controller } SuccessfulCreate: Created pod: csi-nfs-node-4xmg9&#xA;I0402 15:02:58.850058 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for csi-nfs-node: {daemonset-controller } SuccessfulCreate: Created pod: csi-nfs-node-jtj8j&#xA;I0402 15:02:58.850067 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for csi-nfs-node-4xmg9: {default-scheduler } Scheduled: Successfully assigned kube-system/csi-nfs-node-4xmg9 to kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67&#xA;I0402 15:02:58.850075 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for csi-nfs-node-jtj8j: {default-scheduler } Scheduled: Successfully assigned kube-system/csi-nfs-node-jtj8j to kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb&#xA;I0402 15:02:58.850085 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for kube-proxy: {daemonset-controller } SuccessfulCreate: Created pod: kube-proxy-8pc5r&#xA;I0402 15:02:58.850092 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for kube-proxy: {daemonset-controller } SuccessfulCreate: Created pod: kube-proxy-djtqn&#xA;I0402 15:02:58.850100 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for kube-proxy-8pc5r: {default-scheduler } Scheduled: Successfully assigned kube-system/kube-proxy-8pc5r to kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb&#xA;I0402 15:02:58.850108 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for kube-proxy-djtqn: {default-scheduler } Scheduled: Successfully assigned kube-system/kube-proxy-djtqn to kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67&#xA;I0402 15:02:58.850115 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin: {daemonset-controller } SuccessfulCreate: Created pod: manila-csi-openstack-manila-csi-nodeplugin-66z9w&#xA;I0402 15:02:58.850123 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin: {daemonset-controller } SuccessfulCreate: Created pod: manila-csi-openstack-manila-csi-nodeplugin-mnmw8&#xA;I0402 15:02:58.850131 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {default-scheduler } Scheduled: Successfully assigned kube-system/manila-csi-openstack-manila-csi-nodeplugin-mnmw8 to kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67&#xA;I0402 15:02:58.850141 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for openstack-cinder-csi-nodeplugin: {daemonset-controller } SuccessfulCreate: Created pod: openstack-cinder-csi-nodeplugin-lnmpk&#xA;I0402 15:02:58.850150 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for openstack-cinder-csi-nodeplugin: {daemonset-controller } SuccessfulCreate: Created pod: openstack-cinder-csi-nodeplugin-z67qx&#xA;I0402 15:02:58.850158 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {default-scheduler } Scheduled: Successfully assigned kube-system/openstack-cinder-csi-nodeplugin-lnmpk to kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67&#xA;I0402 15:02:58.850167 24 dump.go:53] At 2026-04-02 14:57:58 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {default-scheduler } Scheduled: Successfully assigned kube-system/openstack-cinder-csi-nodeplugin-z67qx to kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb&#xA;I0402 15:02:58.850174 24 dump.go:53] At 2026-04-02 14:57:59 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {default-scheduler } Scheduled: Successfully assigned kube-system/manila-csi-openstack-manila-csi-nodeplugin-66z9w to kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb&#xA;I0402 15:02:58.850183 24 dump.go:53] At 2026-04-02 14:58:09 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulling: Pulling image &#34;quay.io/calico/cni:v3.31.3&#34;&#xA;I0402 15:02:58.850191 24 dump.go:53] At 2026-04-02 14:58:09 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.8.0&#34;&#xA;I0402 15:02:58.850198 24 dump.go:53] At 2026-04-02 14:58:09 +0000 UTC - event for kube-proxy-djtqn: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: kube-proxy&#xA;I0402 15:02:58.850207 24 dump.go:53] At 2026-04-02 14:58:09 +0000 UTC - event for kube-proxy-djtqn: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container kube-proxy&#xA;I0402 15:02:58.850214 24 dump.go:53] At 2026-04-02 14:58:09 +0000 UTC - event for kube-proxy-djtqn: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Container image &#34;registry.k8s.io/kube-proxy:v1.33.10&#34; already present on machine&#xA;I0402 15:02:58.850222 24 dump.go:53] At 2026-04-02 14:58:09 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34;&#xA;I0402 15:02:58.850231 24 dump.go:53] At 2026-04-02 14:58:09 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34;&#xA;I0402 15:02:58.850239 24 dump.go:53] At 2026-04-02 14:58:10 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulling: Pulling image &#34;quay.io/calico/cni:v3.31.3&#34;&#xA;I0402 15:02:58.850247 24 dump.go:53] At 2026-04-02 14:58:10 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.8.0&#34;&#xA;I0402 15:02:58.850256 24 dump.go:53] At 2026-04-02 14:58:10 +0000 UTC - event for kube-proxy-8pc5r: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container kube-proxy&#xA;I0402 15:02:58.850266 24 dump.go:53] At 2026-04-02 14:58:10 +0000 UTC - event for kube-proxy-8pc5r: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: kube-proxy&#xA;I0402 15:02:58.850273 24 dump.go:53] At 2026-04-02 14:58:10 +0000 UTC - event for kube-proxy-8pc5r: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Container image &#34;registry.k8s.io/kube-proxy:v1.33.10&#34; already present on machine&#xA;I0402 15:02:58.850283 24 dump.go:53] At 2026-04-02 14:58:10 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34;&#xA;I0402 15:02:58.850290 24 dump.go:53] At 2026-04-02 14:58:10 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34;&#xA;I0402 15:02:58.850298 24 dump.go:53] At 2026-04-02 14:58:11 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.8.0&#34; in 1.764s (1.764s including waiting). Image size: 8892463 bytes.&#xA;I0402 15:02:58.850306 24 dump.go:53] At 2026-04-02 14:58:11 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container liveness-probe&#xA;I0402 15:02:58.850314 24 dump.go:53] At 2026-04-02 14:58:11 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.2&#34;&#xA;I0402 15:02:58.850321 24 dump.go:53] At 2026-04-02 14:58:11 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: liveness-probe&#xA;I0402 15:02:58.850329 24 dump.go:53] At 2026-04-02 14:58:12 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: nfs-registrar&#xA;I0402 15:02:58.850336 24 dump.go:53] At 2026-04-02 14:58:12 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34; in 2.574s (2.574s including waiting). Image size: 14038309 bytes.&#xA;I0402 15:02:58.850345 24 dump.go:53] At 2026-04-02 14:58:12 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulling: Pulling image &#34;registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#34;&#xA;I0402 15:02:58.850353 24 dump.go:53] At 2026-04-02 14:58:12 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container nfs-registrar&#xA;I0402 15:02:58.850363 24 dump.go:53] At 2026-04-02 14:58:15 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container upgrade-ipam&#xA;I0402 15:02:58.850371 24 dump.go:53] At 2026-04-02 14:58:15 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: upgrade-ipam&#xA;I0402 15:02:58.850379 24 dump.go:53] At 2026-04-02 14:58:15 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Successfully pulled image &#34;quay.io/calico/cni:v3.31.3&#34; in 4.701s (6.405s including waiting). Image size: 72141081 bytes.&#xA;I0402 15:02:58.850387 24 dump.go:53] At 2026-04-02 14:58:15 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.8.0&#34; in 2.571s (5.109s including waiting). Image size: 8892463 bytes.&#xA;I0402 15:02:58.850396 24 dump.go:53] At 2026-04-02 14:58:15 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: liveness-probe&#xA;I0402 15:02:58.850404 24 dump.go:53] At 2026-04-02 14:58:15 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container liveness-probe&#xA;I0402 15:02:58.850412 24 dump.go:53] At 2026-04-02 14:58:15 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.2&#34;&#xA;I0402 15:02:58.850419 24 dump.go:53] At 2026-04-02 14:58:18 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: install-cni&#xA;I0402 15:02:58.850427 24 dump.go:53] At 2026-04-02 14:58:18 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Container image &#34;quay.io/calico/cni:v3.31.3&#34; already present on machine&#xA;I0402 15:02:58.850436 24 dump.go:53] At 2026-04-02 14:58:18 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container install-cni&#xA;I0402 15:02:58.850444 24 dump.go:53] At 2026-04-02 14:58:22 +0000 UTC - event for kube-apiserver-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Readiness probe failed: HTTP probe failed with statuscode: 500&#xA;I0402 15:02:58.850452 24 dump.go:53] At 2026-04-02 14:58:31 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Successfully pulled image &#34;quay.io/calico/cni:v3.31.3&#34; in 16.6s (21.648s including waiting). Image size: 72141081 bytes.&#xA;I0402 15:02:58.850460 24 dump.go:53] At 2026-04-02 14:58:32 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container upgrade-ipam&#xA;I0402 15:02:58.850468 24 dump.go:53] At 2026-04-02 14:58:32 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: upgrade-ipam&#xA;I0402 15:02:58.850475 24 dump.go:53] At 2026-04-02 14:58:33 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34; in 1.57s (23.21s including waiting). Image size: 14038309 bytes.&#xA;I0402 15:02:58.850483 24 dump.go:53] At 2026-04-02 14:58:37 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34; in 22.188s (28.564s including waiting). Image size: 14038309 bytes.&#xA;I0402 15:02:58.850491 24 dump.go:53] At 2026-04-02 14:58:37 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: node-driver-registrar&#xA;I0402 15:02:58.850498 24 dump.go:53] At 2026-04-02 14:58:38 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container install-cni&#xA;I0402 15:02:58.850506 24 dump.go:53] At 2026-04-02 14:58:38 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Container image &#34;quay.io/calico/cni:v3.31.3&#34; already present on machine&#xA;I0402 15:02:58.850515 24 dump.go:53] At 2026-04-02 14:58:38 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: install-cni&#xA;I0402 15:02:58.850525 24 dump.go:53] At 2026-04-02 14:58:38 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34; in 331ms (28.875s including waiting). Image size: 14038309 bytes.&#xA;I0402 15:02:58.850532 24 dump.go:53] At 2026-04-02 14:58:38 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container node-driver-registrar&#xA;I0402 15:02:58.850541 24 dump.go:53] At 2026-04-02 14:58:38 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.14.0&#34;&#xA;I0402 15:02:58.850550 24 dump.go:53] At 2026-04-02 14:58:38 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: node-driver-registrar&#xA;I0402 15:02:58.850557 24 dump.go:53] At 2026-04-02 14:58:38 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.14.0&#34;&#xA;I0402 15:02:58.850565 24 dump.go:53] At 2026-04-02 14:58:38 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container node-driver-registrar&#xA;I0402 15:02:58.850574 24 dump.go:53] At 2026-04-02 14:58:39 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container nfs-registrar&#xA;I0402 15:02:58.850583 24 dump.go:53] At 2026-04-02 14:58:39 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulling: Pulling image &#34;registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#34;&#xA;I0402 15:02:58.850593 24 dump.go:53] At 2026-04-02 14:58:39 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: nfs-registrar&#xA;I0402 15:02:58.850600 24 dump.go:53] At 2026-04-02 14:58:57 +0000 UTC - event for kube-apiserver-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Liveness probe failed: HTTP probe failed with statuscode: 500&#xA;I0402 15:02:58.851023 24 dump.go:53] At 2026-04-02 14:59:08 +0000 UTC - event for kube-controller-manager-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Liveness probe failed: Get &#34;https://:10257/healthz&#34;: dial tcp :10257: connect: connection refused&#xA;I0402 15:02:58.851033 24 dump.go:53] At 2026-04-02 14:59:20 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.2&#34; in 41.916s (1m8.97s including waiting). Image size: 10074312 bytes.&#xA;I0402 15:02:58.851043 24 dump.go:53] At 2026-04-02 14:59:21 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;quay.io/calico/node:v3.31.3&#34;&#xA;I0402 15:02:58.851051 24 dump.go:53] At 2026-04-02 14:59:26 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulling: Pulling image &#34;quay.io/calico/node:v3.31.3&#34;&#xA;I0402 15:02:58.851059 24 dump.go:53] At 2026-04-02 14:59:26 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Successfully pulled image &#34;registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#34; in 52.86s (1m13.452s including waiting). Image size: 21014294 bytes.&#xA;I0402 15:02:58.851178 24 dump.go:53] At 2026-04-02 14:59:39 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulling: Pulling image &#34;quay.io/calico/node:v3.31.3&#34;&#xA;I0402 15:02:58.851191 24 dump.go:53] At 2026-04-02 14:59:40 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/nfsplugin:v4.2.0&#34;&#xA;I0402 15:02:58.851199 24 dump.go:53] At 2026-04-02 14:59:40 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container node-driver-registrar&#xA;I0402 15:02:58.851209 24 dump.go:53] At 2026-04-02 14:59:40 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: node-driver-registrar&#xA;I0402 15:02:58.851217 24 dump.go:53] At 2026-04-02 14:59:41 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: nfs-nodeplugin&#xA;I0402 15:02:58.851224 24 dump.go:53] At 2026-04-02 14:59:42 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container nfs-nodeplugin&#xA;I0402 15:02:58.851232 24 dump.go:53] At 2026-04-02 14:59:43 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Container image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34; already present on machine&#xA;I0402 15:02:58.851240 24 dump.go:53] At 2026-04-02 14:59:43 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Container image &#34;registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#34; already present on machine&#xA;I0402 15:02:58.851247 24 dump.go:53] At 2026-04-02 14:59:44 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} BackOff: Back-off restarting failed container nfs-nodeplugin in pod manila-csi-openstack-manila-csi-nodeplugin-66z9w_kube-system(e6c1c2e9-d6e6-4de1-904a-ef13ae6070b7)&#xA;I0402 15:02:58.851255 24 dump.go:53] At 2026-04-02 14:59:45 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: liveness-probe&#xA;I0402 15:02:58.851264 24 dump.go:53] At 2026-04-02 14:59:45 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container liveness-probe&#xA;I0402 15:02:58.851271 24 dump.go:53] At 2026-04-02 14:59:45 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.14.0&#34; in 25.441s (1m7.556s including waiting). Image size: 14311007 bytes.&#xA;I0402 15:02:58.851281 24 dump.go:53] At 2026-04-02 14:59:45 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulling: Pulling image &#34;registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0&#34;&#xA;I0402 15:02:58.851289 24 dump.go:53] At 2026-04-02 14:59:47 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.2&#34; in 20.995s (1m31.882s including waiting). Image size: 10074312 bytes.&#xA;I0402 15:02:58.851297 24 dump.go:53] At 2026-04-02 14:59:47 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: node-driver-registrar&#xA;I0402 15:02:58.851305 24 dump.go:53] At 2026-04-02 14:59:47 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/nfsplugin:v4.2.0&#34;&#xA;I0402 15:02:58.851313 24 dump.go:53] At 2026-04-02 14:59:47 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container node-driver-registrar&#xA;I0402 15:02:58.851321 24 dump.go:53] At 2026-04-02 14:59:58 +0000 UTC - event for kube-controller-manager: {kube-controller-manager } LeaderElection: kube-uhjhf-67cxd-gq627_1eb2d9ed-e898-4289-bd0f-d705aac6a765 became leader&#xA;I0402 15:02:58.851329 24 dump.go:53] At 2026-04-02 15:00:00 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: nfs-nodeplugin&#xA;I0402 15:02:58.851337 24 dump.go:53] At 2026-04-02 15:00:00 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Successfully pulled image &#34;registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#34; in 14.623s (1m21.042s including waiting). Image size: 21014294 bytes.&#xA;I0402 15:02:58.851345 24 dump.go:53] At 2026-04-02 15:00:00 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container nfs-nodeplugin&#xA;I0402 15:02:58.851355 24 dump.go:53] At 2026-04-02 15:00:00 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.14.0&#34; in 13.561s (1m22.855s including waiting). Image size: 14311007 bytes.&#xA;I0402 15:02:58.851363 24 dump.go:53] At 2026-04-02 15:00:00 +0000 UTC - event for openstack-cloud-controller-manager: {daemonset-controller } SuccessfulCreate: Created pod: openstack-cloud-controller-manager-chxrv&#xA;I0402 15:02:58.851371 24 dump.go:53] At 2026-04-02 15:00:01 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Container image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34; already present on machine&#xA;I0402 15:02:58.851378 24 dump.go:53] At 2026-04-02 15:00:02 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;quay.io/calico/node:v3.31.3&#34; in 41.417s (41.417s including waiting). Image size: 159483847 bytes.&#xA;I0402 15:02:58.851386 24 dump.go:53] At 2026-04-02 15:00:02 +0000 UTC - event for kube-scheduler: {default-scheduler } LeaderElection: kube-uhjhf-67cxd-gq627_f1c8a633-4510-47ec-ad39-bd3e0ead2f22 became leader&#xA;I0402 15:02:58.851394 24 dump.go:53] At 2026-04-02 15:00:03 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: ebpf-bootstrap&#xA;I0402 15:02:58.851403 24 dump.go:53] At 2026-04-02 15:00:03 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container ebpf-bootstrap&#xA;I0402 15:02:58.851412 24 dump.go:53] At 2026-04-02 15:00:03 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulling: Pulling image &#34;registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0&#34;&#xA;I0402 15:02:58.851420 24 dump.go:53] At 2026-04-02 15:00:03 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container liveness-probe&#xA;I0402 15:02:58.851429 24 dump.go:53] At 2026-04-02 15:00:03 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: liveness-probe&#xA;I0402 15:02:58.851437 24 dump.go:53] At 2026-04-02 15:00:04 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Container image &#34;registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#34; already present on machine&#xA;I0402 15:02:58.851446 24 dump.go:53] At 2026-04-02 15:00:09 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} BackOff: Back-off restarting failed container nfs-nodeplugin in pod manila-csi-openstack-manila-csi-nodeplugin-mnmw8_kube-system(a65eb016-b42a-43be-b7f9-8beb8b10c898)&#xA;I0402 15:02:58.851453 24 dump.go:53] At 2026-04-02 15:00:10 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container calico-node&#xA;I0402 15:02:58.851461 24 dump.go:53] At 2026-04-02 15:00:10 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;quay.io/calico/node:v3.31.3&#34; already present on machine&#xA;I0402 15:02:58.851469 24 dump.go:53] At 2026-04-02 15:00:10 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: calico-node&#xA;I0402 15:02:58.851477 24 dump.go:53] At 2026-04-02 15:00:10 +0000 UTC - event for openstack-cloud-controller-manager-chxrv: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/provider-os/openstack-cloud-controller-manager:v1.33.1&#34;&#xA;I0402 15:02:58.851485 24 dump.go:53] At 2026-04-02 15:00:11 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Readiness probe failed: calico/node is not ready: BIRD is not ready: error querying BIRD: unable to connect to BIRDv4 socket: dial unix /var/run/bird/bird.ctl: connect: no such file or directory&#xA;&#xA;I0402 15:02:58.851494 24 dump.go:53] At 2026-04-02 15:00:12 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Readiness probe failed: calico/node is not ready: BIRD is not ready: error querying BIRD: unable to connect to BIRDv4 socket: dial unix /var/run/calico/bird.ctl: connect: connection refused&#xA;&#xA;I0402 15:02:58.851502 24 dump.go:53] At 2026-04-02 15:00:17 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-66z9w: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} BackOff: Back-off restarting failed container nfs-registrar in pod manila-csi-openstack-manila-csi-nodeplugin-66z9w_kube-system(e6c1c2e9-d6e6-4de1-904a-ef13ae6070b7)&#xA;I0402 15:02:58.851511 24 dump.go:53] At 2026-04-02 15:00:24 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Successfully pulled image &#34;quay.io/calico/node:v3.31.3&#34; in 23.921s (44.931s including waiting). Image size: 159483847 bytes.&#xA;I0402 15:02:58.851519 24 dump.go:53] At 2026-04-02 15:00:24 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Successfully pulled image &#34;quay.io/calico/node:v3.31.3&#34; in 24.548s (57.895s including waiting). Image size: 159483847 bytes.&#xA;I0402 15:02:58.851527 24 dump.go:53] At 2026-04-02 15:00:25 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: ebpf-bootstrap&#xA;I0402 15:02:58.851535 24 dump.go:53] At 2026-04-02 15:00:25 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container ebpf-bootstrap&#xA;I0402 15:02:58.851544 24 dump.go:53] At 2026-04-02 15:00:25 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: ebpf-bootstrap&#xA;I0402 15:02:58.851552 24 dump.go:53] At 2026-04-02 15:00:32 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container ebpf-bootstrap&#xA;I0402 15:02:58.851560 24 dump.go:53] At 2026-04-02 15:00:46 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Container image &#34;quay.io/calico/node:v3.31.3&#34; already present on machine&#xA;I0402 15:02:58.851568 24 dump.go:53] At 2026-04-02 15:00:47 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Container image &#34;quay.io/calico/node:v3.31.3&#34; already present on machine&#xA;I0402 15:02:58.851576 24 dump.go:53] At 2026-04-02 15:00:47 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: calico-node&#xA;I0402 15:02:58.851585 24 dump.go:53] At 2026-04-02 15:00:48 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: calico-node&#xA;I0402 15:02:58.851592 24 dump.go:53] At 2026-04-02 15:00:48 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container calico-node&#xA;I0402 15:02:58.851601 24 dump.go:53] At 2026-04-02 15:00:48 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container calico-node&#xA;I0402 15:02:58.851609 24 dump.go:53] At 2026-04-02 15:00:48 +0000 UTC - event for openstack-cloud-controller-manager-chxrv: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container openstack-cloud-controller-manager&#xA;I0402 15:02:58.851618 24 dump.go:53] At 2026-04-02 15:00:48 +0000 UTC - event for openstack-cloud-controller-manager-chxrv: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: openstack-cloud-controller-manager&#xA;I0402 15:02:58.851626 24 dump.go:53] At 2026-04-02 15:00:48 +0000 UTC - event for openstack-cloud-controller-manager-chxrv: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/provider-os/openstack-cloud-controller-manager:v1.33.1&#34; in 38.396s (38.396s including waiting). Image size: 21265731 bytes.&#xA;I0402 15:02:58.851635 24 dump.go:53] At 2026-04-02 15:00:49 +0000 UTC - event for calico-node-6wvr4: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Unhealthy: Readiness probe failed: calico/node is not ready: BIRD is not ready: error querying BIRD: unable to connect to BIRDv4 socket: dial unix /var/run/bird/bird.ctl: connect: no such file or directory&#xA;&#xA;I0402 15:02:58.851644 24 dump.go:53] At 2026-04-02 15:00:49 +0000 UTC - event for calico-node-7pj2s: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Unhealthy: Readiness probe failed: calico/node is not ready: BIRD is not ready: error querying BIRD: unable to connect to BIRDv4 socket: dial unix /var/run/bird/bird.ctl: connect: no such file or directory&#xA;&#xA;I0402 15:02:58.851652 24 dump.go:53] At 2026-04-02 15:00:52 +0000 UTC - event for calico-node-b7l94: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Readiness probe failed: 2026-04-02 15:00:52.521 [INFO][388] node/health.go 206: Number of node(s) with BGP peering established = 0&#xA;calico/node is not ready: BIRD is not ready: BGP not established with 192.168.24.125,192.168.24.48&#xA;&#xA;I0402 15:02:58.851660 24 dump.go:53] At 2026-04-02 15:00:52 +0000 UTC - event for cloud-controller-manager: {cloud-controller-manager } LeaderElection: kube-uhjhf-67cxd-gq627_45ace113-f2e9-456b-85c9-66189fe98ad6 became leader&#xA;I0402 15:02:58.851670 24 dump.go:53] At 2026-04-02 15:00:52 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/nfsplugin:v4.2.0&#34; in 28.148s (1m5.434s including waiting). Image size: 49783016 bytes.&#xA;I0402 15:02:58.851679 24 dump.go:53] At 2026-04-02 15:00:53 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/nfsplugin:v4.2.0&#34; in 28.843s (1m12.652s including waiting). Image size: 49783016 bytes.&#xA;I0402 15:02:58.851687 24 dump.go:53] At 2026-04-02 15:00:53 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: nfs&#xA;I0402 15:02:58.851694 24 dump.go:53] At 2026-04-02 15:00:53 +0000 UTC - event for csi-nfs-node-4xmg9: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container nfs&#xA;I0402 15:02:58.851703 24 dump.go:53] At 2026-04-02 15:00:53 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: nfs&#xA;I0402 15:02:58.851713 24 dump.go:53] At 2026-04-02 15:00:53 +0000 UTC - event for csi-nfs-node-jtj8j: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container nfs&#xA;I0402 15:02:58.851721 24 dump.go:53] At 2026-04-02 15:00:53 +0000 UTC - event for k8s-keystone-auth: {daemonset-controller } SuccessfulCreate: Created pod: k8s-keystone-auth-6dtgn&#xA;I0402 15:02:58.851729 24 dump.go:53] At 2026-04-02 15:00:54 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-provisioner:v3.3.0&#34;&#xA;I0402 15:02:58.851736 24 dump.go:53] At 2026-04-02 15:00:55 +0000 UTC - event for k8s-keystone-auth-6dtgn: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/provider-os/k8s-keystone-auth:v1.29.0&#34;&#xA;I0402 15:02:58.851744 24 dump.go:53] At 2026-04-02 15:00:56 +0000 UTC - event for coredns-674b8bbfcf-hdqcm: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/coredns/coredns:v1.12.0&#34; already present on machine&#xA;I0402 15:02:58.851753 24 dump.go:53] At 2026-04-02 15:00:58 +0000 UTC - event for coredns-674b8bbfcf-hdqcm: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: coredns&#xA;I0402 15:02:58.851761 24 dump.go:53] At 2026-04-02 15:00:59 +0000 UTC - event for calico-kube-controllers-9b54b4c6c-kzbgv: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;quay.io/calico/kube-controllers:v3.31.3&#34;&#xA;I0402 15:02:58.851810 24 dump.go:53] At 2026-04-02 15:00:59 +0000 UTC - event for coredns-674b8bbfcf-hdqcm: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container coredns&#xA;I0402 15:02:58.851820 24 dump.go:53] At 2026-04-02 15:00:59 +0000 UTC - event for coredns-674b8bbfcf-hdqcm: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Readiness probe failed: Get &#34;http://10.100.78.129:8181/ready&#34;: dial tcp 10.100.78.129:8181: connect: connection refused&#xA;I0402 15:02:58.851828 24 dump.go:53] At 2026-04-02 15:00:59 +0000 UTC - event for coredns-674b8bbfcf-pkmgg: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/coredns/coredns:v1.12.0&#34; already present on machine&#xA;I0402 15:02:58.851837 24 dump.go:53] At 2026-04-02 15:00:59 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-provisioner:v5.1.0&#34;&#xA;I0402 15:02:58.851845 24 dump.go:53] At 2026-04-02 15:00:59 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-attacher:v4.7.0&#34;&#xA;I0402 15:02:58.851853 24 dump.go:53] At 2026-04-02 15:01:15 +0000 UTC - event for etcd-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Readiness probe failed: Get &#34;http://0.0.0.0:2381/readyz&#34;: context deadline exceeded (Client.Timeout exceeded while awaiting headers)&#xA;I0402 15:02:58.851862 24 dump.go:53] At 2026-04-02 15:01:20 +0000 UTC - event for kube-controller-manager-kube-uhjhf-67cxd-gq627: {kubelet kube-uhjhf-67cxd-gq627} BackOff: Back-off restarting failed container kube-controller-manager in pod kube-controller-manager-kube-uhjhf-67cxd-gq627_kube-system(f674db109b4d7017ecc7e6af6005c52d)&#xA;I0402 15:02:58.851870 24 dump.go:53] At 2026-04-02 15:01:20 +0000 UTC - event for openstack-cloud-controller-manager-chxrv: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/provider-os/openstack-cloud-controller-manager:v1.33.1&#34; already present on machine&#xA;I0402 15:02:58.851877 24 dump.go:53] At 2026-04-02 15:01:22 +0000 UTC - event for coredns-674b8bbfcf-pkmgg: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: coredns&#xA;I0402 15:02:58.851885 24 dump.go:53] At 2026-04-02 15:01:25 +0000 UTC - event for coredns-674b8bbfcf-pkmgg: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container coredns&#xA;I0402 15:02:58.851894 24 dump.go:53] At 2026-04-02 15:01:27 +0000 UTC - event for manila-csi-openstack-manila-csi-nodeplugin-mnmw8: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} BackOff: Back-off restarting failed container nfs-registrar in pod manila-csi-openstack-manila-csi-nodeplugin-mnmw8_kube-system(a65eb016-b42a-43be-b7f9-8beb8b10c898)&#xA;I0402 15:02:58.851902 24 dump.go:53] At 2026-04-02 15:01:28 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-provisioner:v3.3.0&#34; in 33.874s (33.874s including waiting). Image size: 25491225 bytes.&#xA;I0402 15:02:58.851912 24 dump.go:53] At 2026-04-02 15:01:32 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: csi-provisioner&#xA;I0402 15:02:58.851919 24 dump.go:53] At 2026-04-02 15:01:33 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container csi-provisioner&#xA;I0402 15:02:58.851927 24 dump.go:53] At 2026-04-02 15:01:33 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/sig-storage/livenessprobe:v2.8.0&#34; already present on machine&#xA;I0402 15:02:58.851935 24 dump.go:53] At 2026-04-02 15:01:33 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: liveness-probe&#xA;I0402 15:02:58.851942 24 dump.go:53] At 2026-04-02 15:01:33 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container liveness-probe&#xA;I0402 15:02:58.851950 24 dump.go:53] At 2026-04-02 15:01:33 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/sig-storage/nfsplugin:v4.2.0&#34; already present on machine&#xA;I0402 15:02:58.851958 24 dump.go:53] At 2026-04-02 15:01:34 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container nfs&#xA;I0402 15:02:58.851966 24 dump.go:53] At 2026-04-02 15:01:34 +0000 UTC - event for csi-nfs-controller-7b79548448-npbhx: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: nfs&#xA;I0402 15:02:58.851974 24 dump.go:53] At 2026-04-02 15:01:34 +0000 UTC - event for nfs-csi-k8s-io: {nfs.csi.k8s.io/1775142094232-8081-nfs.csi.k8s.io } LeaderElection: 1775142094232-8081-nfs-csi-k8s-io became leader&#xA;I0402 15:02:58.851983 24 dump.go:53] At 2026-04-02 15:01:35 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Started: Started container cinder-csi-plugin&#xA;I0402 15:02:58.851991 24 dump.go:53] At 2026-04-02 15:01:35 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Created: Created container: cinder-csi-plugin&#xA;I0402 15:02:58.851998 24 dump.go:53] At 2026-04-02 15:01:35 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Successfully pulled image &#34;registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0&#34; in 41.964s (1m49.789s including waiting). Image size: 28718193 bytes.&#xA;I0402 15:02:58.852006 24 dump.go:53] At 2026-04-02 15:01:35 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Successfully pulled image &#34;registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0&#34; in 42.294s (1m31.969s including waiting). Image size: 28718193 bytes.&#xA;I0402 15:02:58.852014 24 dump.go:53] At 2026-04-02 15:01:35 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Created: Created container: cinder-csi-plugin&#xA;I0402 15:02:58.852022 24 dump.go:53] At 2026-04-02 15:01:35 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Started: Started container cinder-csi-plugin&#xA;I0402 15:02:58.852030 24 dump.go:53] At 2026-04-02 15:01:36 +0000 UTC - event for openstack-cinder-csi-nodeplugin-lnmpk: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67} Pulled: Container image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34; already present on machine&#xA;I0402 15:02:58.852038 24 dump.go:53] At 2026-04-02 15:01:36 +0000 UTC - event for openstack-cinder-csi-nodeplugin-z67qx: {kubelet kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb} Pulled: Container image &#34;registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0&#34; already present on machine&#xA;I0402 15:02:58.852046 24 dump.go:53] At 2026-04-02 15:01:41 +0000 UTC - event for k8s-keystone-auth-6dtgn: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/provider-os/k8s-keystone-auth:v1.29.0&#34; in 12.889s (45.167s including waiting). Image size: 14032747 bytes.&#xA;I0402 15:02:58.852053 24 dump.go:53] At 2026-04-02 15:01:41 +0000 UTC - event for k8s-keystone-auth-6dtgn: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: k8s-keystone-auth&#xA;I0402 15:02:58.852061 24 dump.go:53] At 2026-04-02 15:01:41 +0000 UTC - event for k8s-keystone-auth-6dtgn: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container k8s-keystone-auth&#xA;I0402 15:02:58.852068 24 dump.go:53] At 2026-04-02 15:01:48 +0000 UTC - event for cloud-controller-manager: {cloud-controller-manager } LeaderElection: kube-uhjhf-67cxd-gq627_1ec10fd8-42c4-40e3-a573-e12c0e83dcc4 became leader&#xA;I0402 15:02:58.852076 24 dump.go:53] At 2026-04-02 15:02:02 +0000 UTC - event for kube-controller-manager: {kube-controller-manager } LeaderElection: kube-uhjhf-67cxd-gq627_4c52e7ba-b01b-47a1-bb44-7b6cc932fa9c became leader&#xA;I0402 15:02:58.852086 24 dump.go:53] At 2026-04-02 15:02:04 +0000 UTC - event for kube-scheduler: {default-scheduler } LeaderElection: kube-uhjhf-67cxd-gq627_b7766e5e-63d2-4f5c-b674-2801ecba821e became leader&#xA;I0402 15:02:58.852093 24 dump.go:53] At 2026-04-02 15:02:13 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-provisioner:v5.1.0&#34; in 32.828s (1m14.818s including waiting). Image size: 32167411 bytes.&#xA;I0402 15:02:58.852103 24 dump.go:53] At 2026-04-02 15:02:13 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: nfs-provisioner&#xA;I0402 15:02:58.852111 24 dump.go:53] At 2026-04-02 15:02:14 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#34;&#xA;I0402 15:02:58.852119 24 dump.go:53] At 2026-04-02 15:02:14 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container nfs-provisioner&#xA;I0402 15:02:58.852128 24 dump.go:53] At 2026-04-02 15:02:23 +0000 UTC - event for calico-kube-controllers-9b54b4c6c-kzbgv: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;quay.io/calico/kube-controllers:v3.31.3&#34; in 10.049s (1m24.841s including waiting). Image size: 53952979 bytes.&#xA;I0402 15:02:58.852137 24 dump.go:53] At 2026-04-02 15:02:24 +0000 UTC - event for calico-kube-controllers-9b54b4c6c-kzbgv: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: calico-kube-controllers&#xA;I0402 15:02:58.852145 24 dump.go:53] At 2026-04-02 15:02:24 +0000 UTC - event for calico-kube-controllers-9b54b4c6c-kzbgv: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container calico-kube-controllers&#xA;I0402 15:02:58.852152 24 dump.go:53] At 2026-04-02 15:02:27 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: csi-attacher&#xA;I0402 15:02:58.852160 24 dump.go:53] At 2026-04-02 15:02:27 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container csi-attacher&#xA;I0402 15:02:58.852167 24 dump.go:53] At 2026-04-02 15:02:27 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Container image &#34;registry.k8s.io/sig-storage/csi-provisioner:v5.1.0&#34; already present on machine&#xA;I0402 15:02:58.852175 24 dump.go:53] At 2026-04-02 15:02:27 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-attacher:v4.7.0&#34; in 3.377s (1m28.122s including waiting). Image size: 31007188 bytes.&#xA;I0402 15:02:58.852183 24 dump.go:53] At 2026-04-02 15:02:28 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: csi-provisioner&#xA;I0402 15:02:58.852190 24 dump.go:53] At 2026-04-02 15:02:29 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container csi-provisioner&#xA;I0402 15:02:58.852199 24 dump.go:53] At 2026-04-02 15:02:29 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#34;&#xA;I0402 15:02:58.852208 24 dump.go:53] At 2026-04-02 15:02:39 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#34; in 12.18s (25.45s including waiting). Image size: 30555683 bytes.&#xA;I0402 15:02:58.852218 24 dump.go:53] At 2026-04-02 15:02:40 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: nfs-snapshotter&#xA;I0402 15:02:58.852226 24 dump.go:53] At 2026-04-02 15:02:40 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: {kubelet kube-uhjhf-67cxd-gq627} Started: Started container nfs-snapshotter&#xA;I0402 15:02:58.852257 24 dump.go:53] At 2026-04-02 15:02:40 +0000 UTC - event for manila-csi-openstack-manila-csi-controllerplugin-0: {kubelet kube-uhjhf-67cxd-gq627} Pulling: Pulling image &#34;registry.k8s.io/sig-storage/csi-resizer:v1.12.0&#34;&#xA;I0402 15:02:58.852264 24 dump.go:53] At 2026-04-02 15:02:41 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {kubelet kube-uhjhf-67cxd-gq627} Pulled: Successfully pulled image &#34;registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#34; in 1.866s (12.17s including waiting). Image size: 30555683 bytes.&#xA;I0402 15:02:58.852272 24 dump.go:53] At 2026-04-02 15:02:43 +0000 UTC - event for openstack-cinder-csi-controllerplugin-78f966889b-plrvm: {kubelet kube-uhjhf-67cxd-gq627} Created: Created container: csi-snapshotter&#xA;I0402 15:02:58.852280 24 dump.go:53] At 2026-04-02 15:02:53 +0000 UTC - event for calico-kube-controllers-9b54b4c6c-kzbgv: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Liveness probe failed: Error verifying datastore: Get &#34;https://10.254.0.1:443/apis/crd.projectcalico.org/v1/clusterinformations/default&#34;: context deadline exceeded&#xA;&#xA;I0402 15:02:58.852288 24 dump.go:53] At 2026-04-02 15:02:55 +0000 UTC - event for calico-kube-controllers-9b54b4c6c-kzbgv: {kubelet kube-uhjhf-67cxd-gq627} Unhealthy: Readiness probe failed: Error verifying datastore: Get &#34;https://10.254.0.1:443/apis/crd.projectcalico.org/v1/clusterinformations/default&#34;: context deadline exceeded&#xA;&#xA;I0402 15:02:59.058495 24 resource.go:168] POD                                                     NODE                                         PHASE    GRACE  CONDITIONS&#xA;I0402 15:02:59.058550 24 resource.go:175] calico-kube-controllers-9b54b4c6c-kzbgv                 kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:02:24 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:02:24 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:02:24 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:52 +0000 UTC  }]&#xA;I0402 15:02:59.058568 24 resource.go:175] calico-node-6wvr4                                       kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb  Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:32 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:47 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:00 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:00 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  }]&#xA;I0402 15:02:59.058583 24 resource.go:175] calico-node-7pj2s                                       kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67  Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:16 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:46 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:00 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:00 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  }]&#xA;I0402 15:02:59.058599 24 resource.go:175] calico-node-b7l94                                       kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:34 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:10 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:22 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:22 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  }]&#xA;I0402 15:02:59.058615 24 resource.go:175] coredns-674b8bbfcf-hdqcm                                kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:59 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:21 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:21 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:52 +0000 UTC  }]&#xA;I0402 15:02:59.058632 24 resource.go:175] coredns-674b8bbfcf-pkmgg                                kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:26 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:26 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:26 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:52 +0000 UTC  }]&#xA;I0402 15:02:59.058650 24 resource.go:175] csi-nfs-controller-7b79548448-npbhx                     kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:34 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:34 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:34 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:52 +0000 UTC  }]&#xA;I0402 15:02:59.058665 24 resource.go:175] csi-nfs-node-4xmg9                                      kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67  Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:54 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:54 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:54 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  }]&#xA;I0402 15:02:59.058680 24 resource.go:175] csi-nfs-node-jtj8j                                      kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb  Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:54 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:00 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:54 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:54 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  }]&#xA;I0402 15:02:59.058707 24 resource.go:175] csi-nfs-node-vn2br                                      kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:05 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:28 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:05 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:05 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:28 +0000 UTC  }]&#xA;I0402 15:02:59.058723 24 resource.go:175] etcd-kube-uhjhf-67cxd-gq627                             kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:36 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:36 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  }]&#xA;I0402 15:02:59.058740 24 resource.go:175] k8s-keystone-auth-6dtgn                                 kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:41 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:41 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:41 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC  }]&#xA;I0402 15:02:59.058821 24 resource.go:175] kube-apiserver-kube-uhjhf-67cxd-gq627                   kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:06 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  } {Ready 0 False 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:02:58 +0000 UTC ContainersNotReady containers with unready status: [kube-apiserver]} {ContainersReady 0 False 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:02:58 +0000 UTC ContainersNotReady containers with unready status: [kube-apiserver]} {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  }]&#xA;I0402 15:02:59.058840 24 resource.go:175] kube-controller-manager-kube-uhjhf-67cxd-gq627          kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:52 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:52 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  }]&#xA;I0402 15:02:59.058855 24 resource.go:175] kube-proxy-8pc5r                                        kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb  Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:10 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:00 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:10 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:10 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  }]&#xA;I0402 15:02:59.058873 24 resource.go:175] kube-proxy-djtqn                                        kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67  Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:09 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:09 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:09 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  }]&#xA;I0402 15:02:59.058889 24 resource.go:175] kube-proxy-mx8ps                                        kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:30 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:28 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:30 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:30 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  }]&#xA;I0402 15:02:59.058907 24 resource.go:175] kube-scheduler-kube-uhjhf-67cxd-gq627                   kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:02:01 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:02:01 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:27 +0000 UTC  }]&#xA;I0402 15:02:59.058924 24 resource.go:175] manila-csi-openstack-manila-csi-controllerplugin-0      kube-uhjhf-67cxd-gq627                       Pending         [{PodReadyToStartContainers 0 False 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC  } {Ready 0 False 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC ContainersNotReady containers with unready status: [nfs-provisioner nfs-snapshotter nfs-resizer nfs-nodeplugin]} {ContainersReady 0 False 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC ContainersNotReady containers with unready status: [nfs-provisioner nfs-snapshotter nfs-resizer nfs-nodeplugin]} {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:52 +0000 UTC  }]&#xA;I0402 15:02:59.058939 24 resource.go:175] manila-csi-openstack-manila-csi-nodeplugin-66z9w        kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb  Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:59:43 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:00 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:48 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:48 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  }]&#xA;I0402 15:02:59.058955 24 resource.go:175] manila-csi-openstack-manila-csi-nodeplugin-mnmw8        kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67  Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:01 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:42 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:42 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  }]&#xA;I0402 15:02:59.058971 24 resource.go:175] manila-csi-openstack-manila-csi-nodeplugin-pnrtc        kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:51 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:28 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:06 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:06 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:28 +0000 UTC  }]&#xA;I0402 15:02:59.058987 24 resource.go:175] openstack-cinder-csi-controllerplugin-78f966889b-plrvm  kube-uhjhf-67cxd-gq627                       Pending         [{PodReadyToStartContainers 0 False 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC  } {Ready 0 False 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC ContainersNotReady containers with unready status: [csi-attacher csi-provisioner csi-snapshotter csi-resizer liveness-probe cinder-csi-plugin]} {ContainersReady 0 False 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:53 +0000 UTC ContainersNotReady containers with unready status: [csi-attacher csi-provisioner csi-snapshotter csi-resizer liveness-probe cinder-csi-plugin]} {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:52 +0000 UTC  }]&#xA;I0402 15:02:59.059003 24 resource.go:175] openstack-cinder-csi-nodeplugin-h74dd                   kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:55 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:26 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:21 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:21 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:56:26 +0000 UTC  }]&#xA;I0402 15:02:59.059018 24 resource.go:175] openstack-cinder-csi-nodeplugin-lnmpk                   kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67  Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:36 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:39 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:39 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  }]&#xA;I0402 15:02:59.059036 24 resource.go:175] openstack-cinder-csi-nodeplugin-z67qx                   kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb  Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:36 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:58:00 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:39 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:39 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 14:57:58 +0000 UTC  }]&#xA;I0402 15:02:59.059051 24 resource.go:175] openstack-cloud-controller-manager-chxrv                kube-uhjhf-67cxd-gq627                       Running         [{PodReadyToStartContainers 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:49 +0000 UTC  } {Initialized 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:03 +0000 UTC  } {Ready 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:28 +0000 UTC  } {ContainersReady 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:01:28 +0000 UTC  } {PodScheduled 0 True 0001-01-01 00:00:00 +0000 UTC 2026-04-02 15:00:03 +0000 UTC  }]&#xA;I0402 15:02:59.059056 24 resource.go:178] &#xA;I0402 15:03:00.298138 24 dump.go:109] &#xA;Logging node info for node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:00.306481 24 dump.go:114] Node Info: &amp;Node{ObjectMeta:{kube-uhjhf-67cxd-gq627    baf1517a-ebc5-431b-a3bb-cd9f95e92c64 1836 0 2026-04-02 14:55:40 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:m1.large beta.kubernetes.io/os:linux failure-domain.beta.kubernetes.io/region:RegionOne failure-domain.beta.kubernetes.io/zone:nova kubernetes.io/arch:amd64 kubernetes.io/hostname:kube-uhjhf-67cxd-gq627 kubernetes.io/os:linux node-role.kubernetes.io/control-plane: node-role.kubernetes.io/master: node.kubernetes.io/exclude-from-external-load-balancers: node.kubernetes.io/instance-type:m1.large topology.cinder.csi.openstack.org/zone:nova topology.kubernetes.io/region:RegionOne topology.kubernetes.io/zone:nova] map[cluster.x-k8s.io/annotations-from-machine: cluster.x-k8s.io/cluster-name:kube-uhjhf cluster.x-k8s.io/cluster-namespace:magnum-system cluster.x-k8s.io/labels-from-machine:node-role.kubernetes.io/master cluster.x-k8s.io/machine:kube-uhjhf-67cxd-gq627 cluster.x-k8s.io/owner-kind:KubeadmControlPlane cluster.x-k8s.io/owner-name:kube-uhjhf-67cxd csi.volume.kubernetes.io/nodeid:{&#34;cinder.csi.openstack.org&#34;:&#34;f92b3aa6-5ff2-44a1-915a-6f990bb951bf&#34;,&#34;nfs.csi.k8s.io&#34;:&#34;kube-uhjhf-67cxd-gq627&#34;,&#34;nfs.manila.csi.openstack.org&#34;:&#34;f92b3aa6-5ff2-44a1-915a-6f990bb951bf&#34;} kubeadm.alpha.kubernetes.io/cri-socket:unix:///var/run/containerd/containerd.sock node.alpha.kubernetes.io/ttl:0 projectcalico.org/IPv4Address:192.168.24.167/24 projectcalico.org/IPv4IPIPTunnelAddr:10.100.78.128 projectcalico.org/Interfaces:[{&#34;name&#34;:&#34;ens3&#34;,&#34;addresses&#34;:[&#34;192.168.24.167&#34;]}] volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2026-04-02 14:55:40 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}}} } {kubeadm Update v1 2026-04-02 14:56:26 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/control-plane&#34;:{},&#34;f:node.kubernetes.io/exclude-from-external-load-balancers&#34;:{}}}} } {kube-controller-manager Update v1 2026-04-02 14:59:59 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.100.0.0/24\&#34;&#34;:{}}}} } {calico-node Update v1 2026-04-02 15:00:19 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:projectcalico.org/IPv4Address&#34;:{},&#34;f:projectcalico.org/IPv4IPIPTunnelAddr&#34;:{},&#34;f:projectcalico.org/Interfaces&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}} status} {manager Update v1 2026-04-02 15:00:52 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:cluster.x-k8s.io/annotations-from-machine&#34;:{},&#34;f:cluster.x-k8s.io/cluster-name&#34;:{},&#34;f:cluster.x-k8s.io/cluster-namespace&#34;:{},&#34;f:cluster.x-k8s.io/labels-from-machine&#34;:{},&#34;f:cluster.x-k8s.io/machine&#34;:{},&#34;f:cluster.x-k8s.io/owner-kind&#34;:{},&#34;f:cluster.x-k8s.io/owner-name&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}} } {openstack-cloud-controller-manager Update v1 2026-04-02 15:00:52 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:beta.kubernetes.io/instance-type&#34;:{},&#34;f:failure-domain.beta.kubernetes.io/region&#34;:{},&#34;f:failure-domain.beta.kubernetes.io/zone&#34;:{},&#34;f:node.kubernetes.io/instance-type&#34;:{},&#34;f:topology.kubernetes.io/region&#34;:{},&#34;f:topology.kubernetes.io/zone&#34;:{}}},&#34;f:spec&#34;:{&#34;f:providerID&#34;:{},&#34;f:taints&#34;:{}}} } {openstack-cloud-controller-manager Update v1 2026-04-02 15:00:53 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}}}} status} {kubelet Update v1 2026-04-02 15:02:29 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:csi.volume.kubernetes.io/nodeid&#34;:{}},&#34;f:labels&#34;:{&#34;f:topology.cinder.csi.openstack.org/zone&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}} status}]},Spec:NodeSpec{PodCIDR:10.100.0.0/24,DoNotUseExternalID:,ProviderID:openstack:///f92b3aa6-5ff2-44a1-915a-6f990bb951bf,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/control-plane,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.100.0.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{4 0} {&lt;nil&gt;} 4 DecimalSI},ephemeral-storage: {{80514002944 0} {&lt;nil&gt;} 78626956Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{8322887680 0} {&lt;nil&gt;} 8127820Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{4 0} {&lt;nil&gt;} 4 DecimalSI},ephemeral-storage: {{72462602530 0} {&lt;nil&gt;} 72462602530 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{8218030080 0} {&lt;nil&gt;} 8025420Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2026-04-02 15:00:19 +0000 UTC,LastTransitionTime:2026-04-02 15:00:19 +0000 UTC,Reason:CalicoIsUp,Message:Calico is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2026-04-02 15:02:29 +0000 UTC,LastTransitionTime:2026-04-02 14:55:40 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2026-04-02 15:02:29 +0000 UTC,LastTransitionTime:2026-04-02 14:55:40 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2026-04-02 15:02:29 +0000 UTC,LastTransitionTime:2026-04-02 14:55:40 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2026-04-02 15:02:29 +0000 UTC,LastTransitionTime:2026-04-02 14:59:31 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:192.168.24.167,},NodeAddress{Type:Hostname,Address:kube-uhjhf-67cxd-gq627,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:f92b3aa65ff244a1915a6f990bb951bf,SystemUUID:f92b3aa6-5ff2-44a1-915a-6f990bb951bf,BootID:84bbcaff-ba60-424f-ba0b-586c382fbf42,KernelVersion:5.15.0-174-generic,OSImage:Ubuntu 22.04.5 LTS,ContainerRuntimeVersion:containerd://2.2.2,KubeletVersion:v1.33.10,KubeProxyVersion:,OperatingSystem:linux,Architecture:amd64,Swap:nil,},Images:[]ContainerImage{ContainerImage{Names:[quay.io/calico/node@sha256:f2339c4ff3a57228cbc39a1f67ab81abded1997d843e0e0b1e86664c7c4eb6c0 quay.io/calico/node:v3.31.3],SizeBytes:159483847,},ContainerImage{Names:[quay.io/calico/cni@sha256:a6ddb6d00a53f98c5392eb3b525cf00af10cace1f6ca9414230ff652e6700e78 quay.io/calico/cni:v3.31.3],SizeBytes:72141081,},ContainerImage{Names:[quay.io/calico/kube-controllers@sha256:3470040e0995deb6f6e23245a5149c9db1fc698a8a49da376cabe8dece62a0dd quay.io/calico/kube-controllers:v3.31.3],SizeBytes:53952979,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfsplugin@sha256:1d9f524c0fd5d714cf274a0ddf5dceebeb1848349de5687e530c33dd2675e4ba registry.k8s.io/sig-storage/nfsplugin:v4.2.0],SizeBytes:49783016,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f registry.k8s.io/sig-storage/csi-provisioner:v5.1.0],SizeBytes:32167411,},ContainerImage{Names:[registry.k8s.io/kube-proxy@sha256:e8151e38ef22f032dba686cc1bba5a3e525dedbe2d549fa44e653fe79426e261 registry.k8s.io/kube-proxy:v1.33.10],SizeBytes:31827782,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:6e54dae32284f60a2de1645c527715e4c934e5ce7899560c0d845bac957118dc registry.k8s.io/sig-storage/csi-attacher:v4.7.0],SizeBytes:31007188,},ContainerImage{Names:[registry.k8s.io/kube-apiserver@sha256:bbff81e41af4bfca88a1d05a066a48e12e2689c534d073a8c688e3ad6c8701e3 registry.k8s.io/kube-apiserver:v1.33.10],SizeBytes:29986018,},ContainerImage{Names:[registry.k8s.io/provider-os/cinder-csi-plugin@sha256:8dbc1d5c42c143d26ba7fea5ca36a38ab512983b7df16329b486421615381f96 registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0],SizeBytes:28718193,},ContainerImage{Names:[registry.k8s.io/kube-controller-manager@sha256:b0880d6ee19f2b9148d3d37008c5ee9fc73976e8edad4d0709f11d32ab3ee709 registry.k8s.io/kube-controller-manager:v1.33.10],SizeBytes:27552094,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6 registry.k8s.io/etcd:3.5.24-0],SizeBytes:23716032,},ContainerImage{Names:[registry.k8s.io/kube-scheduler@sha256:dc1a1aec3bb0ed126b1adff795935124f719969356b24a159fc1a2a0883b89bc registry.k8s.io/kube-scheduler:v1.33.10],SizeBytes:21692956,},ContainerImage{Names:[registry.k8s.io/provider-os/openstack-cloud-controller-manager@sha256:de8a6da8c31c7b967625451a7169309d6f77aee1ff64b3f8e6ba8d8810ce2a22 registry.k8s.io/provider-os/openstack-cloud-controller-manager:v1.33.1],SizeBytes:21265731,},ContainerImage{Names:[registry.k8s.io/provider-os/manila-csi-plugin@sha256:2cae8c83c215c1f038f50f007d40151d3651a511214827e6dd409b2a919fdb97 registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0],SizeBytes:21014294,},ContainerImage{Names:[registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97 registry.k8s.io/coredns/coredns:v1.12.0],SizeBytes:20939036,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:33692aed26aaf105b4d6e66280cceca9e0463f500c81b5d8c955428a75438f32 registry.k8s.io/sig-storage/livenessprobe:v2.14.0],SizeBytes:14311007,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0],SizeBytes:14038309,},ContainerImage{Names:[registry.k8s.io/provider-os/k8s-keystone-auth@sha256:7088616ac16961be52fbac60c321a87631b8ad2463158b06332e18391aae9a11 registry.k8s.io/provider-os/k8s-keystone-auth:v1.29.0],SizeBytes:14032747,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a13bff2ed69af0cf4270f0cf47bdedf75a56c095cd95b91195ae6c713a9b1845 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.2],SizeBytes:10074312,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:cacee2b5c36dd59d4c7e8469c05c9e4ef53ecb2df9025fa8c10cdaf61bce62f0 registry.k8s.io/sig-storage/livenessprobe:v2.8.0],SizeBytes:8892463,},ContainerImage{Names:[registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097 registry.k8s.io/pause:3.9],SizeBytes:321520,},ContainerImage{Names:[registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a registry.k8s.io/pause:3.10],SizeBytes:320368,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,RuntimeHandlers:[]NodeRuntimeHandler{NodeRuntimeHandler{Name:,Features:&amp;NodeRuntimeHandlerFeatures{RecursiveReadOnlyMounts:*true,UserNamespaces:*true,},},NodeRuntimeHandler{Name:runc,Features:&amp;NodeRuntimeHandlerFeatures{RecursiveReadOnlyMounts:*true,UserNamespaces:*true,},},},Features:&amp;NodeFeatures{SupplementalGroupsPolicy:*true,},},}&#xA;I0402 15:03:00.306577 24 dump.go:116] &#xA;Logging kubelet events for node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:00.861088 24 dump.go:121] &#xA;Logging pods the kubelet thinks are on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:01.239465 24 dump.go:128] kube-system/openstack-cinder-csi-nodeplugin-h74dd started at 2026-04-02 14:56:26 +0000 UTC (0+3 container statuses recorded)&#xA;I0402 15:03:01.239496 24 dump.go:134] &#x9;Container cinder-csi-plugin ready: true, restart count 0&#xA;I0402 15:03:01.239504 24 dump.go:134] &#x9;Container liveness-probe ready: true, restart count 0&#xA;I0402 15:03:01.239512 24 dump.go:134] &#x9;Container node-driver-registrar ready: true, restart count 2&#xA;I0402 15:03:01.239523 24 dump.go:128] kube-system/kube-proxy-mx8ps started at 2026-04-02 14:56:28 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:01.239533 24 dump.go:134] &#x9;Container kube-proxy ready: true, restart count 0&#xA;I0402 15:03:01.239543 24 dump.go:128] kube-system/csi-nfs-node-vn2br started at 2026-04-02 14:56:28 +0000 UTC (0+3 container statuses recorded)&#xA;I0402 15:03:01.239552 24 dump.go:134] &#x9;Container liveness-probe ready: true, restart count 0&#xA;I0402 15:03:01.239561 24 dump.go:134] &#x9;Container nfs ready: true, restart count 0&#xA;I0402 15:03:01.239569 24 dump.go:134] &#x9;Container node-driver-registrar ready: true, restart count 0&#xA;I0402 15:03:01.239580 24 dump.go:128] kube-system/coredns-674b8bbfcf-hdqcm started at 2026-04-02 15:00:53 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:01.239589 24 dump.go:134] &#x9;Container coredns ready: true, restart count 0&#xA;I0402 15:03:01.239599 24 dump.go:128] kube-system/openstack-cinder-csi-controllerplugin-78f966889b-plrvm started at 2026-04-02 15:00:53 +0000 UTC (0+6 container statuses recorded)&#xA;I0402 15:03:01.239606 24 dump.go:134] &#x9;Container cinder-csi-plugin ready: false, restart count 0&#xA;I0402 15:03:01.239637 24 dump.go:134] &#x9;Container csi-attacher ready: false, restart count 0&#xA;I0402 15:03:01.239644 24 dump.go:134] &#x9;Container csi-provisioner ready: false, restart count 0&#xA;I0402 15:03:01.239718 24 dump.go:134] &#x9;Container csi-resizer ready: false, restart count 0&#xA;I0402 15:03:01.239726 24 dump.go:134] &#x9;Container csi-snapshotter ready: false, restart count 0&#xA;I0402 15:03:01.239735 24 dump.go:134] &#x9;Container liveness-probe ready: false, restart count 0&#xA;I0402 15:03:01.239770 24 dump.go:128] kube-system/k8s-keystone-auth-6dtgn started at 2026-04-02 15:00:53 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:01.239776 24 dump.go:134] &#x9;Container k8s-keystone-auth ready: true, restart count 0&#xA;I0402 15:03:01.239785 24 dump.go:128] kube-system/etcd-kube-uhjhf-67cxd-gq627 started at 2026-04-02 14:56:27 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:01.239791 24 dump.go:134] &#x9;Container etcd ready: true, restart count 0&#xA;I0402 15:03:01.239801 24 dump.go:128] kube-system/manila-csi-openstack-manila-csi-nodeplugin-pnrtc started at 2026-04-02 14:56:28 +0000 UTC (0+2 container statuses recorded)&#xA;I0402 15:03:01.239807 24 dump.go:134] &#x9;Container nfs-nodeplugin ready: true, restart count 2&#xA;I0402 15:03:01.239814 24 dump.go:134] &#x9;Container nfs-registrar ready: true, restart count 0&#xA;I0402 15:03:01.239823 24 dump.go:128] kube-system/manila-csi-openstack-manila-csi-controllerplugin-0 started at 2026-04-02 15:00:53 +0000 UTC (0+4 container statuses recorded)&#xA;I0402 15:03:01.239828 24 dump.go:134] &#x9;Container nfs-nodeplugin ready: false, restart count 0&#xA;I0402 15:03:01.239835 24 dump.go:134] &#x9;Container nfs-provisioner ready: false, restart count 0&#xA;I0402 15:03:01.239841 24 dump.go:134] &#x9;Container nfs-resizer ready: false, restart count 0&#xA;I0402 15:03:01.239848 24 dump.go:134] &#x9;Container nfs-snapshotter ready: false, restart count 0&#xA;I0402 15:03:01.239854 24 dump.go:128] kube-system/calico-kube-controllers-9b54b4c6c-kzbgv started at 2026-04-02 15:00:53 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:01.239860 24 dump.go:134] &#x9;Container calico-kube-controllers ready: true, restart count 0&#xA;I0402 15:03:01.239867 24 dump.go:128] kube-system/coredns-674b8bbfcf-pkmgg started at 2026-04-02 15:00:53 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:01.239872 24 dump.go:134] &#x9;Container coredns ready: true, restart count 0&#xA;I0402 15:03:01.239879 24 dump.go:128] kube-system/kube-apiserver-kube-uhjhf-67cxd-gq627 started at 2026-04-02 14:56:27 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:01.239886 24 dump.go:134] &#x9;Container kube-apiserver ready: true, restart count 0&#xA;I0402 15:03:01.239893 24 dump.go:128] kube-system/kube-scheduler-kube-uhjhf-67cxd-gq627 started at 2026-04-02 14:56:27 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:01.239900 24 dump.go:134] &#x9;Container kube-scheduler ready: false, restart count 3&#xA;I0402 15:03:01.239907 24 dump.go:128] kube-system/calico-node-b7l94 started at 2026-04-02 14:56:28 +0000 UTC (3+1 container statuses recorded)&#xA;I0402 15:03:01.239915 24 dump.go:130] &#x9;Init container upgrade-ipam ready: true, restart count 0&#xA;I0402 15:03:01.239920 24 dump.go:130] &#x9;Init container install-cni ready: true, restart count 3&#xA;I0402 15:03:01.239926 24 dump.go:130] &#x9;Init container ebpf-bootstrap ready: true, restart count 0&#xA;I0402 15:03:01.239932 24 dump.go:134] &#x9;Container calico-node ready: true, restart count 0&#xA;I0402 15:03:01.239939 24 dump.go:128] kube-system/openstack-cloud-controller-manager-chxrv started at 2026-04-02 15:00:03 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:01.239947 24 dump.go:134] &#x9;Container openstack-cloud-controller-manager ready: false, restart count 1&#xA;I0402 15:03:01.239955 24 dump.go:128] kube-system/csi-nfs-controller-7b79548448-npbhx started at 2026-04-02 15:00:53 +0000 UTC (0+3 container statuses recorded)&#xA;I0402 15:03:01.239960 24 dump.go:134] &#x9;Container csi-provisioner ready: false, restart count 0&#xA;I0402 15:03:01.239968 24 dump.go:134] &#x9;Container liveness-probe ready: true, restart count 0&#xA;I0402 15:03:01.239974 24 dump.go:134] &#x9;Container nfs ready: true, restart count 0&#xA;I0402 15:03:01.239981 24 dump.go:128] kube-system/kube-controller-manager-kube-uhjhf-67cxd-gq627 started at 2026-04-02 14:56:27 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:01.239988 24 dump.go:134] &#x9;Container kube-controller-manager ready: false, restart count 3&#xA;I0402 15:03:01.691008 24 kubelet_metrics.go:206] &#xA;Latency metrics for node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:01.691039 24 dump.go:109] &#xA;Logging node info for node kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67&#xA;I0402 15:03:01.829436 24 dump.go:114] Node Info: &amp;Node{ObjectMeta:{kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67    8b8122c4-42d4-45a9-a666-b4e1ab1e6f82 1731 0 2026-04-02 14:57:57 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:m1.large beta.kubernetes.io/os:linux failure-domain.beta.kubernetes.io/region:RegionOne failure-domain.beta.kubernetes.io/zone:nova kubernetes.io/arch:amd64 kubernetes.io/hostname:kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67 kubernetes.io/os:linux node-role.kubernetes.io/worker: node.cluster.x-k8s.io/nodegroup:default-worker node.kubernetes.io/instance-type:m1.large topology.cinder.csi.openstack.org/zone:nova topology.kubernetes.io/region:RegionOne topology.kubernetes.io/zone:nova] map[cluster.x-k8s.io/annotations-from-machine: cluster.x-k8s.io/cluster-name:kube-uhjhf cluster.x-k8s.io/cluster-namespace:magnum-system cluster.x-k8s.io/labels-from-machine:node-role.kubernetes.io/worker,node.cluster.x-k8s.io/nodegroup cluster.x-k8s.io/machine:kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67 cluster.x-k8s.io/owner-kind:MachineSet cluster.x-k8s.io/owner-name:kube-uhjhf-default-worker-6qhl5-ctzr5 csi.volume.kubernetes.io/nodeid:{&#34;cinder.csi.openstack.org&#34;:&#34;9a9302ac-1292-4312-af05-f44bd3594e42&#34;,&#34;nfs.csi.k8s.io&#34;:&#34;kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67&#34;,&#34;nfs.manila.csi.openstack.org&#34;:&#34;9a9302ac-1292-4312-af05-f44bd3594e42&#34;} kubeadm.alpha.kubernetes.io/cri-socket:unix:///var/run/containerd/containerd.sock node.alpha.kubernetes.io/ttl:0 projectcalico.org/IPv4Address:192.168.24.125/24 projectcalico.org/IPv4IPIPTunnelAddr:10.100.250.64 projectcalico.org/Interfaces:[{&#34;name&#34;:&#34;ens3&#34;,&#34;addresses&#34;:[&#34;192.168.24.125&#34;]}] volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2026-04-02 14:57:57 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}}} } {kubeadm Update v1 2026-04-02 14:57:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}} } {kube-controller-manager Update v1 2026-04-02 14:59:59 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.100.1.0/24\&#34;&#34;:{}}}} } {manager Update v1 2026-04-02 15:00:53 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:cluster.x-k8s.io/annotations-from-machine&#34;:{},&#34;f:cluster.x-k8s.io/cluster-name&#34;:{},&#34;f:cluster.x-k8s.io/cluster-namespace&#34;:{},&#34;f:cluster.x-k8s.io/labels-from-machine&#34;:{},&#34;f:cluster.x-k8s.io/machine&#34;:{},&#34;f:cluster.x-k8s.io/owner-kind&#34;:{},&#34;f:cluster.x-k8s.io/owner-name&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/worker&#34;:{},&#34;f:node.cluster.x-k8s.io/nodegroup&#34;:{}}}} } {openstack-cloud-controller-manager Update v1 2026-04-02 15:00:53 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:beta.kubernetes.io/instance-type&#34;:{},&#34;f:failure-domain.beta.kubernetes.io/region&#34;:{},&#34;f:failure-domain.beta.kubernetes.io/zone&#34;:{},&#34;f:node.kubernetes.io/instance-type&#34;:{},&#34;f:topology.kubernetes.io/region&#34;:{},&#34;f:topology.kubernetes.io/zone&#34;:{}}},&#34;f:spec&#34;:{&#34;f:providerID&#34;:{}}} } {openstack-cloud-controller-manager Update v1 2026-04-02 15:00:53 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}}}} status} {calico-node Update v1 2026-04-02 15:00:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:projectcalico.org/IPv4Address&#34;:{},&#34;f:projectcalico.org/IPv4IPIPTunnelAddr&#34;:{},&#34;f:projectcalico.org/Interfaces&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}} status} {kubelet Update v1 2026-04-02 15:01:59 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:csi.volume.kubernetes.io/nodeid&#34;:{}},&#34;f:labels&#34;:{&#34;f:topology.cinder.csi.openstack.org/zone&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}} status}]},Spec:NodeSpec{PodCIDR:10.100.1.0/24,DoNotUseExternalID:,ProviderID:openstack:///9a9302ac-1292-4312-af05-f44bd3594e42,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.100.1.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{4 0} {&lt;nil&gt;} 4 DecimalSI},ephemeral-storage: {{80514002944 0} {&lt;nil&gt;} 78626956Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{8322891776 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{4 0} {&lt;nil&gt;} 4 DecimalSI},ephemeral-storage: {{72462602530 0} {&lt;nil&gt;} 72462602530 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{8218034176 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2026-04-02 15:00:58 +0000 UTC,LastTransitionTime:2026-04-02 15:00:58 +0000 UTC,Reason:CalicoIsUp,Message:Calico is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2026-04-02 15:01:59 +0000 UTC,LastTransitionTime:2026-04-02 14:57:57 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2026-04-02 15:01:59 +0000 UTC,LastTransitionTime:2026-04-02 14:57:57 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2026-04-02 15:01:59 +0000 UTC,LastTransitionTime:2026-04-02 14:57:57 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2026-04-02 15:01:59 +0000 UTC,LastTransitionTime:2026-04-02 14:59:23 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:192.168.24.125,},NodeAddress{Type:Hostname,Address:kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:9a9302ac12924312af05f44bd3594e42,SystemUUID:9a9302ac-1292-4312-af05-f44bd3594e42,BootID:ed2e1acc-d5a4-48d0-9a82-85f99021ef05,KernelVersion:5.15.0-174-generic,OSImage:Ubuntu 22.04.5 LTS,ContainerRuntimeVersion:containerd://2.2.2,KubeletVersion:v1.33.10,KubeProxyVersion:,OperatingSystem:linux,Architecture:amd64,Swap:nil,},Images:[]ContainerImage{ContainerImage{Names:[quay.io/calico/node@sha256:f2339c4ff3a57228cbc39a1f67ab81abded1997d843e0e0b1e86664c7c4eb6c0 quay.io/calico/node:v3.31.3],SizeBytes:159483847,},ContainerImage{Names:[quay.io/calico/cni@sha256:a6ddb6d00a53f98c5392eb3b525cf00af10cace1f6ca9414230ff652e6700e78 quay.io/calico/cni:v3.31.3],SizeBytes:72141081,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfsplugin@sha256:1d9f524c0fd5d714cf274a0ddf5dceebeb1848349de5687e530c33dd2675e4ba registry.k8s.io/sig-storage/nfsplugin:v4.2.0],SizeBytes:49783016,},ContainerImage{Names:[registry.k8s.io/kube-proxy@sha256:e8151e38ef22f032dba686cc1bba5a3e525dedbe2d549fa44e653fe79426e261 registry.k8s.io/kube-proxy:v1.33.10],SizeBytes:31827782,},ContainerImage{Names:[registry.k8s.io/kube-apiserver@sha256:bbff81e41af4bfca88a1d05a066a48e12e2689c534d073a8c688e3ad6c8701e3 registry.k8s.io/kube-apiserver:v1.33.10],SizeBytes:29986018,},ContainerImage{Names:[registry.k8s.io/provider-os/cinder-csi-plugin@sha256:8dbc1d5c42c143d26ba7fea5ca36a38ab512983b7df16329b486421615381f96 registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0],SizeBytes:28718193,},ContainerImage{Names:[registry.k8s.io/kube-controller-manager@sha256:b0880d6ee19f2b9148d3d37008c5ee9fc73976e8edad4d0709f11d32ab3ee709 registry.k8s.io/kube-controller-manager:v1.33.10],SizeBytes:27552094,},ContainerImage{Names:[registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6 registry.k8s.io/etcd:3.5.24-0],SizeBytes:23716032,},ContainerImage{Names:[registry.k8s.io/kube-scheduler@sha256:dc1a1aec3bb0ed126b1adff795935124f719969356b24a159fc1a2a0883b89bc registry.k8s.io/kube-scheduler:v1.33.10],SizeBytes:21692956,},ContainerImage{Names:[registry.k8s.io/provider-os/manila-csi-plugin@sha256:2cae8c83c215c1f038f50f007d40151d3651a511214827e6dd409b2a919fdb97 registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0],SizeBytes:21014294,},ContainerImage{Names:[registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97 registry.k8s.io/coredns/coredns:v1.12.0],SizeBytes:20939036,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:33692aed26aaf105b4d6e66280cceca9e0463f500c81b5d8c955428a75438f32 registry.k8s.io/sig-storage/livenessprobe:v2.14.0],SizeBytes:14311007,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0],SizeBytes:14038309,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a13bff2ed69af0cf4270f0cf47bdedf75a56c095cd95b91195ae6c713a9b1845 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.2],SizeBytes:10074312,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:cacee2b5c36dd59d4c7e8469c05c9e4ef53ecb2df9025fa8c10cdaf61bce62f0 registry.k8s.io/sig-storage/livenessprobe:v2.8.0],SizeBytes:8892463,},ContainerImage{Names:[registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097 registry.k8s.io/pause:3.9],SizeBytes:321520,},ContainerImage{Names:[registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a registry.k8s.io/pause:3.10],SizeBytes:320368,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,RuntimeHandlers:[]NodeRuntimeHandler{NodeRuntimeHandler{Name:,Features:&amp;NodeRuntimeHandlerFeatures{RecursiveReadOnlyMounts:*true,UserNamespaces:*true,},},NodeRuntimeHandler{Name:runc,Features:&amp;NodeRuntimeHandlerFeatures{RecursiveReadOnlyMounts:*true,UserNamespaces:*true,},},},Features:&amp;NodeFeatures{SupplementalGroupsPolicy:*true,},},}&#xA;I0402 15:03:01.829481 24 dump.go:116] &#xA;Logging kubelet events for node kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67&#xA;I0402 15:03:01.843248 24 dump.go:121] &#xA;Logging pods the kubelet thinks are on node kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67&#xA;I0402 15:03:02.306544 24 dump.go:128] kube-system/calico-node-7pj2s started at 2026-04-02 14:57:58 +0000 UTC (3+1 container statuses recorded)&#xA;I0402 15:03:02.306579 24 dump.go:130] &#x9;Init container upgrade-ipam ready: true, restart count 0&#xA;I0402 15:03:02.306590 24 dump.go:130] &#x9;Init container install-cni ready: true, restart count 2&#xA;I0402 15:03:02.306599 24 dump.go:130] &#x9;Init container ebpf-bootstrap ready: true, restart count 0&#xA;I0402 15:03:02.306608 24 dump.go:134] &#x9;Container calico-node ready: true, restart count 0&#xA;I0402 15:03:02.306618 24 dump.go:128] kube-system/csi-nfs-node-4xmg9 started at 2026-04-02 14:57:58 +0000 UTC (0+3 container statuses recorded)&#xA;I0402 15:03:02.306625 24 dump.go:134] &#x9;Container liveness-probe ready: true, restart count 0&#xA;I0402 15:03:02.306634 24 dump.go:134] &#x9;Container nfs ready: true, restart count 0&#xA;I0402 15:03:02.306641 24 dump.go:134] &#x9;Container node-driver-registrar ready: true, restart count 0&#xA;I0402 15:03:02.306652 24 dump.go:128] kube-system/kube-proxy-djtqn started at 2026-04-02 14:57:58 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:02.306661 24 dump.go:134] &#x9;Container kube-proxy ready: true, restart count 0&#xA;I0402 15:03:02.306670 24 dump.go:128] kube-system/manila-csi-openstack-manila-csi-nodeplugin-mnmw8 started at 2026-04-02 14:57:58 +0000 UTC (0+2 container statuses recorded)&#xA;I0402 15:03:02.306679 24 dump.go:134] &#x9;Container nfs-nodeplugin ready: true, restart count 4&#xA;I0402 15:03:02.306687 24 dump.go:134] &#x9;Container nfs-registrar ready: true, restart count 3&#xA;I0402 15:03:02.306696 24 dump.go:128] kube-system/openstack-cinder-csi-nodeplugin-lnmpk started at 2026-04-02 14:57:58 +0000 UTC (0+3 container statuses recorded)&#xA;I0402 15:03:02.306705 24 dump.go:134] &#x9;Container cinder-csi-plugin ready: true, restart count 0&#xA;I0402 15:03:02.306713 24 dump.go:134] &#x9;Container liveness-probe ready: true, restart count 0&#xA;I0402 15:03:02.306782 24 dump.go:134] &#x9;Container node-driver-registrar ready: true, restart count 1&#xA;I0402 15:03:02.390464 24 kubelet_metrics.go:206] &#xA;Latency metrics for node kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67&#xA;I0402 15:03:02.390492 24 dump.go:109] &#xA;Logging node info for node kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb&#xA;I0402 15:03:02.567524 24 dump.go:114] Node Info: &amp;Node{ObjectMeta:{kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb    97fd5667-e054-41c9-bc53-42a7e67d1296 1906 0 2026-04-02 14:57:58 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:m1.large beta.kubernetes.io/os:linux failure-domain.beta.kubernetes.io/region:RegionOne failure-domain.beta.kubernetes.io/zone:nova kubernetes.io/arch:amd64 kubernetes.io/hostname:kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb kubernetes.io/os:linux node-role.kubernetes.io/worker: node.cluster.x-k8s.io/nodegroup:default-worker node.kubernetes.io/instance-type:m1.large topology.cinder.csi.openstack.org/zone:nova topology.kubernetes.io/region:RegionOne topology.kubernetes.io/zone:nova] map[cluster.x-k8s.io/annotations-from-machine: cluster.x-k8s.io/cluster-name:kube-uhjhf cluster.x-k8s.io/cluster-namespace:magnum-system cluster.x-k8s.io/labels-from-machine:node-role.kubernetes.io/worker,node.cluster.x-k8s.io/nodegroup cluster.x-k8s.io/machine:kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb cluster.x-k8s.io/owner-kind:MachineSet cluster.x-k8s.io/owner-name:kube-uhjhf-default-worker-6qhl5-ctzr5 csi.volume.kubernetes.io/nodeid:{&#34;cinder.csi.openstack.org&#34;:&#34;e34f1b49-5b09-4eee-ad8b-e5dc8ebe2bdc&#34;,&#34;nfs.csi.k8s.io&#34;:&#34;kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb&#34;,&#34;nfs.manila.csi.openstack.org&#34;:&#34;e34f1b49-5b09-4eee-ad8b-e5dc8ebe2bdc&#34;} kubeadm.alpha.kubernetes.io/cri-socket:unix:///var/run/containerd/containerd.sock node.alpha.kubernetes.io/ttl:0 projectcalico.org/IPv4Address:192.168.24.48/24 projectcalico.org/IPv4IPIPTunnelAddr:10.100.8.64 projectcalico.org/Interfaces:[{&#34;name&#34;:&#34;ens3&#34;,&#34;addresses&#34;:[&#34;192.168.24.48&#34;]}] volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubeadm Update v1 2026-04-02 14:57:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}} } {kubelet Update v1 2026-04-02 14:57:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}}} } {kube-controller-manager Update v1 2026-04-02 14:59:59 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.100.2.0/24\&#34;&#34;:{}}}} } {openstack-cloud-controller-manager Update v1 2026-04-02 15:00:54 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:beta.kubernetes.io/instance-type&#34;:{},&#34;f:failure-domain.beta.kubernetes.io/region&#34;:{},&#34;f:failure-domain.beta.kubernetes.io/zone&#34;:{},&#34;f:node.kubernetes.io/instance-type&#34;:{},&#34;f:topology.kubernetes.io/region&#34;:{},&#34;f:topology.kubernetes.io/zone&#34;:{}}},&#34;f:spec&#34;:{&#34;f:providerID&#34;:{}}} } {manager Update v1 2026-04-02 15:00:56 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:cluster.x-k8s.io/annotations-from-machine&#34;:{},&#34;f:cluster.x-k8s.io/cluster-name&#34;:{},&#34;f:cluster.x-k8s.io/cluster-namespace&#34;:{},&#34;f:cluster.x-k8s.io/labels-from-machine&#34;:{},&#34;f:cluster.x-k8s.io/machine&#34;:{},&#34;f:cluster.x-k8s.io/owner-kind&#34;:{},&#34;f:cluster.x-k8s.io/owner-name&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/worker&#34;:{},&#34;f:node.cluster.x-k8s.io/nodegroup&#34;:{}}}} } {openstack-cloud-controller-manager Update v1 2026-04-02 15:00:57 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}}}} status} {calico-node Update v1 2026-04-02 15:00:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:projectcalico.org/IPv4Address&#34;:{},&#34;f:projectcalico.org/IPv4IPIPTunnelAddr&#34;:{},&#34;f:projectcalico.org/Interfaces&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}} status} {kubelet Update v1 2026-04-02 15:03:01 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:csi.volume.kubernetes.io/nodeid&#34;:{}},&#34;f:labels&#34;:{&#34;f:topology.cinder.csi.openstack.org/zone&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}} status}]},Spec:NodeSpec{PodCIDR:10.100.2.0/24,DoNotUseExternalID:,ProviderID:openstack:///e34f1b49-5b09-4eee-ad8b-e5dc8ebe2bdc,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.100.2.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{4 0} {&lt;nil&gt;} 4 DecimalSI},ephemeral-storage: {{80514002944 0} {&lt;nil&gt;} 78626956Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{8322891776 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{4 0} {&lt;nil&gt;} 4 DecimalSI},ephemeral-storage: {{72462602530 0} {&lt;nil&gt;} 72462602530 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{8218034176 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2026-04-02 15:00:58 +0000 UTC,LastTransitionTime:2026-04-02 15:00:58 +0000 UTC,Reason:CalicoIsUp,Message:Calico is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2026-04-02 15:03:01 +0000 UTC,LastTransitionTime:2026-04-02 14:57:58 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2026-04-02 15:03:01 +0000 UTC,LastTransitionTime:2026-04-02 14:57:58 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2026-04-02 15:03:01 +0000 UTC,LastTransitionTime:2026-04-02 14:57:58 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2026-04-02 15:03:01 +0000 UTC,LastTransitionTime:2026-04-02 14:59:33 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:192.168.24.48,},NodeAddress{Type:Hostname,Address:kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:e34f1b495b094eeead8be5dc8ebe2bdc,SystemUUID:e34f1b49-5b09-4eee-ad8b-e5dc8ebe2bdc,BootID:d53fd5fa-8442-4687-a507-9d87dce39ea3,KernelVersion:5.15.0-174-generic,OSImage:Ubuntu 22.04.5 LTS,ContainerRuntimeVersion:containerd://2.2.2,KubeletVersion:v1.33.10,KubeProxyVersion:,OperatingSystem:linux,Architecture:amd64,Swap:nil,},Images:[]ContainerImage{ContainerImage{Names:[quay.io/calico/node@sha256:f2339c4ff3a57228cbc39a1f67ab81abded1997d843e0e0b1e86664c7c4eb6c0 quay.io/calico/node:v3.31.3],SizeBytes:159483847,},ContainerImage{Names:[registry.k8s.io/conformance@sha256:ae51c08b88bf1f5c96c0cc6a4751ff62ba8267ba3e3e2774d193655f950e20b1 registry.k8s.io/conformance:v1.33.10],SizeBytes:89682008,},ContainerImage{Names:[quay.io/calico/cni@sha256:a6ddb6d00a53f98c5392eb3b525cf00af10cace1f6ca9414230ff652e6700e78 quay.io/calico/cni:v3.31.3],SizeBytes:72141081,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfsplugin@sha256:1d9f524c0fd5d714cf274a0ddf5dceebeb1848349de5687e530c33dd2675e4ba registry.k8s.io/sig-storage/nfsplugin:v4.2.0],SizeBytes:49783016,},ContainerImage{Names:[registry.k8s.io/kube-proxy@sha256:e8151e38ef22f032dba686cc1bba5a3e525dedbe2d549fa44e653fe79426e261 registry.k8s.io/kube-proxy:v1.33.10],SizeBytes:31827782,},ContainerImage{Names:[registry.k8s.io/kube-apiserver@sha256:bbff81e41af4bfca88a1d05a066a48e12e2689c534d073a8c688e3ad6c8701e3 registry.k8s.io/kube-apiserver:v1.33.10],SizeBytes:29986018,},ContainerImage{Names:[registry.k8s.io/provider-os/cinder-csi-plugin@sha256:8dbc1d5c42c143d26ba7fea5ca36a38ab512983b7df16329b486421615381f96 registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0],SizeBytes:28718193,},ContainerImage{Names:[registry.k8s.io/kube-controller-manager@sha256:b0880d6ee19f2b9148d3d37008c5ee9fc73976e8edad4d0709f11d32ab3ee709 registry.k8s.io/kube-controller-manager:v1.33.10],SizeBytes:27552094,},ContainerImage{Names:[registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6 registry.k8s.io/etcd:3.5.24-0],SizeBytes:23716032,},ContainerImage{Names:[registry.k8s.io/kube-scheduler@sha256:dc1a1aec3bb0ed126b1adff795935124f719969356b24a159fc1a2a0883b89bc registry.k8s.io/kube-scheduler:v1.33.10],SizeBytes:21692956,},ContainerImage{Names:[registry.k8s.io/provider-os/manila-csi-plugin@sha256:2cae8c83c215c1f038f50f007d40151d3651a511214827e6dd409b2a919fdb97 registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0],SizeBytes:21014294,},ContainerImage{Names:[registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97 registry.k8s.io/coredns/coredns:v1.12.0],SizeBytes:20939036,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:33692aed26aaf105b4d6e66280cceca9e0463f500c81b5d8c955428a75438f32 registry.k8s.io/sig-storage/livenessprobe:v2.14.0],SizeBytes:14311007,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.12.0],SizeBytes:14038309,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a13bff2ed69af0cf4270f0cf47bdedf75a56c095cd95b91195ae6c713a9b1845 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.2],SizeBytes:10074312,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:cacee2b5c36dd59d4c7e8469c05c9e4ef53ecb2df9025fa8c10cdaf61bce62f0 registry.k8s.io/sig-storage/livenessprobe:v2.8.0],SizeBytes:8892463,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:a9155b13325b2abef48e71de77bb8ac015412a566829f621d06bfae5c699b1b9 registry.k8s.io/e2e-test-images/busybox:1.36.1-1],SizeBytes:2223659,},ContainerImage{Names:[registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097 registry.k8s.io/pause:3.9],SizeBytes:321520,},ContainerImage{Names:[registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a registry.k8s.io/pause:3.10],SizeBytes:320368,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,RuntimeHandlers:[]NodeRuntimeHandler{NodeRuntimeHandler{Name:,Features:&amp;NodeRuntimeHandlerFeatures{RecursiveReadOnlyMounts:*true,UserNamespaces:*true,},},NodeRuntimeHandler{Name:runc,Features:&amp;NodeRuntimeHandlerFeatures{RecursiveReadOnlyMounts:*true,UserNamespaces:*true,},},},Features:&amp;NodeFeatures{SupplementalGroupsPolicy:*true,},},}&#xA;I0402 15:03:02.567580 24 dump.go:116] &#xA;Logging kubelet events for node kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb&#xA;I0402 15:03:02.581347 24 dump.go:121] &#xA;Logging pods the kubelet thinks are on node kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb&#xA;I0402 15:03:02.600928 24 dump.go:128] kube-system/csi-nfs-node-jtj8j started at 2026-04-02 14:58:00 +0000 UTC (0+3 container statuses recorded)&#xA;I0402 15:03:02.600972 24 dump.go:134] &#x9;Container liveness-probe ready: true, restart count 0&#xA;I0402 15:03:02.600979 24 dump.go:134] &#x9;Container nfs ready: true, restart count 0&#xA;I0402 15:03:02.600987 24 dump.go:134] &#x9;Container node-driver-registrar ready: true, restart count 0&#xA;I0402 15:03:02.600995 24 dump.go:128] kube-system/kube-proxy-8pc5r started at 2026-04-02 14:58:00 +0000 UTC (0+1 container statuses recorded)&#xA;I0402 15:03:02.601006 24 dump.go:134] &#x9;Container kube-proxy ready: true, restart count 0&#xA;I0402 15:03:02.601014 24 dump.go:128] kube-system/manila-csi-openstack-manila-csi-nodeplugin-66z9w started at 2026-04-02 14:58:00 +0000 UTC (0+2 container statuses recorded)&#xA;I0402 15:03:02.601021 24 dump.go:134] &#x9;Container nfs-nodeplugin ready: true, restart count 4&#xA;I0402 15:03:02.601028 24 dump.go:134] &#x9;Container nfs-registrar ready: true, restart count 3&#xA;I0402 15:03:02.601035 24 dump.go:128] kube-system/openstack-cinder-csi-nodeplugin-z67qx started at 2026-04-02 14:58:00 +0000 UTC (0+3 container statuses recorded)&#xA;I0402 15:03:02.601041 24 dump.go:134] &#x9;Container cinder-csi-plugin ready: true, restart count 0&#xA;I0402 15:03:02.601047 24 dump.go:134] &#x9;Container liveness-probe ready: true, restart count 0&#xA;I0402 15:03:02.601052 24 dump.go:134] &#x9;Container node-driver-registrar ready: true, restart count 1&#xA;I0402 15:03:02.601059 24 dump.go:128] kube-system/calico-node-6wvr4 started at 2026-04-02 14:58:00 +0000 UTC (3+1 container statuses recorded)&#xA;I0402 15:03:02.601066 24 dump.go:130] &#x9;Init container upgrade-ipam ready: true, restart count 0&#xA;I0402 15:03:02.601072 24 dump.go:130] &#x9;Init container install-cni ready: true, restart count 1&#xA;I0402 15:03:02.601080 24 dump.go:130] &#x9;Init container ebpf-bootstrap ready: true, restart count 0&#xA;I0402 15:03:02.601085 24 dump.go:134] &#x9;Container calico-node ready: true, restart count 0&#xA;I0402 15:03:02.601092 24 dump.go:128] conformance/e2e-conformance-test started at 2026-04-02 15:02:07 +0000 UTC (0+2 container statuses recorded)&#xA;I0402 15:03:02.601098 24 dump.go:134] &#x9;Container conformance-container ready: true, restart count 0&#xA;I0402 15:03:02.601157 24 dump.go:134] &#x9;Container output-container ready: true, restart count 0&#xA;I0402 15:03:02.685161 24 kubelet_metrics.go:206] &#xA;Latency metrics for node kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb&#xA;I0402 15:03:02.700907 24 kubectl_utils.go:109] Running kubectl logs on non-ready containers in kube-system&#xA;I0402 15:03:02.721589 24 kubectl_utils.go:128] Logs of kube-system/kube-controller-manager-kube-uhjhf-67cxd-gq627:kube-controller-manager on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:02.721695 24 kubectl_utils.go:129]  : STARTLOG&#xA;I0402 15:01:42.566331       1 serving.go:386] Generated self-signed cert in-memory&#xA;I0402 15:01:43.353763       1 controllermanager.go:188] &#34;Starting&#34; version=&#34;v1.33.10&#34;&#xA;I0402 15:01:43.353787       1 controllermanager.go:190] &#34;Golang settings&#34; GOGC=&#34;&#34; GOMAXPROCS=&#34;&#34; GOTRACEBACK=&#34;&#34;&#xA;I0402 15:01:43.355046       1 dynamic_cafile_content.go:161] &#34;Starting controller&#34; name=&#34;request-header::/etc/kubernetes/pki/front-proxy-ca.crt&#34;&#xA;I0402 15:01:43.355454       1 dynamic_cafile_content.go:161] &#34;Starting controller&#34; name=&#34;client-ca-bundle::/etc/kubernetes/pki/ca.crt&#34;&#xA;I0402 15:01:43.357209       1 secure_serving.go:211] Serving securely on [::]:10257&#xA;I0402 15:01:43.357352       1 tlsconfig.go:243] &#34;Starting DynamicServingCertificateController&#34;&#xA;I0402 15:01:43.358216       1 leaderelection.go:257] attempting to acquire leader lease kube-system/kube-controller-manager...&#xA;I0402 15:02:02.115245       1 leaderelection.go:271] successfully acquired lease kube-system/kube-controller-manager&#xA;I0402 15:02:02.116024       1 event.go:389] &#34;Event occurred&#34; object=&#34;kube-system/kube-controller-manager&#34; fieldPath=&#34;&#34; kind=&#34;Lease&#34; apiVersion=&#34;coordination.k8s.io/v1&#34; type=&#34;Normal&#34; reason=&#34;LeaderElection&#34; message=&#34;kube-uhjhf-67cxd-gq627_4c52e7ba-b01b-47a1-bb44-7b6cc932fa9c became leader&#34;&#xA;I0402 15:02:02.222680       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;tokens&#34;&#xA;I0402 15:02:02.223462       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;serviceaccount-token-controller&#34;&#xA;I0402 15:02:02.231536       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;disruption-controller&#34;&#xA;I0402 15:02:02.231655       1 disruption.go:455] &#34;Sending events to api server.&#34; logger=&#34;disruption-controller&#34;&#xA;I0402 15:02:02.231697       1 disruption.go:466] &#34;Starting disruption controller&#34; logger=&#34;disruption-controller&#34;&#xA;I0402 15:02:02.231703       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;disruption&#34;&#xA;I0402 15:02:02.307621       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;bootstrap-signer-controller&#34;&#xA;I0402 15:02:02.308190       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;bootstrap_signer&#34;&#xA;I0402 15:02:02.324133       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;tokens&#34;&#xA;I0402 15:02:02.382690       1 range_allocator.go:112] &#34;No Secondary Service CIDR provided. Skipping filtering out secondary service addresses&#34; logger=&#34;node-ipam-controller&#34;&#xA;I0402 15:02:02.382849       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;node-ipam-controller&#34;&#xA;I0402 15:02:02.383301       1 node_ipam_controller.go:141] &#34;Starting ipam controller&#34; logger=&#34;node-ipam-controller&#34;&#xA;I0402 15:02:02.383363       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;node&#34;&#xA;I0402 15:02:02.389461       1 node_lifecycle_controller.go:419] &#34;Controller will reconcile labels&#34; logger=&#34;node-lifecycle-controller&#34;&#xA;I0402 15:02:02.389600       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;node-lifecycle-controller&#34;&#xA;I0402 15:02:02.389731       1 node_lifecycle_controller.go:453] &#34;Sending events to api server&#34; logger=&#34;node-lifecycle-controller&#34;&#xA;I0402 15:02:02.389808       1 node_lifecycle_controller.go:464] &#34;Starting node controller&#34; logger=&#34;node-lifecycle-controller&#34;&#xA;I0402 15:02:02.389858       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;taint&#34;&#xA;I0402 15:02:02.392883       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;ttl-after-finished-controller&#34;&#xA;I0402 15:02:02.393295       1 ttlafterfinished_controller.go:112] &#34;Starting TTL after finished controller&#34; logger=&#34;ttl-after-finished-controller&#34;&#xA;I0402 15:02:02.393596       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;TTL after finished&#34;&#xA;I0402 15:02:02.396056       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;pod-garbage-collector-controller&#34;&#xA;I0402 15:02:02.396285       1 gc_controller.go:99] &#34;Starting GC controller&#34; logger=&#34;pod-garbage-collector-controller&#34;&#xA;I0402 15:02:02.396387       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;GC&#34;&#xA;I0402 15:02:02.398871       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;statefulset-controller&#34;&#xA;I0402 15:02:02.399069       1 stateful_set.go:166] &#34;Starting stateful set controller&#34; logger=&#34;statefulset-controller&#34;&#xA;I0402 15:02:02.399218       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;stateful set&#34;&#xA;I0402 15:02:02.402267       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;persistentvolume-attach-detach-controller&#34;&#xA;I0402 15:02:02.402641       1 attach_detach_controller.go:338] &#34;Starting attach detach controller&#34; logger=&#34;persistentvolume-attach-detach-controller&#34;&#xA;I0402 15:02:02.402680       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;attach detach&#34;&#xA;I0402 15:02:02.405275       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;root-ca-certificate-publisher-controller&#34;&#xA;I0402 15:02:02.405640       1 publisher.go:107] &#34;Starting root CA cert publisher controller&#34; logger=&#34;root-ca-certificate-publisher-controller&#34;&#xA;I0402 15:02:02.405725       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;crt configmap&#34;&#xA;I0402 15:02:02.407564       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;ephemeral-volume-controller&#34;&#xA;I0402 15:02:02.407601       1 controller.go:173] &#34;Starting ephemeral volume controller&#34; logger=&#34;ephemeral-volume-controller&#34;&#xA;I0402 15:02:02.407614       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;ephemeral&#34;&#xA;I0402 15:02:02.407599       1 controllermanager.go:730] &#34;Controller is disabled by a feature gate&#34; controller=&#34;resourceclaim-controller&#34; requiredFeatureGates=[&#34;DynamicResourceAllocation&#34;]&#xA;I0402 15:02:02.454183       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;validatingadmissionpolicy-status-controller&#34;&#xA;I0402 15:02:02.454458       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;validatingadmissionpolicy-status&#34;&#xA;I0402 15:02:02.466552       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;certificatesigningrequest-signing-controller&#34;&#xA;I0402 15:02:02.466724       1 certificate_controller.go:120] &#34;Starting certificate controller&#34; logger=&#34;certificatesigningrequest-signing-controller&#34; name=&#34;csrsigning-kubelet-serving&#34;&#xA;I0402 15:02:02.466735       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;certificate-csrsigning-kubelet-serving&#34;&#xA;I0402 15:02:02.466770       1 certificate_controller.go:120] &#34;Starting certificate controller&#34; logger=&#34;certificatesigningrequest-signing-controller&#34; name=&#34;csrsigning-kubelet-client&#34;&#xA;I0402 15:02:02.466776       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;certificate-csrsigning-kubelet-client&#34;&#xA;I0402 15:02:02.466799       1 certificate_controller.go:120] &#34;Starting certificate controller&#34; logger=&#34;certificatesigningrequest-signing-controller&#34; name=&#34;csrsigning-kube-apiserver-client&#34;&#xA;I0402 15:02:02.466804       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;certificate-csrsigning-kube-apiserver-client&#34;&#xA;I0402 15:02:02.466843       1 certificate_controller.go:120] &#34;Starting certificate controller&#34; logger=&#34;certificatesigningrequest-signing-controller&#34; name=&#34;csrsigning-legacy-unknown&#34;&#xA;I0402 15:02:02.466848       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;certificate-csrsigning-legacy-unknown&#34;&#xA;I0402 15:02:02.466864       1 dynamic_serving_content.go:135] &#34;Starting controller&#34; name=&#34;csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key&#34;&#xA;I0402 15:02:02.467067       1 dynamic_serving_content.go:135] &#34;Starting controller&#34; name=&#34;csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key&#34;&#xA;I0402 15:02:02.467133       1 dynamic_serving_content.go:135] &#34;Starting controller&#34; name=&#34;csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key&#34;&#xA;I0402 15:02:02.467202       1 dynamic_serving_content.go:135] &#34;Starting controller&#34; name=&#34;csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key&#34;&#xA;I0402 15:02:02.473424       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;certificatesigningrequest-approving-controller&#34;&#xA;I0402 15:02:02.473944       1 certificate_controller.go:120] &#34;Starting certificate controller&#34; logger=&#34;certificatesigningrequest-approving-controller&#34; name=&#34;csrapproving&#34;&#xA;I0402 15:02:02.474057       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;certificate-csrapproving&#34;&#xA;I0402 15:02:02.579969       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;persistentvolume-protection-controller&#34;&#xA;I0402 15:02:02.580572       1 pv_protection_controller.go:81] &#34;Starting PV protection controller&#34; logger=&#34;persistentvolume-protection-controller&#34;&#xA;I0402 15:02:02.580697       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;PV protection&#34;&#xA;I0402 15:02:02.586240       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;endpointslice-controller&#34;&#xA;I0402 15:02:02.586861       1 endpointslice_controller.go:281] &#34;Starting endpoint slice controller&#34; logger=&#34;endpointslice-controller&#34;&#xA;I0402 15:02:02.586932       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;endpoint_slice&#34;&#xA;I0402 15:02:02.590504       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;endpointslice-mirroring-controller&#34;&#xA;I0402 15:02:02.591056       1 endpointslicemirroring_controller.go:227] &#34;Starting EndpointSliceMirroring controller&#34; logger=&#34;endpointslice-mirroring-controller&#34;&#xA;I0402 15:02:02.591115       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;endpoint_slice_mirroring&#34;&#xA;I0402 15:02:02.604266       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;garbage-collector-controller&#34;&#xA;I0402 15:02:02.604500       1 garbagecollector.go:144] &#34;Starting controller&#34; logger=&#34;garbage-collector-controller&#34; controller=&#34;garbagecollector&#34;&#xA;I0402 15:02:02.604547       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;garbage collector&#34;&#xA;I0402 15:02:02.604634       1 graph_builder.go:351] &#34;Running&#34; logger=&#34;garbage-collector-controller&#34; component=&#34;GraphBuilder&#34;&#xA;I0402 15:02:02.609483       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;job-controller&#34;&#xA;I0402 15:02:02.610254       1 job_controller.go:243] &#34;Starting job controller&#34; logger=&#34;job-controller&#34;&#xA;I0402 15:02:02.610289       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;job&#34;&#xA;I0402 15:02:02.616461       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;replicaset-controller&#34;&#xA;I0402 15:02:02.617135       1 replica_set.go:219] &#34;Starting controller&#34; logger=&#34;replicaset-controller&#34; name=&#34;replicaset&#34;&#xA;I0402 15:02:02.617401       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;ReplicaSet&#34;&#xA;I0402 15:02:02.628183       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;horizontal-pod-autoscaler-controller&#34;&#xA;I0402 15:02:02.628475       1 horizontal.go:204] &#34;Starting HPA controller&#34; logger=&#34;horizontal-pod-autoscaler-controller&#34;&#xA;I0402 15:02:02.628537       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;HPA&#34;&#xA;I0402 15:02:02.630697       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;certificatesigningrequest-cleaner-controller&#34;&#xA;I0402 15:02:02.630816       1 cleaner.go:83] &#34;Starting CSR cleaner controller&#34; logger=&#34;certificatesigningrequest-cleaner-controller&#34;&#xA;I0402 15:02:02.632822       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;ttl-controller&#34;&#xA;I0402 15:02:02.633068       1 ttl_controller.go:127] &#34;Starting TTL controller&#34; logger=&#34;ttl-controller&#34;&#xA;I0402 15:02:02.634117       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;TTL&#34;&#xA;I0402 15:02:02.650899       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;endpoints&#34;&#xA;I0402 15:02:02.651191       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;networkpolicies.networking.k8s.io&#34;&#xA;I0402 15:02:02.651365       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;rolebindings.rbac.authorization.k8s.io&#34;&#xA;I0402 15:02:02.651481       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;endpointslices.discovery.k8s.io&#34;&#xA;I0402 15:02:02.651539       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;deployments.apps&#34;&#xA;I0402 15:02:02.651616       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;csistoragecapacities.storage.k8s.io&#34;&#xA;I0402 15:02:02.651844       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;stagedkubernetesnetworkpolicies.crd.projectcalico.org&#34;&#xA;I0402 15:02:02.651961       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;podtemplates&#34;&#xA;I0402 15:02:02.652132       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;jobs.batch&#34;&#xA;I0402 15:02:02.652226       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;leases.coordination.k8s.io&#34;&#xA;I0402 15:02:02.652395       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;replicasets.apps&#34;&#xA;I0402 15:02:02.652486       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;daemonsets.apps&#34;&#xA;I0402 15:02:02.652526       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;cronjobs.batch&#34;&#xA;I0402 15:02:02.652595       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;ingresses.networking.k8s.io&#34;&#xA;I0402 15:02:02.652686       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;horizontalpodautoscalers.autoscaling&#34;&#xA;I0402 15:02:02.652728       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;controllerrevisions.apps&#34;&#xA;I0402 15:02:02.652766       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;statefulsets.apps&#34;&#xA;I0402 15:02:02.652859       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;stagednetworkpolicies.crd.projectcalico.org&#34;&#xA;I0402 15:02:02.652925       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;limitranges&#34;&#xA;I0402 15:02:02.653067       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;roles.rbac.authorization.k8s.io&#34;&#xA;I0402 15:02:02.653189       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;networksets.crd.projectcalico.org&#34;&#xA;I0402 15:02:02.653245       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;networkpolicies.crd.projectcalico.org&#34;&#xA;I0402 15:02:02.653311       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;poddisruptionbudgets.policy&#34;&#xA;I0402 15:02:02.653655       1 resource_quota_monitor.go:228] &#34;QuotaMonitor created object count evaluator&#34; logger=&#34;resourcequota-controller&#34; resource=&#34;serviceaccounts&#34;&#xA;I0402 15:02:02.653750       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;resourcequota-controller&#34;&#xA;I0402 15:02:02.653863       1 resource_quota_controller.go:300] &#34;Starting resource quota controller&#34; logger=&#34;resourcequota-controller&#34;&#xA;I0402 15:02:02.653877       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;resource quota&#34;&#xA;I0402 15:02:02.653898       1 resource_quota_monitor.go:309] &#34;QuotaMonitor running&#34; logger=&#34;resourcequota-controller&#34;&#xA;I0402 15:02:02.659602       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;serviceaccount-controller&#34;&#xA;I0402 15:02:02.659640       1 serviceaccounts_controller.go:114] &#34;Starting service account controller&#34; logger=&#34;serviceaccount-controller&#34;&#xA;I0402 15:02:02.659645       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;service account&#34;&#xA;I0402 15:02:02.970505       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;cronjob-controller&#34;&#xA;I0402 15:02:02.970575       1 controllermanager.go:736] &#34;Skipping a cloud provider controller&#34; controller=&#34;node-route-controller&#34;&#xA;I0402 15:02:02.971089       1 cronjob_controllerv2.go:145] &#34;Starting cronjob controller v2&#34; logger=&#34;cronjob-controller&#34;&#xA;I0402 15:02:02.971116       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;cronjob&#34;&#xA;I0402 15:02:02.975458       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;persistentvolume-expander-controller&#34;&#xA;I0402 15:02:02.975888       1 expand_controller.go:329] &#34;Starting expand controller&#34; logger=&#34;persistentvolume-expander-controller&#34;&#xA;I0402 15:02:02.976122       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;expand&#34;&#xA;I0402 15:02:02.981598       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;clusterrole-aggregation-controller&#34;&#xA;I0402 15:02:02.981619       1 controllermanager.go:730] &#34;Controller is disabled by a feature gate&#34; controller=&#34;volumeattributesclass-protection-controller&#34; requiredFeatureGates=[&#34;VolumeAttributesClass&#34;]&#xA;I0402 15:02:02.981639       1 controllermanager.go:730] &#34;Controller is disabled by a feature gate&#34; controller=&#34;storageversion-garbage-collector-controller&#34; requiredFeatureGates=[&#34;APIServerIdentity&#34;,&#34;StorageVersionAPI&#34;]&#xA;I0402 15:02:02.981874       1 clusterroleaggregation_controller.go:194] &#34;Starting ClusterRoleAggregator controller&#34; logger=&#34;clusterrole-aggregation-controller&#34;&#xA;I0402 15:02:02.981931       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;ClusterRoleAggregator&#34;&#xA;I0402 15:02:02.986008       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;persistentvolume-binder-controller&#34;&#xA;I0402 15:02:02.986061       1 controllermanager.go:730] &#34;Controller is disabled by a feature gate&#34; controller=&#34;device-taint-eviction-controller&#34; requiredFeatureGates=[&#34;DynamicResourceAllocation&#34;,&#34;DRADeviceTaints&#34;]&#xA;I0402 15:02:02.986558       1 pv_controller_base.go:308] &#34;Starting persistent volume controller&#34; logger=&#34;persistentvolume-binder-controller&#34;&#xA;I0402 15:02:02.986597       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;persistent volume&#34;&#xA;I0402 15:02:02.990822       1 legacy_serviceaccount_token_cleaner.go:103] &#34;Starting legacy service account token cleaner controller&#34; logger=&#34;legacy-serviceaccount-token-cleaner-controller&#34;&#xA;I0402 15:02:02.990910       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;legacy-service-account-token-cleaner&#34;&#xA;I0402 15:02:02.990984       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;legacy-serviceaccount-token-cleaner-controller&#34;&#xA;I0402 15:02:02.994646       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;service-cidr-controller&#34;&#xA;I0402 15:02:02.994690       1 controllermanager.go:756] &#34;Warning: skipping controller&#34; controller=&#34;storage-version-migrator-controller&#34;&#xA;I0402 15:02:02.994978       1 servicecidrs_controller.go:136] &#34;Starting&#34; logger=&#34;service-cidr-controller&#34; controller=&#34;service-cidr-controller&#34;&#xA;I0402 15:02:02.995043       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;service-cidr-controller&#34;&#xA;I0402 15:02:03.093086       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;namespace-controller&#34;&#xA;I0402 15:02:03.093168       1 namespace_controller.go:202] &#34;Starting namespace controller&#34; logger=&#34;namespace-controller&#34;&#xA;I0402 15:02:03.093179       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;namespace&#34;&#xA;I0402 15:02:03.131443       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;deployment-controller&#34;&#xA;I0402 15:02:03.131493       1 controllermanager.go:736] &#34;Skipping a cloud provider controller&#34; controller=&#34;service-lb-controller&#34;&#xA;I0402 15:02:03.131509       1 controllermanager.go:730] &#34;Controller is disabled by a feature gate&#34; controller=&#34;kube-apiserver-serving-clustertrustbundle-publisher-controller&#34; requiredFeatureGates=[&#34;ClusterTrustBundle&#34;]&#xA;I0402 15:02:03.132104       1 deployment_controller.go:173] &#34;Starting controller&#34; logger=&#34;deployment-controller&#34; controller=&#34;deployment&#34;&#xA;I0402 15:02:03.132172       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;deployment&#34;&#xA;I0402 15:02:03.180798       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;taint-eviction-controller&#34;&#xA;I0402 15:02:03.180957       1 taint_eviction.go:282] &#34;Starting&#34; logger=&#34;taint-eviction-controller&#34; controller=&#34;taint-eviction-controller&#34;&#xA;I0402 15:02:03.181706       1 taint_eviction.go:288] &#34;Sending events to api server&#34; logger=&#34;taint-eviction-controller&#34;&#xA;I0402 15:02:03.182046       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;taint-eviction-controller&#34;&#xA;I0402 15:02:03.229885       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;replicationcontroller-controller&#34;&#xA;I0402 15:02:03.230223       1 replica_set.go:219] &#34;Starting controller&#34; logger=&#34;replicationcontroller-controller&#34; name=&#34;replicationcontroller&#34;&#xA;I0402 15:02:03.230264       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;ReplicationController&#34;&#xA;I0402 15:02:03.280589       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;token-cleaner-controller&#34;&#xA;I0402 15:02:03.280661       1 controllermanager.go:736] &#34;Skipping a cloud provider controller&#34; controller=&#34;cloud-node-lifecycle-controller&#34;&#xA;I0402 15:02:03.281005       1 tokencleaner.go:117] &#34;Starting token cleaner controller&#34; logger=&#34;token-cleaner-controller&#34;&#xA;I0402 15:02:03.281263       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;token_cleaner&#34;&#xA;I0402 15:02:03.281826       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;token_cleaner&#34;&#xA;I0402 15:02:03.329117       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;persistentvolumeclaim-protection-controller&#34;&#xA;I0402 15:02:03.329228       1 controllermanager.go:741] &#34;Warning: controller is disabled&#34; controller=&#34;selinux-warning-controller&#34;&#xA;I0402 15:02:03.329765       1 pvc_protection_controller.go:168] &#34;Starting PVC protection controller&#34; logger=&#34;persistentvolumeclaim-protection-controller&#34;&#xA;I0402 15:02:03.329808       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;PVC protection&#34;&#xA;I0402 15:02:03.380698       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;endpoints-controller&#34;&#xA;I0402 15:02:03.381608       1 endpoints_controller.go:187] &#34;Starting endpoint controller&#34; logger=&#34;endpoints-controller&#34;&#xA;I0402 15:02:03.381667       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;endpoint&#34;&#xA;I0402 15:02:03.431114       1 controllermanager.go:778] &#34;Started controller&#34; controller=&#34;daemonset-controller&#34;&#xA;I0402 15:02:03.433753       1 daemon_controller.go:310] &#34;Starting daemon sets controller&#34; logger=&#34;daemonset-controller&#34;&#xA;I0402 15:02:03.433811       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;daemon sets&#34;&#xA;I0402 15:02:03.453141       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;resource quota&#34;&#xA;I0402 15:02:03.466899       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;certificate-csrsigning-legacy-unknown&#34;&#xA;I0402 15:02:03.472097       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;certificate-csrsigning-kubelet-client&#34;&#xA;I0402 15:02:03.472722       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;certificate-csrsigning-kubelet-serving&#34;&#xA;I0402 15:02:03.473389       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;certificate-csrsigning-kube-apiserver-client&#34;&#xA;I0402 15:02:03.474274       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;certificate-csrapproving&#34;&#xA;I0402 15:02:03.476572       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;expand&#34;&#xA;I0402 15:02:03.483539       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;garbage collector&#34;&#xA;I0402 15:02:03.491390       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;endpoint_slice_mirroring&#34;&#xA;I0402 15:02:03.494089       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;namespace&#34;&#xA;I0402 15:02:03.494210       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;TTL after finished&#34;&#xA;I0402 15:02:03.495111       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;service-cidr-controller&#34;&#xA;I0402 15:02:03.506190       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;crt configmap&#34;&#xA;I0402 15:02:03.508479       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;bootstrap_signer&#34;&#xA;I0402 15:02:03.518697       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;ReplicaSet&#34;&#xA;I0402 15:02:03.529661       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;HPA&#34;&#xA;I0402 15:02:03.530147       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;PVC protection&#34;&#xA;I0402 15:02:03.530850       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;ReplicationController&#34;&#xA;I0402 15:02:03.532510       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;deployment&#34;&#xA;I0402 15:02:03.554885       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;validatingadmissionpolicy-status&#34;&#xA;I0402 15:02:03.560858       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;service account&#34;&#xA;I0402 15:02:03.582039       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;endpoint&#34;&#xA;I0402 15:02:03.594187       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;legacy-service-account-token-cleaner&#34;&#xA;I0402 15:02:03.600055       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;stateful set&#34;&#xA;I0402 15:02:03.607900       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;ephemeral&#34;&#xA;I0402 15:02:03.611843       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;job&#34;&#xA;I0402 15:02:03.632556       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;disruption&#34;&#xA;I0402 15:02:03.654624       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;resource quota&#34;&#xA;I0402 15:02:03.654835       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;resource quota&#34;&#xA;I0402 15:02:03.672815       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;cronjob&#34;&#xA;I0402 15:02:03.682202       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;ClusterRoleAggregator&#34;&#xA;I0402 15:02:03.781854       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;PV protection&#34;&#xA;I0402 15:02:03.788632       1 actual_state_of_world.go:541] &#34;Failed to update statusUpdateNeeded field in actual state of world&#34; logger=&#34;persistentvolume-attach-detach-controller&#34; err=&#34;Failed to set statusUpdateNeeded to needed true, because nodeName=\&#34;kube-uhjhf-67cxd-gq627\&#34; does not exist&#34;&#xA;I0402 15:02:03.788981       1 actual_state_of_world.go:541] &#34;Failed to update statusUpdateNeeded field in actual state of world&#34; logger=&#34;persistentvolume-attach-detach-controller&#34; err=&#34;Failed to set statusUpdateNeeded to needed true, because nodeName=\&#34;kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67\&#34; does not exist&#34;&#xA;I0402 15:02:03.789067       1 actual_state_of_world.go:541] &#34;Failed to update statusUpdateNeeded field in actual state of world&#34; logger=&#34;persistentvolume-attach-detach-controller&#34; err=&#34;Failed to set statusUpdateNeeded to needed true, because nodeName=\&#34;kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb\&#34; does not exist&#34;&#xA;I0402 15:02:03.791183       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;taint&#34;&#xA;I0402 15:02:03.791316       1 node_lifecycle_controller.go:1221] &#34;Initializing eviction metric for zone&#34; logger=&#34;node-lifecycle-controller&#34; zone=&#34;RegionOne:\x00:nova&#34;&#xA;I0402 15:02:03.791474       1 node_lifecycle_controller.go:873] &#34;Missing timestamp for Node. Assuming now as a timestamp&#34; logger=&#34;node-lifecycle-controller&#34; node=&#34;kube-uhjhf-67cxd-gq627&#34;&#xA;I0402 15:02:03.791706       1 node_lifecycle_controller.go:873] &#34;Missing timestamp for Node. Assuming now as a timestamp&#34; logger=&#34;node-lifecycle-controller&#34; node=&#34;kube-uhjhf-default-worker-6qhl5-ctzr5-n9flb&#34;&#xA;I0402 15:02:03.791786       1 node_lifecycle_controller.go:873] &#34;Missing timestamp for Node. Assuming now as a timestamp&#34; logger=&#34;node-lifecycle-controller&#34; node=&#34;kube-uhjhf-default-worker-6qhl5-ctzr5-8lq67&#34;&#xA;I0402 15:02:03.792068       1 node_lifecycle_controller.go:1067] &#34;Controller detected that zone is now in new state&#34; logger=&#34;node-lifecycle-controller&#34; zone=&#34;RegionOne:\x00:nova&#34; newState=&#34;Normal&#34;&#xA;I0402 15:02:03.796617       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;GC&#34;&#xA;I0402 15:02:03.803251       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;attach detach&#34;&#xA;I0402 15:02:03.834299       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;TTL&#34;&#xA;I0402 15:02:03.835673       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;daemon sets&#34;&#xA;I0402 15:02:03.882716       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;taint-eviction-controller&#34;&#xA;I0402 15:02:03.883489       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;node&#34;&#xA;I0402 15:02:03.883559       1 range_allocator.go:177] &#34;Sending events to api server&#34; logger=&#34;node-ipam-controller&#34;&#xA;I0402 15:02:03.883617       1 range_allocator.go:183] &#34;Starting range CIDR allocator&#34; logger=&#34;node-ipam-controller&#34;&#xA;I0402 15:02:03.883650       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;cidrallocator&#34;&#xA;I0402 15:02:03.883660       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;cidrallocator&#34;&#xA;I0402 15:02:03.887050       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;persistent volume&#34;&#xA;I0402 15:02:03.887301       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;endpoint_slice&#34;&#xA;I0402 15:02:04.284247       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;garbage collector&#34;&#xA;I0402 15:02:04.306474       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;garbage collector&#34;&#xA;I0402 15:02:04.306809       1 garbagecollector.go:154] &#34;Garbage collector: all resource monitors have synced&#34; logger=&#34;garbage-collector-controller&#34;&#xA;I0402 15:02:04.307097       1 garbagecollector.go:157] &#34;Proceeding to collect garbage&#34; logger=&#34;garbage-collector-controller&#34;&#xA;E0402 15:02:49.885820       1 leaderelection.go:429] Failed to update lock optimistically: Put &#34;https://192.168.24.167:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=5s&#34;: net/http: request canceled (Client.Timeout exceeded while awaiting headers), falling back to slow path&#xA;E0402 15:02:54.885209       1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get &#34;https://192.168.24.167:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=5s&#34;: context deadline exceeded&#xA;I0402 15:02:54.885321       1 leaderelection.go:297] failed to renew lease kube-system/kube-controller-manager: context deadline exceeded&#xA;E0402 15:02:54.885616       1 controllermanager.go:340] &#34;leaderelection lost&#34;&#xA;&#xA;ENDLOG for container kube-system:kube-controller-manager-kube-uhjhf-67cxd-gq627:kube-controller-manager&#xA;I0402 15:03:02.853394 24 kubectl_utils.go:128] Logs of kube-system/kube-scheduler-kube-uhjhf-67cxd-gq627:kube-scheduler on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:02.853459 24 kubectl_utils.go:129]  : STARTLOG&#xA;I0402 15:01:45.912060       1 serving.go:386] Generated self-signed cert in-memory&#xA;I0402 15:01:46.907276       1 server.go:171] &#34;Starting Kubernetes Scheduler&#34; version=&#34;v1.33.10&#34;&#xA;I0402 15:01:46.907366       1 server.go:173] &#34;Golang settings&#34; GOGC=&#34;&#34; GOMAXPROCS=&#34;&#34; GOTRACEBACK=&#34;&#34;&#xA;I0402 15:01:46.913986       1 requestheader_controller.go:180] Starting RequestHeaderAuthRequestController&#xA;I0402 15:01:46.914046       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;RequestHeaderAuthRequestController&#34;&#xA;I0402 15:01:46.914133       1 configmap_cafile_content.go:205] &#34;Starting controller&#34; name=&#34;client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file&#34;&#xA;I0402 15:01:46.914142       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file&#34;&#xA;I0402 15:01:46.914138       1 configmap_cafile_content.go:205] &#34;Starting controller&#34; name=&#34;client-ca::kube-system::extension-apiserver-authentication::client-ca-file&#34;&#xA;I0402 15:01:46.914163       1 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;client-ca::kube-system::extension-apiserver-authentication::client-ca-file&#34;&#xA;I0402 15:01:46.919581       1 secure_serving.go:211] Serving securely on [::]:10259&#xA;I0402 15:01:46.919742       1 tlsconfig.go:243] &#34;Starting DynamicServingCertificateController&#34;&#xA;I0402 15:01:47.014666       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file&#34;&#xA;I0402 15:01:47.014688       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;RequestHeaderAuthRequestController&#34;&#xA;I0402 15:01:47.014968       1 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;client-ca::kube-system::extension-apiserver-authentication::client-ca-file&#34;&#xA;I0402 15:01:47.021433       1 leaderelection.go:257] attempting to acquire leader lease kube-system/kube-scheduler...&#xA;I0402 15:02:04.832484       1 leaderelection.go:271] successfully acquired lease kube-system/kube-scheduler&#xA;E0402 15:02:49.911117       1 leaderelection.go:429] Failed to update lock optimistically: Put &#34;https://192.168.24.167:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-scheduler?timeout=5s&#34;: context deadline exceeded, falling back to slow path&#xA;E0402 15:02:54.910108       1 leaderelection.go:436] error retrieving resource lock kube-system/kube-scheduler: Get &#34;https://192.168.24.167:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-scheduler?timeout=5s&#34;: context deadline exceeded&#xA;I0402 15:02:54.911398       1 leaderelection.go:297] failed to renew lease kube-system/kube-scheduler: context deadline exceeded&#xA;E0402 15:02:58.845909       1 leaderelection.go:322] Failed to release lock: Operation cannot be fulfilled on leases.coordination.k8s.io &#34;kube-scheduler&#34;: the object has been modified; please apply your changes to the latest version and try again&#xA;E0402 15:02:58.845990       1 server.go:303] &#34;Leaderelection lost&#34;&#xA;&#xA;ENDLOG for container kube-system:kube-scheduler-kube-uhjhf-67cxd-gq627:kube-scheduler&#xA;I0402 15:03:03.250543 24 kubectl_utils.go:125] Failed to get logs of pod manila-csi-openstack-manila-csi-controllerplugin-0, container nfs-provisioner, err: the server rejected our request for an unknown reason (get pods manila-csi-openstack-manila-csi-controllerplugin-0)&#xA;I0402 15:03:03.250587 24 kubectl_utils.go:128] Logs of kube-system/manila-csi-openstack-manila-csi-controllerplugin-0:nfs-provisioner on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:03.250605 24 kubectl_utils.go:129]  : STARTLOG&#xA;&#xA;ENDLOG for container kube-system:manila-csi-openstack-manila-csi-controllerplugin-0:nfs-provisioner&#xA;I0402 15:03:03.653576 24 kubectl_utils.go:125] Failed to get logs of pod manila-csi-openstack-manila-csi-controllerplugin-0, container nfs-snapshotter, err: the server rejected our request for an unknown reason (get pods manila-csi-openstack-manila-csi-controllerplugin-0)&#xA;I0402 15:03:03.653613 24 kubectl_utils.go:128] Logs of kube-system/manila-csi-openstack-manila-csi-controllerplugin-0:nfs-snapshotter on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:03.653693 24 kubectl_utils.go:129]  : STARTLOG&#xA;&#xA;ENDLOG for container kube-system:manila-csi-openstack-manila-csi-controllerplugin-0:nfs-snapshotter&#xA;I0402 15:03:04.053221 24 kubectl_utils.go:125] Failed to get logs of pod manila-csi-openstack-manila-csi-controllerplugin-0, container nfs-resizer, err: the server rejected our request for an unknown reason (get pods manila-csi-openstack-manila-csi-controllerplugin-0)&#xA;I0402 15:03:04.053276 24 kubectl_utils.go:128] Logs of kube-system/manila-csi-openstack-manila-csi-controllerplugin-0:nfs-resizer on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:04.053289 24 kubectl_utils.go:129]  : STARTLOG&#xA;&#xA;ENDLOG for container kube-system:manila-csi-openstack-manila-csi-controllerplugin-0:nfs-resizer&#xA;I0402 15:03:04.449595 24 kubectl_utils.go:125] Failed to get logs of pod manila-csi-openstack-manila-csi-controllerplugin-0, container nfs-nodeplugin, err: the server rejected our request for an unknown reason (get pods manila-csi-openstack-manila-csi-controllerplugin-0)&#xA;I0402 15:03:04.449643 24 kubectl_utils.go:128] Logs of kube-system/manila-csi-openstack-manila-csi-controllerplugin-0:nfs-nodeplugin on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:04.449664 24 kubectl_utils.go:129]  : STARTLOG&#xA;&#xA;ENDLOG for container kube-system:manila-csi-openstack-manila-csi-controllerplugin-0:nfs-nodeplugin&#xA;I0402 15:03:04.850403 24 kubectl_utils.go:125] Failed to get logs of pod openstack-cinder-csi-controllerplugin-78f966889b-plrvm, container csi-attacher, err: the server rejected our request for an unknown reason (get pods openstack-cinder-csi-controllerplugin-78f966889b-plrvm)&#xA;I0402 15:03:04.850439 24 kubectl_utils.go:128] Logs of kube-system/openstack-cinder-csi-controllerplugin-78f966889b-plrvm:csi-attacher on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:04.850450 24 kubectl_utils.go:129]  : STARTLOG&#xA;&#xA;ENDLOG for container kube-system:openstack-cinder-csi-controllerplugin-78f966889b-plrvm:csi-attacher&#xA;I0402 15:03:05.252790 24 kubectl_utils.go:125] Failed to get logs of pod openstack-cinder-csi-controllerplugin-78f966889b-plrvm, container csi-provisioner, err: the server rejected our request for an unknown reason (get pods openstack-cinder-csi-controllerplugin-78f966889b-plrvm)&#xA;I0402 15:03:05.252827 24 kubectl_utils.go:128] Logs of kube-system/openstack-cinder-csi-controllerplugin-78f966889b-plrvm:csi-provisioner on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:05.252837 24 kubectl_utils.go:129]  : STARTLOG&#xA;&#xA;ENDLOG for container kube-system:openstack-cinder-csi-controllerplugin-78f966889b-plrvm:csi-provisioner&#xA;I0402 15:03:05.648737 24 kubectl_utils.go:125] Failed to get logs of pod openstack-cinder-csi-controllerplugin-78f966889b-plrvm, container csi-snapshotter, err: the server rejected our request for an unknown reason (get pods openstack-cinder-csi-controllerplugin-78f966889b-plrvm)&#xA;I0402 15:03:05.648774 24 kubectl_utils.go:128] Logs of kube-system/openstack-cinder-csi-controllerplugin-78f966889b-plrvm:csi-snapshotter on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:05.648784 24 kubectl_utils.go:129]  : STARTLOG&#xA;&#xA;ENDLOG for container kube-system:openstack-cinder-csi-controllerplugin-78f966889b-plrvm:csi-snapshotter&#xA;I0402 15:03:06.050461 24 kubectl_utils.go:125] Failed to get logs of pod openstack-cinder-csi-controllerplugin-78f966889b-plrvm, container csi-resizer, err: the server rejected our request for an unknown reason (get pods openstack-cinder-csi-controllerplugin-78f966889b-plrvm)&#xA;I0402 15:03:06.050520 24 kubectl_utils.go:128] Logs of kube-system/openstack-cinder-csi-controllerplugin-78f966889b-plrvm:csi-resizer on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:06.050539 24 kubectl_utils.go:129]  : STARTLOG&#xA;&#xA;ENDLOG for container kube-system:openstack-cinder-csi-controllerplugin-78f966889b-plrvm:csi-resizer&#xA;I0402 15:03:06.452606 24 kubectl_utils.go:125] Failed to get logs of pod openstack-cinder-csi-controllerplugin-78f966889b-plrvm, container liveness-probe, err: the server rejected our request for an unknown reason (get pods openstack-cinder-csi-controllerplugin-78f966889b-plrvm)&#xA;I0402 15:03:06.452644 24 kubectl_utils.go:128] Logs of kube-system/openstack-cinder-csi-controllerplugin-78f966889b-plrvm:liveness-probe on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:06.452710 24 kubectl_utils.go:129]  : STARTLOG&#xA;&#xA;ENDLOG for container kube-system:openstack-cinder-csi-controllerplugin-78f966889b-plrvm:liveness-probe&#xA;I0402 15:03:06.851717 24 kubectl_utils.go:125] Failed to get logs of pod openstack-cinder-csi-controllerplugin-78f966889b-plrvm, container cinder-csi-plugin, err: the server rejected our request for an unknown reason (get pods openstack-cinder-csi-controllerplugin-78f966889b-plrvm)&#xA;I0402 15:03:06.851779 24 kubectl_utils.go:128] Logs of kube-system/openstack-cinder-csi-controllerplugin-78f966889b-plrvm:cinder-csi-plugin on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:06.851796 24 kubectl_utils.go:129]  : STARTLOG&#xA;&#xA;ENDLOG for container kube-system:openstack-cinder-csi-controllerplugin-78f966889b-plrvm:cinder-csi-plugin&#xA;I0402 15:03:07.056242 24 kubectl_utils.go:128] Logs of kube-system/openstack-cloud-controller-manager-chxrv:openstack-cloud-controller-manager on node kube-uhjhf-67cxd-gq627&#xA;I0402 15:03:07.056298 24 kubectl_utils.go:129]  : STARTLOG&#xA;2026/04/02 15:01:27 Running command:&#xA;Command env: (log-file=, also-stdout=false, redirect-stderr=true)&#xA;Run from directory: &#xA;Executable path: /bin/openstack-cloud-controller-manager&#xA;Args (comma-delimited): /bin/openstack-cloud-controller-manager,--v=2,--cloud-config=/etc/config/cloud.conf,--cluster-name=a5757e38-87f8-44aa-b5e9-2a9e6bc4c1c3,--cloud-provider=openstack,--use-service-account-credentials=false,--controllers=cloud-node,cloud-node-lifecycle,route,service,--bind-address=127.0.0.1&#xA;2026/04/02 15:01:27 Now listening for interrupts&#xA;I0402 15:01:27.656766      12 serving.go:386] Generated self-signed cert in-memory&#xA;I0402 15:01:28.633184      12 serving.go:386] Generated self-signed cert in-memory&#xA;W0402 15:01:28.633264      12 client_config.go:667] Neither --kubeconfig nor --master was specified.  Using the inClusterConfig.  This might not work.&#xA;I0402 15:01:32.077667      12 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController&#xA;I0402 15:01:32.140752      12 controllermanager.go:160] Version: v1.33.1&#xA;I0402 15:01:33.637115      12 tlsconfig.go:203] &#34;Loaded serving cert&#34; certName=&#34;Generated self signed cert&#34; certDetail=&#34;\&#34;localhost@1775142087\&#34; [serving] validServingFor=[127.0.0.1,127.0.0.1,localhost] issuer=\&#34;localhost-ca@1775142087\&#34; (2026-04-02 14:01:27 +0000 UTC to 2027-04-02 14:01:27 +0000 UTC (now=2026-04-02 15:01:33.637081347 +0000 UTC))&#34;&#xA;I0402 15:01:33.642729      12 named_certificates.go:53] &#34;Loaded SNI cert&#34; index=0 certName=&#34;self-signed loopback&#34; certDetail=&#34;\&#34;apiserver-loopback-client@1775142089\&#34; [serving] validServingFor=[apiserver-loopback-client] issuer=\&#34;apiserver-loopback-client-ca@1775142089\&#34; (2026-04-02 14:01:28 +0000 UTC to 2029-04-02 14:01:28 +0000 UTC (now=2026-04-02 15:01:33.642696586 +0000 UTC))&#34;&#xA;I0402 15:01:33.642821      12 secure_serving.go:211] Serving securely on 127.0.0.1:10258&#xA;I0402 15:01:33.643181      12 leaderelection.go:257] attempting to acquire leader lease kube-system/cloud-controller-manager...&#xA;I0402 15:01:33.643176      12 configmap_cafile_content.go:205] &#34;Starting controller&#34; name=&#34;client-ca::kube-system::extension-apiserver-authentication::client-ca-file&#34;&#xA;I0402 15:01:33.643160      12 requestheader_controller.go:180] Starting RequestHeaderAuthRequestController&#xA;I0402 15:01:33.643260      12 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;client-ca::kube-system::extension-apiserver-authentication::client-ca-file&#34;&#xA;I0402 15:01:33.643375      12 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;RequestHeaderAuthRequestController&#34;&#xA;I0402 15:01:33.643994      12 configmap_cafile_content.go:205] &#34;Starting controller&#34; name=&#34;client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file&#34;&#xA;I0402 15:01:33.644014      12 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file&#34;&#xA;I0402 15:01:33.644643      12 tlsconfig.go:243] &#34;Starting DynamicServingCertificateController&#34;&#xA;I0402 15:01:33.654069      12 reflector.go:430] &#34;Caches populated&#34; type=&#34;*v1.ConfigMap&#34; reflector=&#34;k8s.io/client-go@v0.33.3/tools/cache/reflector.go:285&#34;&#xA;I0402 15:01:33.654089      12 reflector.go:430] &#34;Caches populated&#34; type=&#34;*v1.ConfigMap&#34; reflector=&#34;k8s.io/client-go@v0.33.3/tools/cache/reflector.go:285&#34;&#xA;I0402 15:01:33.654649      12 reflector.go:430] &#34;Caches populated&#34; type=&#34;*v1.ConfigMap&#34; reflector=&#34;k8s.io/client-go@v0.33.3/tools/cache/reflector.go:285&#34;&#xA;I0402 15:01:33.744076      12 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;RequestHeaderAuthRequestController&#34;&#xA;I0402 15:01:33.744608      12 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;client-ca::kube-system::extension-apiserver-authentication::client-ca-file&#34;&#xA;I0402 15:01:33.744746      12 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file&#34;&#xA;I0402 15:01:33.745561      12 tlsconfig.go:181] &#34;Loaded client CA&#34; index=0 certName=&#34;client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file&#34; certDetail=&#34;\&#34;k8s-cluster\&#34; [] issuer=\&#34;&lt;self&gt;\&#34; (2026-04-01 14:53:12 +0000 UTC to 2031-04-01 14:53:12 +0000 UTC (now=2026-04-02 15:01:33.745474223 +0000 UTC))&#34;&#xA;I0402 15:01:33.745901      12 tlsconfig.go:203] &#34;Loaded serving cert&#34; certName=&#34;Generated self signed cert&#34; certDetail=&#34;\&#34;localhost@1775142087\&#34; [serving] validServingFor=[127.0.0.1,127.0.0.1,localhost] issuer=\&#34;localhost-ca@1775142087\&#34; (2026-04-02 14:01:27 +0000 UTC to 2027-04-02 14:01:27 +0000 UTC (now=2026-04-02 15:01:33.74588601 +0000 UTC))&#34;&#xA;I0402 15:01:33.746299      12 named_certificates.go:53] &#34;Loaded SNI cert&#34; index=0 certName=&#34;self-signed loopback&#34; certDetail=&#34;\&#34;apiserver-loopback-client@1775142089\&#34; [serving] validServingFor=[apiserver-loopback-client] issuer=\&#34;apiserver-loopback-client-ca@1775142089\&#34; (2026-04-02 14:01:28 +0000 UTC to 2029-04-02 14:01:28 +0000 UTC (now=2026-04-02 15:01:33.746282468 +0000 UTC))&#34;&#xA;I0402 15:01:33.746400      12 tlsconfig.go:181] &#34;Loaded client CA&#34; index=0 certName=&#34;client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file&#34; certDetail=&#34;\&#34;k8s-cluster\&#34; [] issuer=\&#34;&lt;self&gt;\&#34; (2026-04-01 14:53:10 +0000 UTC to 2031-04-01 14:53:10 +0000 UTC (now=2026-04-02 15:01:33.74638848 +0000 UTC))&#34;&#xA;I0402 15:01:33.746447      12 tlsconfig.go:181] &#34;Loaded client CA&#34; index=1 certName=&#34;client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file&#34; certDetail=&#34;\&#34;k8s-cluster\&#34; [] issuer=\&#34;&lt;self&gt;\&#34; (2026-04-01 14:53:12 +0000 UTC to 2031-04-01 14:53:12 +0000 UTC (now=2026-04-02 15:01:33.746438791 +0000 UTC))&#34;&#xA;I0402 15:01:33.746700      12 tlsconfig.go:203] &#34;Loaded serving cert&#34; certName=&#34;Generated self signed cert&#34; certDetail=&#34;\&#34;localhost@1775142087\&#34; [serving] validServingFor=[127.0.0.1,127.0.0.1,localhost] issuer=\&#34;localhost-ca@1775142087\&#34; (2026-04-02 14:01:27 +0000 UTC to 2027-04-02 14:01:27 +0000 UTC (now=2026-04-02 15:01:33.746689155 +0000 UTC))&#34;&#xA;I0402 15:01:33.746926      12 named_certificates.go:53] &#34;Loaded SNI cert&#34; index=0 certName=&#34;self-signed loopback&#34; certDetail=&#34;\&#34;apiserver-loopback-client@1775142089\&#34; [serving] validServingFor=[apiserver-loopback-client] issuer=\&#34;apiserver-loopback-client-ca@1775142089\&#34; (2026-04-02 14:01:28 +0000 UTC to 2029-04-02 14:01:28 +0000 UTC (now=2026-04-02 15:01:33.746915569 +0000 UTC))&#34;&#xA;I0402 15:01:48.769354      12 leaderelection.go:271] successfully acquired lease kube-system/cloud-controller-manager&#xA;I0402 15:01:48.769740      12 event.go:389] &#34;Event occurred&#34; object=&#34;kube-system/cloud-controller-manager&#34; fieldPath=&#34;&#34; kind=&#34;Lease&#34; apiVersion=&#34;coordination.k8s.io/v1&#34; type=&#34;Normal&#34; reason=&#34;LeaderElection&#34; message=&#34;kube-uhjhf-67cxd-gq627_1ec10fd8-42c4-40e3-a573-e12c0e83dcc4 became leader&#34;&#xA;I0402 15:01:48.776960      12 openstack.go:424] Setting up informers for Cloud&#xA;I0402 15:01:48.777433      12 controllermanager.go:310] Starting &#34;node-route-controller&#34;&#xA;W0402 15:01:48.835161      12 openstack.go:408] Error initialising Routes support: router-id not set in cloud provider config&#xA;W0402 15:01:48.835281      12 core.go:111] --configure-cloud-routes is set, but cloud provider does not support routes. Will not configure cloud provider routes.&#xA;W0402 15:01:48.835339      12 controllermanager.go:317] Skipping &#34;node-route-controller&#34;&#xA;I0402 15:01:48.835350      12 controllermanager.go:310] Starting &#34;cloud-node-controller&#34;&#xA;I0402 15:01:48.835675      12 controllermanager.go:329] Started &#34;cloud-node-controller&#34;&#xA;I0402 15:01:48.835708      12 controllermanager.go:310] Starting &#34;cloud-node-lifecycle-controller&#34;&#xA;I0402 15:01:48.835866      12 controllermanager.go:329] Started &#34;cloud-node-lifecycle-controller&#34;&#xA;I0402 15:01:48.835872      12 controllermanager.go:310] Starting &#34;service-lb-controller&#34;&#xA;I0402 15:01:48.836149      12 openstack.go:373] Claiming to support LoadBalancer&#xA;I0402 15:01:48.836173      12 controllermanager.go:329] Started &#34;service-lb-controller&#34;&#xA;I0402 15:01:48.836648      12 node_controller.go:176] Sending events to api server.&#xA;I0402 15:01:48.836679      12 node_lifecycle_controller.go:112] Sending events to api server&#xA;I0402 15:01:48.836738      12 node_controller.go:185] Waiting for informer caches to sync&#xA;I0402 15:01:48.836801      12 controller.go:234] Starting service controller&#xA;I0402 15:01:48.836812      12 shared_informer.go:350] &#34;Waiting for caches to sync&#34; controller=&#34;service&#34;&#xA;I0402 15:01:48.841013      12 reflector.go:430] &#34;Caches populated&#34; type=&#34;*v1.Node&#34; reflector=&#34;k8s.io/client-go@v0.33.3/tools/cache/reflector.go:285&#34;&#xA;I0402 15:01:48.842532      12 reflector.go:430] &#34;Caches populated&#34; type=&#34;*v1.Service&#34; reflector=&#34;k8s.io/client-go@v0.33.3/tools/cache/reflector.go:285&#34;&#xA;I0402 15:01:48.937810      12 shared_informer.go:357] &#34;Caches are synced&#34; controller=&#34;service&#34;&#xA;I0402 15:01:48.937957      12 controller.go:722] Syncing backends for all LB services.&#xA;I0402 15:01:48.937974      12 controller.go:726] Successfully updated 0 out of 0 load balancers to direct traffic to the updated set of nodes&#xA;I0402 15:01:48.937990      12 controller.go:722] Syncing backends for all LB services.&#xA;I0402 15:01:48.937995      12 controller.go:726] Successfully updated 0 out of 0 load balancers to direct traffic to the updated set of nodes&#xA;I0402 15:01:48.938003      12 controller.go:722] Syncing backends for all LB services.&#xA;I0402 15:01:48.938008      12 controller.go:726] Successfully updated 0 out of 0 load balancers to direct traffic to the updated set of nodes&#xA;I0402 15:01:50.046021      12 node_controller.go:271] Update 3 nodes status took 1.108059562s.&#xA;E0402 15:02:49.907475      12 leaderelection.go:429] Failed to update lock optimistically: Put &#34;https://10.254.0.1:443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/cloud-controller-manager?timeout=5s&#34;: context deadline exceeded, falling back to slow path&#xA;E0402 15:02:54.910854      12 leaderelection.go:436] error retrieving resource lock kube-system/cloud-controller-manager: Get &#34;https://10.254.0.1:443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/cloud-controller-manager?timeout=5s&#34;: net/http: request canceled (Client.Timeout exceeded while awaiting headers)&#xA;I0402 15:02:54.910968      12 leaderelection.go:297] failed to renew lease kube-system/cloud-controller-manager: context deadline exceeded&#xA;E0402 15:02:54.911140      12 controllermanager.go:265] &#34;leaderelection lost&#34;&#xA;2026/04/02 15:02:54 running command: exit status 1&#xA;&#xA;ENDLOG for container kube-system:openstack-cloud-controller-manager-chxrv:openstack-cloud-controller-manager&#xA;[FAILED] Error waiting for all pods to be running and ready: Told to stop trying after 18.147s.&#xA;Unexpected final error while getting *pod.state: listing replication controllers in namespace kube-system: etcdserver: request timed out&#xA;At one point, however, the function did return successfully.&#xA;Yet, Eventually failed because the matcher was not satisfied:&#xA;Expected all pods (need at least 0) in namespace &#34;kube-system&#34; to be running and ready (except for 0).&#xA;25 / 27 pods were running and ready.&#xA;Expected 5 pod replicas, 4 are Running and Ready.&#xA;Pods that were neither completed nor running:&#xA;    &lt;[]v1.Pod | len:2, cap:2&gt;: &#xA;        - metadata:&#xA;            annotations:&#xA;              cni.projectcalico.org/containerID: 18f9c5d3bd5023ece4b6d3fe23b14c208d2d7c525b8a31ba893326fcb3cc63a0&#xA;              cni.projectcalico.org/podIP: 10.100.78.131/32&#xA;              cni.projectcalico.org/podIPs: 10.100.78.131/32&#xA;            creationTimestamp: &#34;2026-04-02T14:56:28Z&#34;&#xA;            generateName: manila-csi-openstack-manila-csi-controllerplugin-&#xA;            generation: 1&#xA;            labels:&#xA;              app: openstack-manila-csi&#xA;              apps.kubernetes.io/pod-index: &#34;0&#34;&#xA;              chart: openstack-manila-csi-2.32.0&#xA;              component: controllerplugin&#xA;              controller-revision-hash: manila-csi-openstack-manila-csi-controllerplugin-6f488f9967&#xA;              heritage: Helm&#xA;              release: manila-csi&#xA;              statefulset.kubernetes.io/pod-name: manila-csi-openstack-manila-csi-controllerplugin-0&#xA;            managedFields:&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:metadata:&#xA;                  f:generateName: {}&#xA;                  f:labels:&#xA;                    .: {}&#xA;                    f:app: {}&#xA;                    f:apps.kubernetes.io/pod-index: {}&#xA;                    f:chart: {}&#xA;                    f:component: {}&#xA;                    f:controller-revision-hash: {}&#xA;                    f:heritage: {}&#xA;                    f:release: {}&#xA;                    f:statefulset.kubernetes.io/pod-name: {}&#xA;                  f:ownerReferences:&#xA;                    .: {}&#xA;                    k:{&#34;uid&#34;:&#34;1e454624-fa6c-46e9-8a02-a1769df05b94&#34;}: {}&#xA;                f:spec:&#xA;                  f:containers:&#xA;                    k:{&#34;name&#34;:&#34;nfs-nodeplugin&#34;}:&#xA;                      .: {}&#xA;                      f:command: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;CSI_ENDPOINT&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                        k:{&#34;name&#34;:&#34;DRIVER_NAME&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                        k:{&#34;name&#34;:&#34;FWD_CSI_ENDPOINT&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                        k:{&#34;name&#34;:&#34;MANILA_SHARE_PROTO&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext:&#xA;                        .: {}&#xA;                        f:allowPrivilegeEscalation: {}&#xA;                        f:capabilities:&#xA;                          .: {}&#xA;                          f:add: {}&#xA;                        f:privileged: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/etc/config&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                          f:readOnly: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/plugins/csi-nfsplugin&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/pods&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:mountPropagation: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;nfs-provisioner&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;nfs-resizer&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;nfs-snapshotter&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                  f:dnsPolicy: {}&#xA;                  f:enableServiceLinks: {}&#xA;                  f:hostname: {}&#xA;                  f:restartPolicy: {}&#xA;                  f:schedulerName: {}&#xA;                  f:securityContext: {}&#xA;                  f:serviceAccount: {}&#xA;                  f:serviceAccountName: {}&#xA;                  f:subdomain: {}&#xA;                  f:terminationGracePeriodSeconds: {}&#xA;                  f:tolerations: {}&#xA;                  f:volumes:&#xA;                    .: {}&#xA;                    k:{&#34;name&#34;:&#34;cloud-config&#34;}:&#xA;                      .: {}&#xA;                      f:name: {}&#xA;                      f:secret:&#xA;                        .: {}&#xA;                        f:defaultMode: {}&#xA;                        f:items: {}&#xA;                        f:secretName: {}&#xA;                    k:{&#34;name&#34;:&#34;nfs-fwd-plugin-dir&#34;}:&#xA;                      .: {}&#xA;                      f:hostPath:&#xA;                        .: {}&#xA;                        f:path: {}&#xA;                        f:type: {}&#xA;                      f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;nfs-plugin-dir&#34;}:&#xA;                      .: {}&#xA;                      f:hostPath:&#xA;                        .: {}&#xA;                        f:path: {}&#xA;                        f:type: {}&#xA;                      f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;pod-mounts&#34;}:&#xA;                      .: {}&#xA;                      f:hostPath:&#xA;                        .: {}&#xA;                        f:path: {}&#xA;                        f:type: {}&#xA;                      f:name: {}&#xA;              manager: kube-controller-manager&#xA;              operation: Update&#xA;              time: &#34;2026-04-02T14:56:28Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:status:&#xA;                  f:conditions:&#xA;                    .: {}&#xA;                    k:{&#34;type&#34;:&#34;PodScheduled&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;              manager: kube-scheduler&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:03Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:status:&#xA;                  f:conditions:&#xA;                    k:{&#34;type&#34;:&#34;ContainersReady&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;Initialized&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;PodReadyToStartContainers&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;Ready&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                  f:containerStatuses: {}&#xA;                  f:startTime: {}&#xA;              manager: kubelet&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:53Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:metadata:&#xA;                  f:annotations:&#xA;                    .: {}&#xA;                    f:cni.projectcalico.org/containerID: {}&#xA;                    f:cni.projectcalico.org/podIP: {}&#xA;                    f:cni.projectcalico.org/podIPs: {}&#xA;              manager: calico&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:55Z&#34;&#xA;            name: manila-csi-openstack-manila-csi-controllerplugin-0&#xA;            namespace: kube-system&#xA;            ownerReferences:&#xA;            - apiVersion: apps/v1&#xA;              blockOwnerDeletion: true&#xA;              controller: true&#xA;              kind: StatefulSet&#xA;              name: manila-csi-openstack-manila-csi-controllerplugin&#xA;              uid: 1e454624-fa6c-46e9-8a02-a1769df05b94&#xA;            resourceVersion: &#34;1549&#34;&#xA;            uid: f3c4b392-db35-43b3-8e28-3e3d429adcf7&#xA;          spec:&#xA;            containers:&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: unix:///var/lib/kubelet/plugins/nfs.manila.csi.openstack.org/csi-controllerplugin.sock&#xA;              image: registry.k8s.io/sig-storage/csi-provisioner:v5.1.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: nfs-provisioner&#xA;              resources: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: unix:///var/lib/kubelet/plugins/nfs.manila.csi.openstack.org/csi-controllerplugin.sock&#xA;              image: registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: nfs-snapshotter&#xA;              resources: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              - --handle-volume-inuse-error=false&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: unix:///var/lib/kubelet/plugins/nfs.manila.csi.openstack.org/csi-controllerplugin.sock&#xA;              image: registry.k8s.io/sig-storage/csi-resizer:v1.12.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: nfs-resizer&#xA;              resources: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;            - command:&#xA;              - /bin/sh&#xA;              - -c&#xA;              - /bin/manila-csi-plugin -v=2 --endpoint=$(CSI_ENDPOINT) --drivername=$(DRIVER_NAME)&#xA;                --share-protocol-selector=$(MANILA_SHARE_PROTO) --fwdendpoint=$(FWD_CSI_ENDPOINT)&#xA;                --cluster-id=&#34;&#34;&#xA;              env:&#xA;              - name: DRIVER_NAME&#xA;                value: nfs.manila.csi.openstack.org&#xA;              - name: CSI_ENDPOINT&#xA;                value: unix:///var/lib/kubelet/plugins/nfs.manila.csi.openstack.org/csi-controllerplugin.sock&#xA;              - name: FWD_CSI_ENDPOINT&#xA;                value: unix:///var/lib/kubelet/plugins/csi-nfsplugin/csi.sock&#xA;              - name: MANILA_SHARE_PROTO&#xA;                value: NFS&#xA;              image: registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: nfs-nodeplugin&#xA;              resources: {}&#xA;              securityContext:&#xA;                allowPrivilegeEscalation: true&#xA;                capabilities:&#xA;                  add:&#xA;                  - SYS_ADMIN&#xA;                privileged: true&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/lib/kubelet/plugins/csi-nfsplugin&#xA;                name: nfs-fwd-plugin-dir&#xA;              - mountPath: /var/lib/kubelet/pods&#xA;                mountPropagation: Bidirectional&#xA;                name: pod-mounts&#xA;              - mountPath: /etc/config&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;            dnsPolicy: ClusterFirst&#xA;            enableServiceLinks: true&#xA;            hostname: manila-csi-openstack-manila-csi-controllerplugin-0&#xA;            nodeName: kube-uhjhf-67cxd-gq627&#xA;            preemptionPolicy: PreemptLowerPriority&#xA;            priority: 0&#xA;            restartPolicy: Always&#xA;            schedulerName: default-scheduler&#xA;            securityContext: {}&#xA;            serviceAccount: manila-csi-openstack-manila-csi-controllerplugin&#xA;            serviceAccountName: manila-csi-openstack-manila-csi-controllerplugin&#xA;            subdomain: manila-csi-openstack-manila-csi-controllerplugin&#xA;            terminationGracePeriodSeconds: 30&#xA;            tolerations:&#xA;            - effect: NoSchedule&#xA;              key: node-role.kubernetes.io/master&#xA;            - effect: NoSchedule&#xA;              key: node-role.kubernetes.io/control-plane&#xA;            - effect: NoExecute&#xA;              key: node.kubernetes.io/not-ready&#xA;              operator: Exists&#xA;              tolerationSeconds: 300&#xA;            - effect: NoExecute&#xA;              key: node.kubernetes.io/unreachable&#xA;              operator: Exists&#xA;              tolerationSeconds: 300&#xA;            volumes:&#xA;            - hostPath:&#xA;                path: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                type: DirectoryOrCreate&#xA;              name: nfs-plugin-dir&#xA;            - hostPath:&#xA;                path: /var/lib/kubelet/plugins/csi-nfsplugin&#xA;                type: Directory&#xA;              name: nfs-fwd-plugin-dir&#xA;            - hostPath:&#xA;                path: /var/lib/kubelet/pods&#xA;                type: Directory&#xA;              name: pod-mounts&#xA;            - name: cloud-config&#xA;              secret:&#xA;                defaultMode: 420&#xA;                items:&#xA;                - key: ca.crt&#xA;                  path: ca.crt&#xA;                secretName: cloud-config&#xA;            - name: kube-api-access-tq4nx&#xA;              projected:&#xA;                defaultMode: 420&#xA;                sources:&#xA;                - serviceAccountToken:&#xA;                    expirationSeconds: 3607&#xA;                    path: token&#xA;                - configMap:&#xA;                    items:&#xA;                    - key: ca.crt&#xA;                      path: ca.crt&#xA;                    name: kube-root-ca.crt&#xA;                - downwardAPI:&#xA;                    items:&#xA;                    - fieldRef:&#xA;                        apiVersion: v1&#xA;                        fieldPath: metadata.namespace&#xA;                      path: namespace&#xA;          status:&#xA;            conditions:&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              status: &#34;False&#34;&#xA;              type: PodReadyToStartContainers&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              status: &#34;True&#34;&#xA;              type: Initialized&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              message: &#39;containers with unready status: [nfs-provisioner nfs-snapshotter nfs-resizer&#xA;                nfs-nodeplugin]&#39;&#xA;              reason: ContainersNotReady&#xA;              status: &#34;False&#34;&#xA;              type: Ready&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              message: &#39;containers with unready status: [nfs-provisioner nfs-snapshotter nfs-resizer&#xA;                nfs-nodeplugin]&#39;&#xA;              reason: ContainersNotReady&#xA;              status: &#34;False&#34;&#xA;              type: ContainersReady&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:52Z&#34;&#xA;              status: &#34;True&#34;&#xA;              type: PodScheduled&#xA;            containerStatuses:&#xA;            - image: registry.k8s.io/provider-os/manila-csi-plugin:v1.32.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: nfs-nodeplugin&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/lib/kubelet/plugins/csi-nfsplugin&#xA;                name: nfs-fwd-plugin-dir&#xA;              - mountPath: /var/lib/kubelet/pods&#xA;                name: pod-mounts&#xA;              - mountPath: /etc/config&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-provisioner:v5.1.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: nfs-provisioner&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-resizer:v1.12.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: nfs-resizer&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: nfs-snapshotter&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/kubelet/plugins/nfs.manila.csi.openstack.org&#xA;                name: nfs-plugin-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-tq4nx&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            phase: Pending&#xA;            qosClass: BestEffort&#xA;            startTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;        - metadata:&#xA;            annotations:&#xA;              cni.projectcalico.org/containerID: cf0f08707afe3c7479ac743e82356119aae0fadac55f61f29ca480473dd08907&#xA;              cni.projectcalico.org/podIP: 10.100.78.132/32&#xA;              cni.projectcalico.org/podIPs: 10.100.78.132/32&#xA;            creationTimestamp: &#34;2026-04-02T14:56:27Z&#34;&#xA;            generateName: openstack-cinder-csi-controllerplugin-78f966889b-&#xA;            generation: 1&#xA;            labels:&#xA;              app: openstack-cinder-csi&#xA;              chart: openstack-cinder-csi-2.32.0&#xA;              component: controllerplugin&#xA;              heritage: Helm&#xA;              pod-template-hash: 78f966889b&#xA;              release: cinder-csi&#xA;            managedFields:&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:metadata:&#xA;                  f:generateName: {}&#xA;                  f:labels:&#xA;                    .: {}&#xA;                    f:app: {}&#xA;                    f:chart: {}&#xA;                    f:component: {}&#xA;                    f:heritage: {}&#xA;                    f:pod-template-hash: {}&#xA;                    f:release: {}&#xA;                  f:ownerReferences:&#xA;                    .: {}&#xA;                    k:{&#34;uid&#34;:&#34;3bd9f6ed-6669-41c3-9246-73d536880313&#34;}: {}&#xA;                f:spec:&#xA;                  f:affinity: {}&#xA;                  f:containers:&#xA;                    k:{&#34;name&#34;:&#34;cinder-csi-plugin&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;CLOUD_CONFIG&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                        k:{&#34;name&#34;:&#34;CLUSTER_NAME&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                        k:{&#34;name&#34;:&#34;CSI_ENDPOINT&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:livenessProbe:&#xA;                        .: {}&#xA;                        f:failureThreshold: {}&#xA;                        f:httpGet:&#xA;                          .: {}&#xA;                          f:path: {}&#xA;                          f:port: {}&#xA;                          f:scheme: {}&#xA;                        f:initialDelaySeconds: {}&#xA;                        f:periodSeconds: {}&#xA;                        f:successThreshold: {}&#xA;                        f:timeoutSeconds: {}&#xA;                      f:name: {}&#xA;                      f:ports:&#xA;                        .: {}&#xA;                        k:{&#34;containerPort&#34;:9808,&#34;protocol&#34;:&#34;TCP&#34;}:&#xA;                          .: {}&#xA;                          f:containerPort: {}&#xA;                          f:name: {}&#xA;                          f:protocol: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/csi&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/etc/config&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                          f:readOnly: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/etc/kubernetes&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                          f:readOnly: {}&#xA;                    k:{&#34;name&#34;:&#34;csi-attacher&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/csi/sockets/pluginproxy/&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;csi-provisioner&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/csi/sockets/pluginproxy/&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;csi-resizer&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/csi/sockets/pluginproxy/&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;csi-snapshotter&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/csi/sockets/pluginproxy/&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                    k:{&#34;name&#34;:&#34;liveness-probe&#34;}:&#xA;                      .: {}&#xA;                      f:args: {}&#xA;                      f:env:&#xA;                        .: {}&#xA;                        k:{&#34;name&#34;:&#34;ADDRESS&#34;}:&#xA;                          .: {}&#xA;                          f:name: {}&#xA;                          f:value: {}&#xA;                      f:image: {}&#xA;                      f:imagePullPolicy: {}&#xA;                      f:name: {}&#xA;                      f:resources: {}&#xA;                      f:securityContext: {}&#xA;                      f:terminationMessagePath: {}&#xA;                      f:terminationMessagePolicy: {}&#xA;                      f:volumeMounts:&#xA;                        .: {}&#xA;                        k:{&#34;mountPath&#34;:&#34;/var/lib/csi/sockets/pluginproxy/&#34;}:&#xA;                          .: {}&#xA;                          f:mountPath: {}&#xA;                          f:name: {}&#xA;                  f:dnsPolicy: {}&#xA;                  f:enableServiceLinks: {}&#xA;                  f:restartPolicy: {}&#xA;                  f:schedulerName: {}&#xA;                  f:securityContext: {}&#xA;                  f:serviceAccount: {}&#xA;                  f:serviceAccountName: {}&#xA;                  f:terminationGracePeriodSeconds: {}&#xA;                  f:tolerations: {}&#xA;                  f:volumes:&#xA;                    .: {}&#xA;                    k:{&#34;name&#34;:&#34;cloud-config&#34;}:&#xA;                      .: {}&#xA;                      f:name: {}&#xA;                      f:secret:&#xA;                        .: {}&#xA;                        f:defaultMode: {}&#xA;                        f:secretName: {}&#xA;                    k:{&#34;name&#34;:&#34;socket-dir&#34;}:&#xA;                      .: {}&#xA;                      f:emptyDir: {}&#xA;                      f:name: {}&#xA;              manager: kube-controller-manager&#xA;              operation: Update&#xA;              time: &#34;2026-04-02T14:56:27Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:status:&#xA;                  f:conditions:&#xA;                    .: {}&#xA;                    k:{&#34;type&#34;:&#34;PodScheduled&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;              manager: kube-scheduler&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:03Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:status:&#xA;                  f:conditions:&#xA;                    k:{&#34;type&#34;:&#34;ContainersReady&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;Initialized&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;PodReadyToStartContainers&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                    k:{&#34;type&#34;:&#34;Ready&#34;}:&#xA;                      .: {}&#xA;                      f:lastProbeTime: {}&#xA;                      f:lastTransitionTime: {}&#xA;                      f:message: {}&#xA;                      f:reason: {}&#xA;                      f:status: {}&#xA;                      f:type: {}&#xA;                  f:containerStatuses: {}&#xA;                  f:hostIP: {}&#xA;                  f:hostIPs: {}&#xA;                  f:startTime: {}&#xA;              manager: kubelet&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:53Z&#34;&#xA;            - apiVersion: v1&#xA;              fieldsType: FieldsV1&#xA;              fieldsV1:&#xA;                f:metadata:&#xA;                  f:annotations:&#xA;                    .: {}&#xA;                    f:cni.projectcalico.org/containerID: {}&#xA;                    f:cni.projectcalico.org/podIP: {}&#xA;                    f:cni.projectcalico.org/podIPs: {}&#xA;              manager: calico&#xA;              operation: Update&#xA;              subresource: status&#xA;              time: &#34;2026-04-02T15:00:58Z&#34;&#xA;            name: openstack-cinder-csi-controllerplugin-78f966889b-plrvm&#xA;            namespace: kube-system&#xA;            ownerReferences:&#xA;            - apiVersion: apps/v1&#xA;              blockOwnerDeletion: true&#xA;              controller: true&#xA;              kind: ReplicaSet&#xA;              name: openstack-cinder-csi-controllerplugin-78f966889b&#xA;              uid: 3bd9f6ed-6669-41c3-9246-73d536880313&#xA;            resourceVersion: &#34;1562&#34;&#xA;            uid: 4f0b1844-d34b-44cc-a923-06bca632e956&#xA;          spec:&#xA;            affinity: {}&#xA;            containers:&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              - --timeout=3m&#xA;              - --leader-election=true&#xA;              - --default-fstype=ext4&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: /var/lib/csi/sockets/pluginproxy/csi.sock&#xA;              image: registry.k8s.io/sig-storage/csi-attacher:v4.7.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: csi-attacher&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              - --timeout=3m&#xA;              - --leader-election=true&#xA;              - --default-fstype=ext4&#xA;              - --feature-gates=Topology=true&#xA;              - --extra-create-metadata&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: /var/lib/csi/sockets/pluginproxy/csi.sock&#xA;              image: registry.k8s.io/sig-storage/csi-provisioner:v5.1.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: csi-provisioner&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              - --timeout=3m&#xA;              - --leader-election=true&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: /var/lib/csi/sockets/pluginproxy/csi.sock&#xA;              image: registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: csi-snapshotter&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              - --timeout=3m&#xA;              - --handle-volume-inuse-error=false&#xA;              - --leader-election=true&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: /var/lib/csi/sockets/pluginproxy/csi.sock&#xA;              image: registry.k8s.io/sig-storage/csi-resizer:v1.12.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: csi-resizer&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            - args:&#xA;              - -v=2&#xA;              - --csi-address=$(ADDRESS)&#xA;              env:&#xA;              - name: ADDRESS&#xA;                value: /var/lib/csi/sockets/pluginproxy/csi.sock&#xA;              image: registry.k8s.io/sig-storage/livenessprobe:v2.14.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              name: liveness-probe&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            - args:&#xA;              - /bin/cinder-csi-plugin&#xA;              - -v=2&#xA;              - --endpoint=$(CSI_ENDPOINT)&#xA;              - --cloud-config=$(CLOUD_CONFIG)&#xA;              - --cluster=$(CLUSTER_NAME)&#xA;              - --provide-node-service=false&#xA;              env:&#xA;              - name: CSI_ENDPOINT&#xA;                value: unix://csi/csi.sock&#xA;              - name: CLOUD_CONFIG&#xA;                value: /etc/kubernetes/cloud.conf&#xA;              - name: CLUSTER_NAME&#xA;                value: a5757e38-87f8-44aa-b5e9-2a9e6bc4c1c3&#xA;              image: registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0&#xA;              imagePullPolicy: IfNotPresent&#xA;              livenessProbe:&#xA;                failureThreshold: 5&#xA;                httpGet:&#xA;                  path: /healthz&#xA;                  port: healthz&#xA;                  scheme: HTTP&#xA;                initialDelaySeconds: 10&#xA;                periodSeconds: 60&#xA;                successThreshold: 1&#xA;                timeoutSeconds: 10&#xA;              name: cinder-csi-plugin&#xA;              ports:&#xA;              - containerPort: 9808&#xA;                name: healthz&#xA;                protocol: TCP&#xA;              resources: {}&#xA;              securityContext: {}&#xA;              terminationMessagePath: /dev/termination-log&#xA;              terminationMessagePolicy: File&#xA;              volumeMounts:&#xA;              - mountPath: /csi&#xA;                name: socket-dir&#xA;              - mountPath: /etc/kubernetes&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;              - mountPath: /etc/config&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;            dnsPolicy: ClusterFirst&#xA;            enableServiceLinks: true&#xA;            nodeName: kube-uhjhf-67cxd-gq627&#xA;            preemptionPolicy: PreemptLowerPriority&#xA;            priority: 0&#xA;            restartPolicy: Always&#xA;            schedulerName: default-scheduler&#xA;            securityContext: {}&#xA;            serviceAccount: csi-cinder-controller-sa&#xA;            serviceAccountName: csi-cinder-controller-sa&#xA;            terminationGracePeriodSeconds: 30&#xA;            tolerations:&#xA;            - effect: NoSchedule&#xA;              key: node-role.kubernetes.io/master&#xA;            - effect: NoSchedule&#xA;              key: node-role.kubernetes.io/control-plane&#xA;            - effect: NoExecute&#xA;              key: node.kubernetes.io/not-ready&#xA;              operator: Exists&#xA;              tolerationSeconds: 300&#xA;            - effect: NoExecute&#xA;              key: node.kubernetes.io/unreachable&#xA;              operator: Exists&#xA;              tolerationSeconds: 300&#xA;            volumes:&#xA;            - emptyDir: {}&#xA;              name: socket-dir&#xA;            - name: cloud-config&#xA;              secret:&#xA;                defaultMode: 420&#xA;                secretName: cloud-config&#xA;            - name: kube-api-access-6tzzg&#xA;              projected:&#xA;                defaultMode: 420&#xA;                sources:&#xA;                - serviceAccountToken:&#xA;                    expirationSeconds: 3607&#xA;                    path: token&#xA;                - configMap:&#xA;                    items:&#xA;                    - key: ca.crt&#xA;                      path: ca.crt&#xA;                    name: kube-root-ca.crt&#xA;                - downwardAPI:&#xA;                    items:&#xA;                    - fieldRef:&#xA;                        apiVersion: v1&#xA;                        fieldPath: metadata.namespace&#xA;                      path: namespace&#xA;          status:&#xA;            conditions:&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              status: &#34;False&#34;&#xA;              type: PodReadyToStartContainers&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              status: &#34;True&#34;&#xA;              type: Initialized&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              message: &#39;containers with unready status: [csi-attacher csi-provisioner csi-snapshotter&#xA;                csi-resizer liveness-probe cinder-csi-plugin]&#39;&#xA;              reason: ContainersNotReady&#xA;              status: &#34;False&#34;&#xA;              type: Ready&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;              message: &#39;containers with unready status: [csi-attacher csi-provisioner csi-snapshotter&#xA;                csi-resizer liveness-probe cinder-csi-plugin]&#39;&#xA;              reason: ContainersNotReady&#xA;              status: &#34;False&#34;&#xA;              type: ContainersReady&#xA;            - lastProbeTime: null&#xA;              lastTransitionTime: &#34;2026-04-02T15:00:52Z&#34;&#xA;              status: &#34;True&#34;&#xA;              type: PodScheduled&#xA;            containerStatuses:&#xA;            - image: registry.k8s.io/provider-os/cinder-csi-plugin:v1.32.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: cinder-csi-plugin&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /csi&#xA;                name: socket-dir&#xA;              - mountPath: /etc/kubernetes&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;              - mountPath: /etc/config&#xA;                name: cloud-config&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-attacher:v4.7.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: csi-attacher&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-provisioner:v5.1.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: csi-provisioner&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-resizer:v1.12.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: csi-resizer&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/csi-snapshotter:v8.1.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: csi-snapshotter&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            - image: registry.k8s.io/sig-storage/livenessprobe:v2.14.0&#xA;              imageID: &#34;&#34;&#xA;              lastState: {}&#xA;              name: liveness-probe&#xA;              ready: false&#xA;              restartCount: 0&#xA;              started: false&#xA;              state:&#xA;                waiting:&#xA;                  reason: ContainerCreating&#xA;              volumeMounts:&#xA;              - mountPath: /var/lib/csi/sockets/pluginproxy/&#xA;                name: socket-dir&#xA;              - mountPath: /var/run/secrets/kubernetes.io/serviceaccount&#xA;                name: kube-api-access-6tzzg&#xA;                readOnly: true&#xA;                recursiveReadOnly: Disabled&#xA;            hostIP: 192.168.24.167&#xA;            hostIPs:&#xA;            - ip: 192.168.24.167&#xA;            phase: Pending&#xA;            qosClass: BestEffort&#xA;            startTime: &#34;2026-04-02T15:00:53Z&#34;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:232 @ 04/02/26 15:03:07.057&#xA;&lt; Exit [SynchronizedBeforeSuite] TOP-LEVEL - k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.057 (33.302s)&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.167714828"></testcase>
          <testcase name="[ReportAfterSuite] Kubernetes e2e suite report" classname="Kubernetes e2e suite" status="passed" time="0.001887995"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.062035179">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.074&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.074&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.000450521"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="32.924925812">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.074&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.074&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.001397691"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.292520774">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.074&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.074&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.002153087"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.191624881">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.074&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.074&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.001232032"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.304867263">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.074&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.074&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.000639731"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="32.829552496">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.073&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.073&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.001956415"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.300515877">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.06&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.06&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.001874526"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.269718341">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.073&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.073&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.003091638"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.336333752">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.137&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.137&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.000388011"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.310472015">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.089&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.089&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.001679526"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="32.862574215">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.097&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.097&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.005013985"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="32.972864343">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.139&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.139&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.001999967"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.030979695">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.14&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.14&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.001981367"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.330515127">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.089&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.089&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.006046226"></testcase>
          <testcase name="[SynchronizedBeforeSuite]" classname="Kubernetes e2e suite" status="failed" time="33.225001393">
              <failure message="" type="failed">[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.146&#xA;</failure>
              <system-err>[FAILED] �[1m�[38;5;9mSynchronizedBeforeSuite failed on Ginkgo parallel process #1�[0m&#xA;  The first SynchronizedBeforeSuite function running on Ginkgo parallel process&#xA;  #1 failed.  This suite will now abort.&#xA;&#xA;In [SynchronizedBeforeSuite] at: k8s.io/kubernetes/test/e2e/e2e.go:69 @ 04/02/26 15:03:07.146&#xA;</system-err>
          </testcase>
          <testcase name="[SynchronizedAfterSuite]" classname="Kubernetes e2e suite" status="passed" time="0.004315654"></testcase>
      </testsuite>
  </testsuites>