I0421 19:17:05.700912 1 serving.go:348] Generated self-signed cert in-memory I0421 19:17:06.376968 1 controllermanager.go:189] "Starting" version="v1.28.13" I0421 19:17:06.376988 1 controllermanager.go:191] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" I0421 19:17:06.378244 1 secure_serving.go:213] Serving securely on [::]:10257 I0421 19:17:06.378247 1 dynamic_cafile_content.go:157] "Starting controller" name="request-header::/etc/kubernetes/pki/front-proxy-ca.crt" I0421 19:17:06.378396 1 leaderelection.go:250] attempting to acquire leader lease kube-system/kube-controller-manager... I0421 19:17:06.378452 1 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" I0421 19:17:06.378622 1 tlsconfig.go:240] "Starting DynamicServingCertificateController" E0421 19:17:07.466981 1 leaderelection.go:332] error retrieving resource lock kube-system/kube-controller-manager: leases.coordination.k8s.io "kube-controller-manager" is forbidden: User "system:kube-controller-manager" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-system" I0421 19:17:11.494596 1 leaderelection.go:260] successfully acquired lease kube-system/kube-controller-manager I0421 19:17:11.495416 1 event.go:307] "Event occurred" object="kube-system/kube-controller-manager" fieldPath="" kind="Lease" apiVersion="coordination.k8s.io/v1" type="Normal" reason="LeaderElection" message="instance_922be1a9-2cbb-44dd-a2fc-d3f02d18abfd became leader" I0421 19:17:11.504033 1 shared_informer.go:311] Waiting for caches to sync for tokens I0421 19:17:11.512593 1 controllermanager.go:642] "Started controller" controller="replicaset-controller" I0421 19:17:11.512762 1 replica_set.go:214] "Starting controller" name="replicaset" I0421 19:17:11.512784 1 shared_informer.go:311] Waiting for caches to sync for ReplicaSet I0421 19:17:11.518104 1 controllermanager.go:642] "Started controller" controller="ttl-after-finished-controller" I0421 19:17:11.518335 1 ttlafterfinished_controller.go:109] "Starting TTL after finished controller" I0421 19:17:11.518352 1 shared_informer.go:311] Waiting for caches to sync for TTL after finished I0421 19:17:11.525618 1 controllermanager.go:642] "Started controller" controller="token-cleaner-controller" I0421 19:17:11.525672 1 tokencleaner.go:112] "Starting token cleaner controller" I0421 19:17:11.525724 1 shared_informer.go:311] Waiting for caches to sync for token_cleaner I0421 19:17:11.525735 1 shared_informer.go:318] Caches are synced for token_cleaner I0421 19:17:11.525682 1 controllermanager.go:620] "Warning: skipping controller" controller="node-ipam-controller" I0421 19:17:11.525874 1 core.go:224] "Will not configure cloud provider routes for allocate-node-cidrs" CIDRs=false routes=true I0421 19:17:11.525906 1 controllermanager.go:620] "Warning: skipping controller" controller="node-route-controller" I0421 19:17:11.538704 1 controllermanager.go:642] "Started controller" controller="endpoints-controller" I0421 19:17:11.538872 1 endpoints_controller.go:177] "Starting endpoint controller" I0421 19:17:11.538898 1 shared_informer.go:311] Waiting for caches to sync for endpoint I0421 19:17:11.545138 1 controllermanager.go:642] "Started controller" controller="pod-garbage-collector-controller" I0421 19:17:11.545423 1 gc_controller.go:101] "Starting GC controller" I0421 19:17:11.545442 1 shared_informer.go:311] Waiting for caches to sync for GC I0421 19:17:11.559904 1 controllermanager.go:642] "Started controller" controller="horizontal-pod-autoscaler-controller" I0421 19:17:11.560006 1 horizontal.go:200] "Starting HPA controller" I0421 19:17:11.560038 1 shared_informer.go:311] Waiting for caches to sync for HPA I0421 19:17:11.566954 1 controllermanager.go:642] "Started controller" controller="statefulset-controller" I0421 19:17:11.567114 1 stateful_set.go:163] "Starting stateful set controller" I0421 19:17:11.567137 1 shared_informer.go:311] Waiting for caches to sync for stateful set I0421 19:17:11.574347 1 controllermanager.go:642] "Started controller" controller="ttl-controller" I0421 19:17:11.574491 1 ttl_controller.go:124] "Starting TTL controller" I0421 19:17:11.574514 1 shared_informer.go:311] Waiting for caches to sync for TTL I0421 19:17:11.583286 1 controllermanager.go:642] "Started controller" controller="persistentvolume-attach-detach-controller" I0421 19:17:11.583573 1 attach_detach_controller.go:337] "Starting attach detach controller" I0421 19:17:11.583608 1 shared_informer.go:311] Waiting for caches to sync for attach detach I0421 19:17:11.604968 1 shared_informer.go:318] Caches are synced for tokens I0421 19:17:11.700244 1 controllermanager.go:642] "Started controller" controller="ephemeral-volume-controller" I0421 19:17:11.700292 1 controller.go:169] "Starting ephemeral volume controller" I0421 19:17:11.700316 1 shared_informer.go:311] Waiting for caches to sync for ephemeral I0421 19:17:11.958882 1 controllermanager.go:642] "Started controller" controller="namespace-controller" I0421 19:17:11.958936 1 namespace_controller.go:197] "Starting namespace controller" I0421 19:17:11.958944 1 shared_informer.go:311] Waiting for caches to sync for namespace I0421 19:17:12.100422 1 controllermanager.go:642] "Started controller" controller="job-controller" I0421 19:17:12.100492 1 job_controller.go:226] "Starting job controller" I0421 19:17:12.100499 1 shared_informer.go:311] Waiting for caches to sync for job E0421 19:17:12.249451 1 core.go:213] "Failed to start cloud node lifecycle controller" err="no cloud provider provided" I0421 19:17:12.249477 1 controllermanager.go:620] "Warning: skipping controller" controller="cloud-node-lifecycle-controller" I0421 19:17:12.400817 1 controllermanager.go:642] "Started controller" controller="persistentvolume-binder-controller" I0421 19:17:12.400992 1 pv_controller_base.go:319] "Starting persistent volume controller" I0421 19:17:12.401025 1 shared_informer.go:311] Waiting for caches to sync for persistent volume I0421 19:17:12.550508 1 controllermanager.go:642] "Started controller" controller="clusterrole-aggregation-controller" I0421 19:17:12.550595 1 clusterroleaggregation_controller.go:189] "Starting ClusterRoleAggregator controller" I0421 19:17:12.550622 1 shared_informer.go:311] Waiting for caches to sync for ClusterRoleAggregator I0421 19:17:12.699998 1 controllermanager.go:642] "Started controller" controller="cronjob-controller" I0421 19:17:12.700113 1 cronjob_controllerv2.go:139] "Starting cronjob controller v2" I0421 19:17:12.700150 1 shared_informer.go:311] Waiting for caches to sync for cronjob I0421 19:17:12.749253 1 node_lifecycle_controller.go:431] "Controller will reconcile labels" I0421 19:17:12.749321 1 controllermanager.go:642] "Started controller" controller="node-lifecycle-controller" I0421 19:17:12.749615 1 node_lifecycle_controller.go:465] "Sending events to api server" I0421 19:17:12.749648 1 node_lifecycle_controller.go:476] "Starting node controller" I0421 19:17:12.749660 1 shared_informer.go:311] Waiting for caches to sync for taint I0421 19:17:12.908148 1 controllermanager.go:642] "Started controller" controller="daemonset-controller" I0421 19:17:12.908430 1 daemon_controller.go:291] "Starting daemon sets controller" I0421 19:17:12.908466 1 shared_informer.go:311] Waiting for caches to sync for daemon sets I0421 19:17:13.050801 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-kubelet-serving" I0421 19:17:13.050821 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-kubelet-serving I0421 19:17:13.050837 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0421 19:17:13.051261 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-kubelet-client" I0421 19:17:13.051277 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-kubelet-client I0421 19:17:13.051303 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0421 19:17:13.051905 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-kube-apiserver-client" I0421 19:17:13.051932 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-kube-apiserver-client I0421 19:17:13.051971 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0421 19:17:13.052255 1 controllermanager.go:642] "Started controller" controller="certificatesigningrequest-signing-controller" I0421 19:17:13.052312 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-legacy-unknown" I0421 19:17:13.052321 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-legacy-unknown I0421 19:17:13.052348 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0421 19:17:13.098454 1 controllermanager.go:642] "Started controller" controller="certificatesigningrequest-approving-controller" I0421 19:17:13.098538 1 certificate_controller.go:115] "Starting certificate controller" name="csrapproving" I0421 19:17:13.098553 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrapproving E0421 19:17:13.250597 1 core.go:92] "Failed to start service controller" err="WARNING: no cloud provider provided, services of type LoadBalancer will fail" I0421 19:17:13.250820 1 controllermanager.go:620] "Warning: skipping controller" controller="service-lb-controller" I0421 19:17:13.399638 1 controllermanager.go:642] "Started controller" controller="root-ca-certificate-publisher-controller" I0421 19:17:13.399745 1 publisher.go:102] "Starting root CA cert publisher controller" I0421 19:17:13.399757 1 shared_informer.go:311] Waiting for caches to sync for crt configmap I0421 19:17:13.550642 1 controllermanager.go:642] "Started controller" controller="endpointslice-mirroring-controller" I0421 19:17:13.550800 1 endpointslicemirroring_controller.go:223] "Starting EndpointSliceMirroring controller" I0421 19:17:13.550828 1 shared_informer.go:311] Waiting for caches to sync for endpoint_slice_mirroring I0421 19:17:13.699655 1 controllermanager.go:642] "Started controller" controller="deployment-controller" I0421 19:17:13.699909 1 deployment_controller.go:168] "Starting controller" controller="deployment" I0421 19:17:13.699953 1 shared_informer.go:311] Waiting for caches to sync for deployment I0421 19:17:13.850833 1 controllermanager.go:642] "Started controller" controller="bootstrap-signer-controller" I0421 19:17:13.851054 1 shared_informer.go:311] Waiting for caches to sync for bootstrap_signer I0421 19:17:13.999812 1 controllermanager.go:642] "Started controller" controller="persistentvolume-protection-controller" I0421 19:17:13.999916 1 pv_protection_controller.go:78] "Starting PV protection controller" I0421 19:17:13.999948 1 shared_informer.go:311] Waiting for caches to sync for PV protection I0421 19:17:14.307435 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="deployments.apps" I0421 19:17:14.307508 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="jobs.batch" I0421 19:17:14.307534 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="cronjobs.batch" I0421 19:17:14.307569 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="rolebindings.rbac.authorization.k8s.io" I0421 19:17:14.307614 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="csistoragecapacities.storage.k8s.io" I0421 19:17:14.307637 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="limitranges" I0421 19:17:14.307669 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="replicasets.apps" I0421 19:17:14.307697 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="leases.coordination.k8s.io" I0421 19:17:14.307734 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="endpoints" I0421 19:17:14.307812 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="serviceaccounts" I0421 19:17:14.307856 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="endpointslices.discovery.k8s.io" W0421 19:17:14.307873 1 shared_informer.go:593] resyncPeriod 12h20m5.4483969s is smaller than resyncCheckPeriod 14h46m23.196848172s and the informer has already started. Changing it to 14h46m23.196848172s I0421 19:17:14.307935 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="horizontalpodautoscalers.autoscaling" I0421 19:17:14.307965 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="poddisruptionbudgets.policy" I0421 19:17:14.307990 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="podtemplates" I0421 19:17:14.308028 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ingresses.networking.k8s.io" I0421 19:17:14.308061 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="roles.rbac.authorization.k8s.io" I0421 19:17:14.308113 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="controllerrevisions.apps" I0421 19:17:14.308137 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="daemonsets.apps" I0421 19:17:14.308160 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="statefulsets.apps" I0421 19:17:14.308207 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="networkpolicies.networking.k8s.io" I0421 19:17:14.308239 1 controllermanager.go:642] "Started controller" controller="resourcequota-controller" I0421 19:17:14.308412 1 resource_quota_controller.go:294] "Starting resource quota controller" I0421 19:17:14.308451 1 shared_informer.go:311] Waiting for caches to sync for resource quota I0421 19:17:14.308531 1 resource_quota_monitor.go:305] "QuotaMonitor running" I0421 19:17:14.451114 1 controllermanager.go:642] "Started controller" controller="serviceaccount-controller" I0421 19:17:14.451252 1 serviceaccounts_controller.go:111] "Starting service account controller" I0421 19:17:14.451272 1 shared_informer.go:311] Waiting for caches to sync for service account I0421 19:17:14.698926 1 controllermanager.go:642] "Started controller" controller="garbage-collector-controller" I0421 19:17:14.698950 1 garbagecollector.go:155] "Starting controller" controller="garbagecollector" I0421 19:17:14.698982 1 shared_informer.go:311] Waiting for caches to sync for garbage collector I0421 19:17:14.699031 1 graph_builder.go:294] "Running" component="GraphBuilder" I0421 19:17:15.001428 1 controllermanager.go:642] "Started controller" controller="disruption-controller" I0421 19:17:15.001470 1 disruption.go:433] "Sending events to api server." I0421 19:17:15.001520 1 disruption.go:444] "Starting disruption controller" I0421 19:17:15.001534 1 shared_informer.go:311] Waiting for caches to sync for disruption I0421 19:17:15.048690 1 controllermanager.go:642] "Started controller" controller="certificatesigningrequest-cleaner-controller" I0421 19:17:15.048808 1 cleaner.go:83] "Starting CSR cleaner controller" I0421 19:17:15.201149 1 controllermanager.go:642] "Started controller" controller="endpointslice-controller" I0421 19:17:15.201297 1 endpointslice_controller.go:264] "Starting endpoint slice controller" I0421 19:17:15.201433 1 shared_informer.go:311] Waiting for caches to sync for endpoint_slice I0421 19:17:15.351525 1 controllermanager.go:642] "Started controller" controller="replicationcontroller-controller" I0421 19:17:15.351840 1 replica_set.go:214] "Starting controller" name="replicationcontroller" I0421 19:17:15.351877 1 shared_informer.go:311] Waiting for caches to sync for ReplicationController I0421 19:17:15.505416 1 controllermanager.go:642] "Started controller" controller="persistentvolume-expander-controller" I0421 19:17:15.505491 1 expand_controller.go:328] "Starting expand controller" I0421 19:17:15.505501 1 shared_informer.go:311] Waiting for caches to sync for expand I0421 19:17:15.650600 1 controllermanager.go:642] "Started controller" controller="persistentvolumeclaim-protection-controller" I0421 19:17:15.653552 1 pvc_protection_controller.go:102] "Starting PVC protection controller" I0421 19:17:15.653575 1 shared_informer.go:311] Waiting for caches to sync for PVC protection I0421 19:17:15.657059 1 shared_informer.go:311] Waiting for caches to sync for resource quota I0421 19:17:15.663303 1 actual_state_of_world.go:547] "Failed to update statusUpdateNeeded field in actual state of world" err="Failed to set statusUpdateNeeded to needed true, because nodeName=\"instance\" does not exist" I0421 19:17:15.663739 1 shared_informer.go:318] Caches are synced for namespace I0421 19:17:15.669440 1 shared_informer.go:311] Waiting for caches to sync for garbage collector I0421 19:17:15.675620 1 shared_informer.go:318] Caches are synced for TTL I0421 19:17:15.683957 1 shared_informer.go:318] Caches are synced for attach detach I0421 19:17:15.699284 1 shared_informer.go:318] Caches are synced for certificate-csrapproving I0421 19:17:15.699951 1 shared_informer.go:318] Caches are synced for crt configmap I0421 19:17:15.699987 1 shared_informer.go:318] Caches are synced for deployment I0421 19:17:15.700095 1 shared_informer.go:318] Caches are synced for PV protection I0421 19:17:15.700591 1 shared_informer.go:318] Caches are synced for ephemeral I0421 19:17:15.700635 1 shared_informer.go:318] Caches are synced for job I0421 19:17:15.701201 1 shared_informer.go:318] Caches are synced for persistent volume I0421 19:17:15.701503 1 shared_informer.go:318] Caches are synced for endpoint_slice I0421 19:17:15.701681 1 shared_informer.go:318] Caches are synced for disruption I0421 19:17:15.706159 1 shared_informer.go:318] Caches are synced for expand I0421 19:17:15.712977 1 shared_informer.go:318] Caches are synced for ReplicaSet I0421 19:17:15.719291 1 shared_informer.go:318] Caches are synced for TTL after finished I0421 19:17:15.739509 1 shared_informer.go:318] Caches are synced for endpoint I0421 19:17:15.745794 1 shared_informer.go:318] Caches are synced for GC I0421 19:17:15.752406 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-kube-apiserver-client I0421 19:17:15.753643 1 shared_informer.go:318] Caches are synced for PVC protection I0421 19:17:15.753790 1 shared_informer.go:318] Caches are synced for service account I0421 19:17:15.753820 1 shared_informer.go:318] Caches are synced for ReplicationController I0421 19:17:15.753841 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-kubelet-serving I0421 19:17:15.753858 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-kubelet-client I0421 19:17:15.753933 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-legacy-unknown I0421 19:17:15.753948 1 shared_informer.go:318] Caches are synced for bootstrap_signer I0421 19:17:15.754009 1 shared_informer.go:318] Caches are synced for ClusterRoleAggregator I0421 19:17:15.754053 1 shared_informer.go:318] Caches are synced for endpoint_slice_mirroring I0421 19:17:15.760414 1 shared_informer.go:318] Caches are synced for HPA I0421 19:17:15.767257 1 shared_informer.go:318] Caches are synced for stateful set I0421 19:17:15.800906 1 shared_informer.go:318] Caches are synced for cronjob I0421 19:17:15.809016 1 shared_informer.go:318] Caches are synced for daemon sets I0421 19:17:15.850387 1 shared_informer.go:318] Caches are synced for taint I0421 19:17:15.850561 1 node_lifecycle_controller.go:1225] "Initializing eviction metric for zone" zone="" I0421 19:17:15.850689 1 node_lifecycle_controller.go:877] "Missing timestamp for Node. Assuming now as a timestamp" node="instance" I0421 19:17:15.850765 1 node_lifecycle_controller.go:1029] "Controller detected that all Nodes are not-Ready. Entering master disruption mode" I0421 19:17:15.850823 1 taint_manager.go:205] "Starting NoExecuteTaintManager" I0421 19:17:15.850936 1 taint_manager.go:210] "Sending events to api server" I0421 19:17:15.851061 1 event.go:307] "Event occurred" object="instance" fieldPath="" kind="Node" apiVersion="v1" type="Normal" reason="RegisteredNode" message="Node instance event: Registered Node instance in Controller" I0421 19:17:15.857569 1 shared_informer.go:318] Caches are synced for resource quota I0421 19:17:15.908650 1 shared_informer.go:318] Caches are synced for resource quota I0421 19:17:16.269593 1 shared_informer.go:318] Caches are synced for garbage collector I0421 19:17:16.299501 1 shared_informer.go:318] Caches are synced for garbage collector I0421 19:17:16.299598 1 garbagecollector.go:166] "All resource monitors have synced. Proceeding to collect garbage" I0421 19:17:16.415006 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set coredns-7c96b6546b to 2" I0421 19:17:16.716232 1 event.go:307] "Event occurred" object="kube-system/kube-proxy" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: kube-proxy-x8fth" I0421 19:17:16.864195 1 event.go:307] "Event occurred" object="kube-system/coredns-7c96b6546b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-7c96b6546b-ctkjr" I0421 19:17:16.876823 1 event.go:307] "Event occurred" object="kube-system/coredns-7c96b6546b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-7c96b6546b-x5228" I0421 19:17:16.887611 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="473.787564ms" I0421 19:17:16.926054 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="38.371748ms" I0421 19:17:16.926124 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="39.501µs" I0421 19:17:21.910441 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set coredns-67659f764b to 1" I0421 19:17:21.917340 1 event.go:307] "Event occurred" object="kube-system/coredns-67659f764b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-67659f764b-zl69j" I0421 19:17:21.926325 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="15.523443ms" I0421 19:17:21.928145 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled down replica set coredns-7c96b6546b to 1 from 2" I0421 19:17:21.935108 1 event.go:307] "Event occurred" object="kube-system/coredns-7c96b6546b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulDelete" message="Deleted pod: coredns-7c96b6546b-x5228" I0421 19:17:21.940317 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set coredns-67659f764b to 2 from 1" I0421 19:17:21.944638 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="17.008653ms" I0421 19:17:21.948356 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="21.990319ms" E0421 19:17:21.948397 1 replica_set.go:557] sync "kube-system/coredns-67659f764b" failed with Operation cannot be fulfilled on replicasets.apps "coredns-67659f764b": the object has been modified; please apply your changes to the latest version and try again I0421 19:17:21.953522 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="8.843062ms" I0421 19:17:21.953621 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="63.851µs" I0421 19:17:21.957028 1 event.go:307] "Event occurred" object="kube-system/coredns-67659f764b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-67659f764b-vskz2" I0421 19:17:21.962629 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="14.176766ms" I0421 19:17:21.972101 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="9.428463ms" I0421 19:17:21.972165 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="39.891µs" I0421 19:17:34.186635 1 event.go:307] "Event occurred" object="kube-system/cilium-operator" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set cilium-operator-869df985b8 to 1" I0421 19:17:34.191614 1 event.go:307] "Event occurred" object="kube-system/cilium" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: cilium-69fwr" I0421 19:17:34.194639 1 event.go:307] "Event occurred" object="kube-system/cilium-operator-869df985b8" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: cilium-operator-869df985b8-k9762" I0421 19:17:34.209065 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="22.725642ms" I0421 19:17:34.217904 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="8.777186ms" I0421 19:17:34.218029 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="82.612µs" I0421 19:17:34.234337 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="87.692µs" I0421 19:17:42.806145 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:17:42.823369 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:17:42.823492 1 event.go:307] "Event occurred" object="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" fieldPath="" kind="Job" apiVersion="batch/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: envoy-gateway-gateway-helm-certgen-vcl74" I0421 19:17:42.832646 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:17:42.835938 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:17:44.047636 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="15.192419ms" I0421 19:17:44.047752 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="84.842µs" I0421 19:17:45.865540 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="backends.gateway.envoyproxy.io" I0421 19:17:45.865581 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="securitypolicies.gateway.envoyproxy.io" I0421 19:17:45.865603 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="gateways.gateway.networking.k8s.io" I0421 19:17:45.865631 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="envoyextensionpolicies.gateway.envoyproxy.io" I0421 19:17:45.865660 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="httproutes.gateway.networking.k8s.io" I0421 19:17:45.865685 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="referencegrants.gateway.networking.k8s.io" I0421 19:17:45.865715 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="backendtrafficpolicies.gateway.envoyproxy.io" I0421 19:17:45.865847 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="udproutes.gateway.networking.k8s.io" I0421 19:17:45.865874 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="httproutefilters.gateway.envoyproxy.io" I0421 19:17:45.865964 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ciliumendpoints.cilium.io" I0421 19:17:45.866084 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ciliumnodeconfigs.cilium.io" I0421 19:17:45.866188 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="tlsroutes.gateway.networking.k8s.io" I0421 19:17:45.866225 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="tcproutes.gateway.networking.k8s.io" I0421 19:17:45.866252 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="backendtlspolicies.gateway.networking.k8s.io" I0421 19:17:45.866282 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="clienttrafficpolicies.gateway.envoyproxy.io" I0421 19:17:45.866358 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="envoypatchpolicies.gateway.envoyproxy.io" I0421 19:17:45.866439 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="xlistenersets.gateway.networking.x-k8s.io" I0421 19:17:45.866502 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="xbackendtrafficpolicies.gateway.networking.x-k8s.io" I0421 19:17:45.866535 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="envoyproxies.gateway.envoyproxy.io" I0421 19:17:45.866629 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="grpcroutes.gateway.networking.k8s.io" I0421 19:17:45.866848 1 shared_informer.go:311] Waiting for caches to sync for resource quota I0421 19:17:46.283014 1 shared_informer.go:311] Waiting for caches to sync for garbage collector I0421 19:17:47.067597 1 shared_informer.go:318] Caches are synced for resource quota I0421 19:17:47.284192 1 shared_informer.go:318] Caches are synced for garbage collector I0421 19:17:48.754026 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="83.362µs" I0421 19:17:48.757524 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:17:48.768559 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="294.526µs" I0421 19:17:48.778870 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:17:50.856636 1 node_lifecycle_controller.go:1048] "Controller detected that some Nodes are Ready. Exiting master disruption mode" I0421 19:17:54.071745 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="242.985µs" I0421 19:17:54.093021 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="9.378499ms" I0421 19:17:54.093798 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="663.692µs" I0421 19:17:58.094206 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:18:00.089139 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:18:01.097687 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:18:01.110116 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:18:01.121690 1 event.go:307] "Event occurred" object="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" fieldPath="" kind="Job" apiVersion="batch/v1" type="Normal" reason="Completed" message="Job completed" I0421 19:18:01.121895 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:18:01.174306 1 event.go:307] "Event occurred" object="envoy-gateway-system/envoy-gateway" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set envoy-gateway-78446f96c9 to 1" I0421 19:18:01.192880 1 event.go:307] "Event occurred" object="envoy-gateway-system/envoy-gateway-78446f96c9" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: envoy-gateway-78446f96c9-hxj2p" I0421 19:18:01.208606 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="34.628158ms" I0421 19:18:01.224397 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="15.740868ms" I0421 19:18:01.224505 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="77.331µs" I0421 19:18:01.224604 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="54.071µs" I0421 19:18:02.105889 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="132.872µs" I0421 19:18:04.411400 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="65.851µs" I0421 19:18:04.411933 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="53.281µs" I0421 19:18:04.426617 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="163.353µs" I0421 19:18:04.440281 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="73.151µs" I0421 19:18:06.127359 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="77.981µs" I0421 19:18:06.149260 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="8.908859ms" I0421 19:18:06.149425 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="67.711µs" I0421 19:18:06.161792 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled down replica set coredns-7c96b6546b to 0 from 1" I0421 19:18:06.169891 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="12.09032ms" I0421 19:18:06.170047 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="115.933µs" I0421 19:18:06.175602 1 event.go:307] "Event occurred" object="kube-system/coredns-7c96b6546b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulDelete" message="Deleted pod: coredns-7c96b6546b-ctkjr" I0421 19:18:06.185604 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="23.945924ms" I0421 19:18:06.199629 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="13.971555ms" I0421 19:18:06.199757 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="83.581µs" I0421 19:18:06.199802 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="24.41µs" I0421 19:18:11.448220 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="102.622µs" I0421 19:18:11.533109 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="11.115091ms" I0421 19:18:11.533207 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="66.931µs" I0421 19:18:12.157088 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="101.642µs" I0421 19:18:12.164509 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="93.471µs" I0421 19:18:17.075079 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ciliumnetworkpolicies.cilium.io" I0421 19:18:17.075182 1 shared_informer.go:311] Waiting for caches to sync for resource quota I0421 19:18:17.176053 1 shared_informer.go:318] Caches are synced for resource quota I0421 19:18:17.300031 1 shared_informer.go:311] Waiting for caches to sync for garbage collector I0421 19:18:17.400590 1 shared_informer.go:318] Caches are synced for garbage collector I0421 19:18:31.023882 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:18:31.051529 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0421 19:19:06.046857 1 event.go:307] "Event occurred" object="kube-system/ceph-csi-rbd-provisioner" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set ceph-csi-rbd-provisioner-69cb97fb6 to 1" I0421 19:19:06.054682 1 event.go:307] "Event occurred" object="kube-system/ceph-csi-rbd-nodeplugin" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: ceph-csi-rbd-nodeplugin-ngzfj" I0421 19:19:06.060939 1 event.go:307] "Event occurred" object="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: ceph-csi-rbd-provisioner-69cb97fb6-c7n2q" I0421 19:19:06.073238 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="27.574351ms" I0421 19:19:06.085443 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="12.014487ms" I0421 19:19:06.085626 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="135.862µs" I0421 19:19:06.108167 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="108.982µs" I0421 19:19:07.332687 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="85.482µs" I0421 19:19:19.905318 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="110.592µs" I0421 19:19:31.903910 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="134.542µs" I0421 19:19:43.906758 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="125.773µs" I0421 19:19:57.901198 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="120.452µs" I0421 19:20:11.907711 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="117.702µs" I0421 19:20:40.905232 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="72.302µs" I0421 19:20:54.926092 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="69.421µs" I0421 19:22:11.903625 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="142.962µs" I0421 19:22:22.905240 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="271.175µs" I0421 19:22:36.055085 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:22:45.717017 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:23:00.717156 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:23:15.718325 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:23:30.717570 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:23:45.717851 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:24:00.718253 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:24:15.719165 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:24:30.720307 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:24:45.720262 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:25:00.720477 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:25:02.900529 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="195.785µs" I0421 19:25:15.721395 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0421 19:25:15.904971 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="135.932µs"