I0423 07:44:03.945273 1 serving.go:348] Generated self-signed cert in-memory I0423 07:44:04.209173 1 controllermanager.go:189] "Starting" version="v1.28.13" I0423 07:44:04.209196 1 controllermanager.go:191] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" I0423 07:44:04.226180 1 dynamic_cafile_content.go:157] "Starting controller" name="request-header::/etc/kubernetes/pki/front-proxy-ca.crt" I0423 07:44:04.226245 1 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" I0423 07:44:04.226773 1 secure_serving.go:213] Serving securely on [::]:10257 I0423 07:44:04.226884 1 tlsconfig.go:240] "Starting DynamicServingCertificateController" I0423 07:44:04.227236 1 leaderelection.go:250] attempting to acquire leader lease kube-system/kube-controller-manager... E0423 07:44:05.829865 1 leaderelection.go:332] error retrieving resource lock kube-system/kube-controller-manager: leases.coordination.k8s.io "kube-controller-manager" is forbidden: User "system:kube-controller-manager" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-system" E0423 07:44:08.002831 1 leaderelection.go:332] error retrieving resource lock kube-system/kube-controller-manager: leases.coordination.k8s.io "kube-controller-manager" is forbidden: User "system:kube-controller-manager" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-system" I0423 07:44:11.622961 1 leaderelection.go:260] successfully acquired lease kube-system/kube-controller-manager I0423 07:44:11.623363 1 event.go:307] "Event occurred" object="kube-system/kube-controller-manager" fieldPath="" kind="Lease" apiVersion="coordination.k8s.io/v1" type="Normal" reason="LeaderElection" message="instance_b74c7a6e-2d31-46a0-983e-39660a41bea8 became leader" I0423 07:44:11.630334 1 shared_informer.go:311] Waiting for caches to sync for tokens I0423 07:44:11.730916 1 shared_informer.go:318] Caches are synced for tokens I0423 07:44:14.047043 1 controllermanager.go:642] "Started controller" controller="replicaset-controller" I0423 07:44:14.047173 1 replica_set.go:214] "Starting controller" name="replicaset" I0423 07:44:14.047186 1 shared_informer.go:311] Waiting for caches to sync for ReplicaSet I0423 07:44:14.122184 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-kubelet-serving" I0423 07:44:14.122223 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-kubelet-serving I0423 07:44:14.122399 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0423 07:44:14.122845 1 controllermanager.go:642] "Started controller" controller="certificatesigningrequest-signing-controller" I0423 07:44:14.123007 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-kubelet-client" I0423 07:44:14.123030 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-kubelet-client I0423 07:44:14.123043 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-kube-apiserver-client" I0423 07:44:14.123062 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0423 07:44:14.123087 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0423 07:44:14.123063 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-kube-apiserver-client I0423 07:44:14.123186 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-legacy-unknown" I0423 07:44:14.123199 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-legacy-unknown I0423 07:44:14.123401 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0423 07:44:14.379100 1 node_lifecycle_controller.go:431] "Controller will reconcile labels" I0423 07:44:14.379175 1 controllermanager.go:642] "Started controller" controller="node-lifecycle-controller" I0423 07:44:14.379418 1 node_lifecycle_controller.go:465] "Sending events to api server" I0423 07:44:14.379464 1 node_lifecycle_controller.go:476] "Starting node controller" I0423 07:44:14.379472 1 shared_informer.go:311] Waiting for caches to sync for taint I0423 07:44:15.847796 1 controllermanager.go:642] "Started controller" controller="namespace-controller" I0423 07:44:15.847879 1 namespace_controller.go:197] "Starting namespace controller" I0423 07:44:15.847904 1 shared_informer.go:311] Waiting for caches to sync for namespace I0423 07:44:16.058545 1 controllermanager.go:642] "Started controller" controller="ttl-controller" I0423 07:44:16.058738 1 ttl_controller.go:124] "Starting TTL controller" I0423 07:44:16.058761 1 shared_informer.go:311] Waiting for caches to sync for TTL I0423 07:44:16.067144 1 controllermanager.go:642] "Started controller" controller="persistentvolume-attach-detach-controller" I0423 07:44:16.067336 1 attach_detach_controller.go:337] "Starting attach detach controller" I0423 07:44:16.067349 1 shared_informer.go:311] Waiting for caches to sync for attach detach I0423 07:44:16.075525 1 controllermanager.go:642] "Started controller" controller="persistentvolumeclaim-protection-controller" I0423 07:44:16.075676 1 pvc_protection_controller.go:102] "Starting PVC protection controller" I0423 07:44:16.075705 1 shared_informer.go:311] Waiting for caches to sync for PVC protection I0423 07:44:16.082124 1 controllermanager.go:642] "Started controller" controller="daemonset-controller" I0423 07:44:16.082622 1 daemon_controller.go:291] "Starting daemon sets controller" I0423 07:44:16.082649 1 shared_informer.go:311] Waiting for caches to sync for daemon sets I0423 07:44:16.091821 1 controllermanager.go:642] "Started controller" controller="cronjob-controller" I0423 07:44:16.092066 1 cronjob_controllerv2.go:139] "Starting cronjob controller v2" I0423 07:44:16.092091 1 shared_informer.go:311] Waiting for caches to sync for cronjob I0423 07:44:16.093975 1 controllermanager.go:642] "Started controller" controller="certificatesigningrequest-cleaner-controller" I0423 07:44:16.094072 1 cleaner.go:83] "Starting CSR cleaner controller" I0423 07:44:16.102677 1 controllermanager.go:642] "Started controller" controller="bootstrap-signer-controller" I0423 07:44:16.102808 1 shared_informer.go:311] Waiting for caches to sync for bootstrap_signer I0423 07:44:16.110660 1 controllermanager.go:642] "Started controller" controller="token-cleaner-controller" I0423 07:44:16.110770 1 tokencleaner.go:112] "Starting token cleaner controller" I0423 07:44:16.110878 1 shared_informer.go:311] Waiting for caches to sync for token_cleaner I0423 07:44:16.110892 1 shared_informer.go:318] Caches are synced for token_cleaner E0423 07:44:16.118896 1 core.go:92] "Failed to start service controller" err="WARNING: no cloud provider provided, services of type LoadBalancer will fail" I0423 07:44:16.118933 1 controllermanager.go:620] "Warning: skipping controller" controller="service-lb-controller" I0423 07:44:16.126475 1 controllermanager.go:642] "Started controller" controller="endpoints-controller" I0423 07:44:16.126584 1 endpoints_controller.go:177] "Starting endpoint controller" I0423 07:44:16.126632 1 shared_informer.go:311] Waiting for caches to sync for endpoint I0423 07:44:16.134619 1 controllermanager.go:642] "Started controller" controller="pod-garbage-collector-controller" I0423 07:44:16.134696 1 gc_controller.go:101] "Starting GC controller" I0423 07:44:16.134711 1 shared_informer.go:311] Waiting for caches to sync for GC I0423 07:44:16.203122 1 controllermanager.go:642] "Started controller" controller="serviceaccount-controller" I0423 07:44:16.203218 1 serviceaccounts_controller.go:111] "Starting service account controller" I0423 07:44:16.203232 1 shared_informer.go:311] Waiting for caches to sync for service account I0423 07:44:16.353792 1 controllermanager.go:642] "Started controller" controller="persistentvolume-protection-controller" I0423 07:44:16.353936 1 pv_protection_controller.go:78] "Starting PV protection controller" I0423 07:44:16.353992 1 shared_informer.go:311] Waiting for caches to sync for PV protection I0423 07:44:16.505668 1 controllermanager.go:642] "Started controller" controller="ephemeral-volume-controller" I0423 07:44:16.505756 1 controller.go:169] "Starting ephemeral volume controller" I0423 07:44:16.505772 1 shared_informer.go:311] Waiting for caches to sync for ephemeral I0423 07:44:16.657701 1 controllermanager.go:642] "Started controller" controller="clusterrole-aggregation-controller" I0423 07:44:16.658066 1 clusterroleaggregation_controller.go:189] "Starting ClusterRoleAggregator controller" I0423 07:44:16.658276 1 shared_informer.go:311] Waiting for caches to sync for ClusterRoleAggregator I0423 07:44:16.807839 1 controllermanager.go:642] "Started controller" controller="endpointslice-controller" I0423 07:44:16.807966 1 endpointslice_controller.go:264] "Starting endpoint slice controller" I0423 07:44:16.807982 1 shared_informer.go:311] Waiting for caches to sync for endpoint_slice I0423 07:44:16.953173 1 controllermanager.go:642] "Started controller" controller="replicationcontroller-controller" I0423 07:44:16.953338 1 replica_set.go:214] "Starting controller" name="replicationcontroller" I0423 07:44:16.953359 1 shared_informer.go:311] Waiting for caches to sync for ReplicationController I0423 07:44:17.260209 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="statefulsets.apps" I0423 07:44:17.260261 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="horizontalpodautoscalers.autoscaling" I0423 07:44:17.260296 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="networkpolicies.networking.k8s.io" I0423 07:44:17.260351 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="jobs.batch" I0423 07:44:17.260382 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ingresses.networking.k8s.io" I0423 07:44:17.260438 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="serviceaccounts" I0423 07:44:17.260473 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="deployments.apps" I0423 07:44:17.260528 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="controllerrevisions.apps" I0423 07:44:17.260556 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="replicasets.apps" I0423 07:44:17.260585 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="roles.rbac.authorization.k8s.io" I0423 07:44:17.260632 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="endpoints" I0423 07:44:17.260660 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="leases.coordination.k8s.io" I0423 07:44:17.260702 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="podtemplates" I0423 07:44:17.260776 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="rolebindings.rbac.authorization.k8s.io" I0423 07:44:17.260802 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="limitranges" I0423 07:44:17.260825 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="cronjobs.batch" I0423 07:44:17.260858 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="csistoragecapacities.storage.k8s.io" I0423 07:44:17.260882 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="endpointslices.discovery.k8s.io" I0423 07:44:17.260925 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="daemonsets.apps" I0423 07:44:17.260953 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="poddisruptionbudgets.policy" I0423 07:44:17.260981 1 controllermanager.go:642] "Started controller" controller="resourcequota-controller" I0423 07:44:17.261061 1 resource_quota_controller.go:294] "Starting resource quota controller" I0423 07:44:17.261121 1 shared_informer.go:311] Waiting for caches to sync for resource quota I0423 07:44:17.261191 1 resource_quota_monitor.go:305] "QuotaMonitor running" I0423 07:44:17.403229 1 controllermanager.go:642] "Started controller" controller="deployment-controller" I0423 07:44:17.403423 1 deployment_controller.go:168] "Starting controller" controller="deployment" I0423 07:44:17.403445 1 shared_informer.go:311] Waiting for caches to sync for deployment I0423 07:44:17.553612 1 controllermanager.go:642] "Started controller" controller="statefulset-controller" I0423 07:44:17.553736 1 controllermanager.go:620] "Warning: skipping controller" controller="node-ipam-controller" I0423 07:44:17.553735 1 stateful_set.go:163] "Starting stateful set controller" I0423 07:44:17.553914 1 shared_informer.go:311] Waiting for caches to sync for stateful set I0423 07:44:17.703608 1 controllermanager.go:642] "Started controller" controller="persistentvolume-binder-controller" I0423 07:44:17.703752 1 pv_controller_base.go:319] "Starting persistent volume controller" I0423 07:44:17.703769 1 shared_informer.go:311] Waiting for caches to sync for persistent volume E0423 07:44:17.751582 1 core.go:213] "Failed to start cloud node lifecycle controller" err="no cloud provider provided" I0423 07:44:17.751682 1 controllermanager.go:620] "Warning: skipping controller" controller="cloud-node-lifecycle-controller" I0423 07:44:17.905528 1 controllermanager.go:642] "Started controller" controller="ttl-after-finished-controller" I0423 07:44:17.905618 1 ttlafterfinished_controller.go:109] "Starting TTL after finished controller" I0423 07:44:17.905633 1 shared_informer.go:311] Waiting for caches to sync for TTL after finished I0423 07:44:18.054877 1 controllermanager.go:642] "Started controller" controller="endpointslice-mirroring-controller" I0423 07:44:18.055125 1 endpointslicemirroring_controller.go:223] "Starting EndpointSliceMirroring controller" I0423 07:44:18.055151 1 shared_informer.go:311] Waiting for caches to sync for endpoint_slice_mirroring I0423 07:44:18.302859 1 controllermanager.go:642] "Started controller" controller="garbage-collector-controller" I0423 07:44:18.302857 1 garbagecollector.go:155] "Starting controller" controller="garbagecollector" I0423 07:44:18.302947 1 shared_informer.go:311] Waiting for caches to sync for garbage collector I0423 07:44:18.303012 1 graph_builder.go:294] "Running" component="GraphBuilder" I0423 07:44:18.602644 1 controllermanager.go:642] "Started controller" controller="disruption-controller" I0423 07:44:18.602702 1 core.go:224] "Will not configure cloud provider routes for allocate-node-cidrs" CIDRs=false routes=true I0423 07:44:18.602716 1 controllermanager.go:620] "Warning: skipping controller" controller="node-route-controller" I0423 07:44:18.602972 1 disruption.go:433] "Sending events to api server." I0423 07:44:18.603013 1 disruption.go:444] "Starting disruption controller" I0423 07:44:18.603024 1 shared_informer.go:311] Waiting for caches to sync for disruption I0423 07:44:18.756023 1 controllermanager.go:642] "Started controller" controller="root-ca-certificate-publisher-controller" I0423 07:44:18.756205 1 publisher.go:102] "Starting root CA cert publisher controller" I0423 07:44:18.756269 1 shared_informer.go:311] Waiting for caches to sync for crt configmap I0423 07:44:18.905463 1 job_controller.go:226] "Starting job controller" I0423 07:44:18.905495 1 shared_informer.go:311] Waiting for caches to sync for job I0423 07:44:18.905329 1 controllermanager.go:642] "Started controller" controller="job-controller" I0423 07:44:19.202138 1 controllermanager.go:642] "Started controller" controller="horizontal-pod-autoscaler-controller" I0423 07:44:19.202231 1 horizontal.go:200] "Starting HPA controller" I0423 07:44:19.202273 1 shared_informer.go:311] Waiting for caches to sync for HPA I0423 07:44:19.251096 1 controllermanager.go:642] "Started controller" controller="certificatesigningrequest-approving-controller" I0423 07:44:19.251190 1 certificate_controller.go:115] "Starting certificate controller" name="csrapproving" I0423 07:44:19.251201 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrapproving I0423 07:44:19.403764 1 controllermanager.go:642] "Started controller" controller="persistentvolume-expander-controller" I0423 07:44:19.403851 1 expand_controller.go:328] "Starting expand controller" I0423 07:44:19.403871 1 shared_informer.go:311] Waiting for caches to sync for expand I0423 07:44:19.408955 1 shared_informer.go:311] Waiting for caches to sync for resource quota I0423 07:44:19.419950 1 actual_state_of_world.go:547] "Failed to update statusUpdateNeeded field in actual state of world" err="Failed to set statusUpdateNeeded to needed true, because nodeName=\"instance\" does not exist" I0423 07:44:19.426488 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-legacy-unknown I0423 07:44:19.426548 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-kube-apiserver-client I0423 07:44:19.426554 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-kubelet-serving I0423 07:44:19.426573 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-kubelet-client I0423 07:44:19.426832 1 shared_informer.go:318] Caches are synced for endpoint I0423 07:44:19.429150 1 shared_informer.go:311] Waiting for caches to sync for garbage collector I0423 07:44:19.435522 1 shared_informer.go:318] Caches are synced for GC I0423 07:44:19.447768 1 shared_informer.go:318] Caches are synced for ReplicaSet I0423 07:44:19.449039 1 shared_informer.go:318] Caches are synced for namespace I0423 07:44:19.451347 1 shared_informer.go:318] Caches are synced for certificate-csrapproving I0423 07:44:19.453954 1 shared_informer.go:318] Caches are synced for ReplicationController I0423 07:44:19.455222 1 shared_informer.go:318] Caches are synced for endpoint_slice_mirroring I0423 07:44:19.456450 1 shared_informer.go:318] Caches are synced for crt configmap I0423 07:44:19.458928 1 shared_informer.go:318] Caches are synced for TTL I0423 07:44:19.459481 1 shared_informer.go:318] Caches are synced for ClusterRoleAggregator I0423 07:44:19.476122 1 shared_informer.go:318] Caches are synced for PVC protection I0423 07:44:19.480495 1 shared_informer.go:318] Caches are synced for taint I0423 07:44:19.480639 1 taint_manager.go:205] "Starting NoExecuteTaintManager" I0423 07:44:19.480677 1 node_lifecycle_controller.go:1225] "Initializing eviction metric for zone" zone="" I0423 07:44:19.480707 1 taint_manager.go:210] "Sending events to api server" I0423 07:44:19.480782 1 node_lifecycle_controller.go:877] "Missing timestamp for Node. Assuming now as a timestamp" node="instance" I0423 07:44:19.480859 1 node_lifecycle_controller.go:1029] "Controller detected that all Nodes are not-Ready. Entering master disruption mode" I0423 07:44:19.481458 1 event.go:307] "Event occurred" object="instance" fieldPath="" kind="Node" apiVersion="v1" type="Normal" reason="RegisteredNode" message="Node instance event: Registered Node instance in Controller" I0423 07:44:19.483660 1 shared_informer.go:318] Caches are synced for daemon sets I0423 07:44:19.492963 1 shared_informer.go:318] Caches are synced for cronjob I0423 07:44:19.502631 1 shared_informer.go:318] Caches are synced for HPA I0423 07:44:19.503797 1 shared_informer.go:318] Caches are synced for service account I0423 07:44:19.503876 1 shared_informer.go:318] Caches are synced for disruption I0423 07:44:19.503966 1 shared_informer.go:318] Caches are synced for expand I0423 07:44:19.503973 1 shared_informer.go:318] Caches are synced for bootstrap_signer I0423 07:44:19.504021 1 shared_informer.go:318] Caches are synced for deployment I0423 07:44:19.505617 1 shared_informer.go:318] Caches are synced for job I0423 07:44:19.505698 1 shared_informer.go:318] Caches are synced for TTL after finished I0423 07:44:19.506957 1 shared_informer.go:318] Caches are synced for ephemeral I0423 07:44:19.508403 1 shared_informer.go:318] Caches are synced for endpoint_slice I0423 07:44:19.554778 1 shared_informer.go:318] Caches are synced for stateful set I0423 07:44:19.554875 1 shared_informer.go:318] Caches are synced for PV protection I0423 07:44:19.568480 1 shared_informer.go:318] Caches are synced for attach detach I0423 07:44:19.604787 1 shared_informer.go:318] Caches are synced for persistent volume I0423 07:44:19.661361 1 shared_informer.go:318] Caches are synced for resource quota I0423 07:44:19.709465 1 shared_informer.go:318] Caches are synced for resource quota I0423 07:44:20.029625 1 shared_informer.go:318] Caches are synced for garbage collector I0423 07:44:20.103152 1 shared_informer.go:318] Caches are synced for garbage collector I0423 07:44:20.103207 1 garbagecollector.go:166] "All resource monitors have synced. Proceeding to collect garbage" I0423 07:44:20.270286 1 event.go:307] "Event occurred" object="kube-system/kube-vip-instance" fieldPath="" kind="Pod" apiVersion="v1" type="Warning" reason="NodeNotReady" message="Node is not ready" I0423 07:44:20.272719 1 event.go:307] "Event occurred" object="kube-system/etcd-instance" fieldPath="" kind="Pod" apiVersion="v1" type="Warning" reason="NodeNotReady" message="Node is not ready" I0423 07:44:20.273121 1 event.go:307] "Event occurred" object="kube-system/kube-controller-manager-instance" fieldPath="" kind="Pod" apiVersion="v1" type="Warning" reason="NodeNotReady" message="Node is not ready" I0423 07:44:20.275131 1 event.go:307] "Event occurred" object="kube-system/kube-apiserver-instance" fieldPath="" kind="Pod" apiVersion="v1" type="Warning" reason="NodeNotReady" message="Node is not ready" I0423 07:44:20.316910 1 event.go:307] "Event occurred" object="kube-system/kube-proxy" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: kube-proxy-r7jkh" I0423 07:44:20.467520 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set coredns-7c96b6546b to 2" I0423 07:44:20.615076 1 event.go:307] "Event occurred" object="kube-system/coredns-7c96b6546b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-7c96b6546b-8txxj" I0423 07:44:20.624084 1 event.go:307] "Event occurred" object="kube-system/coredns-7c96b6546b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-7c96b6546b-j8z2r" I0423 07:44:20.632973 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="165.718877ms" I0423 07:44:20.657648 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="24.617505ms" I0423 07:44:20.657796 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="99.243µs" I0423 07:44:24.181216 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set coredns-67659f764b to 1" I0423 07:44:24.191556 1 event.go:307] "Event occurred" object="kube-system/coredns-67659f764b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-67659f764b-6t9nz" I0423 07:44:24.204722 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="24.44694ms" I0423 07:44:24.206366 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled down replica set coredns-7c96b6546b to 1 from 2" I0423 07:44:24.220161 1 event.go:307] "Event occurred" object="kube-system/coredns-7c96b6546b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulDelete" message="Deleted pod: coredns-7c96b6546b-8txxj" I0423 07:44:24.221980 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="17.190522ms" I0423 07:44:24.222171 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="140.023µs" I0423 07:44:24.229246 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="23.361313ms" I0423 07:44:24.236945 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="7.636967ms" I0423 07:44:24.237126 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="130.134µs" I0423 07:44:24.239468 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set coredns-67659f764b to 2 from 1" I0423 07:44:24.248007 1 event.go:307] "Event occurred" object="kube-system/coredns-67659f764b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-67659f764b-ss9qr" I0423 07:44:24.262017 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="23.382935ms" I0423 07:44:24.278608 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="16.525556ms" I0423 07:44:24.278751 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="94.603µs" I0423 07:44:24.278856 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="70.322µs" I0423 07:44:35.519370 1 event.go:307] "Event occurred" object="kube-system/cilium-operator" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set cilium-operator-869df985b8 to 1" I0423 07:44:35.525032 1 event.go:307] "Event occurred" object="kube-system/cilium" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: cilium-xt9lw" I0423 07:44:35.528627 1 event.go:307] "Event occurred" object="kube-system/cilium-operator-869df985b8" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: cilium-operator-869df985b8-mnqms" I0423 07:44:35.539018 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="20.097644ms" I0423 07:44:35.548616 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="9.535134ms" I0423 07:44:35.549992 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="120.903µs" I0423 07:44:35.577354 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="52.491µs" I0423 07:44:46.962512 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:44:48.933392 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:44:48.934009 1 event.go:307] "Event occurred" object="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" fieldPath="" kind="Job" apiVersion="batch/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: envoy-gateway-gateway-helm-certgen-99dxr" I0423 07:44:49.010872 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:44:49.011553 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:44:49.718326 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="backendtlspolicies.gateway.networking.k8s.io" I0423 07:44:49.718401 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="clienttrafficpolicies.gateway.envoyproxy.io" I0423 07:44:49.718442 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="referencegrants.gateway.networking.k8s.io" I0423 07:44:49.718494 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="udproutes.gateway.networking.k8s.io" I0423 07:44:49.718532 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="xlistenersets.gateway.networking.x-k8s.io" I0423 07:44:49.718565 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="backendtrafficpolicies.gateway.envoyproxy.io" I0423 07:44:49.718604 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="backends.gateway.envoyproxy.io" I0423 07:44:49.718646 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="httproutes.gateway.networking.k8s.io" I0423 07:44:49.718702 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="tcproutes.gateway.networking.k8s.io" I0423 07:44:49.718768 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="httproutefilters.gateway.envoyproxy.io" I0423 07:44:49.718810 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="securitypolicies.gateway.envoyproxy.io" I0423 07:44:49.718857 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="gateways.gateway.networking.k8s.io" I0423 07:44:49.718889 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="tlsroutes.gateway.networking.k8s.io" I0423 07:44:49.718921 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="xbackendtrafficpolicies.gateway.networking.x-k8s.io" I0423 07:44:49.718955 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="envoyextensionpolicies.gateway.envoyproxy.io" I0423 07:44:49.719002 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="envoyproxies.gateway.envoyproxy.io" I0423 07:44:49.719340 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="envoypatchpolicies.gateway.envoyproxy.io" I0423 07:44:49.719377 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="grpcroutes.gateway.networking.k8s.io" I0423 07:44:49.720888 1 shared_informer.go:311] Waiting for caches to sync for resource quota I0423 07:44:50.045739 1 shared_informer.go:311] Waiting for caches to sync for garbage collector I0423 07:44:50.921913 1 shared_informer.go:318] Caches are synced for resource quota I0423 07:44:50.946061 1 shared_informer.go:318] Caches are synced for garbage collector I0423 07:44:51.876864 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="115.242µs" I0423 07:44:52.924646 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="9.292178ms" I0423 07:44:52.924822 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-869df985b8" duration="116.662µs" I0423 07:44:55.681449 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:44:55.685421 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="82.552µs" I0423 07:44:55.693222 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:44:55.705290 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="32.751µs" I0423 07:44:59.488300 1 node_lifecycle_controller.go:1048] "Controller detected that some Nodes are Ready. Exiting master disruption mode" I0423 07:45:01.939776 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="107.943µs" I0423 07:45:01.973412 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="11.939904ms" I0423 07:45:01.973503 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="61.071µs" I0423 07:45:05.676784 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:45:05.960560 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:45:07.970649 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:45:08.983431 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:45:08.998616 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:45:09.007380 1 event.go:307] "Event occurred" object="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" fieldPath="" kind="Job" apiVersion="batch/v1" type="Normal" reason="Completed" message="Job completed" I0423 07:45:09.007688 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:45:09.073586 1 event.go:307] "Event occurred" object="envoy-gateway-system/envoy-gateway" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set envoy-gateway-78446f96c9 to 1" I0423 07:45:09.083846 1 event.go:307] "Event occurred" object="envoy-gateway-system/envoy-gateway-78446f96c9" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: envoy-gateway-78446f96c9-5rwm4" I0423 07:45:09.096390 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="22.370801ms" I0423 07:45:09.105841 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="9.378051ms" I0423 07:45:09.106013 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="91.992µs" I0423 07:45:09.108448 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="64.511µs" I0423 07:45:09.979316 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="120.083µs" I0423 07:45:12.250440 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="78.032µs" I0423 07:45:12.256380 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="104.052µs" I0423 07:45:12.272884 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="94.652µs" I0423 07:45:12.287038 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="276.287µs" I0423 07:45:12.997163 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="64.851µs" I0423 07:45:13.999833 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="69.081µs" I0423 07:45:14.009188 1 endpointslice_controller.go:310] "Error syncing endpoint slices for service, retrying" key="kube-system/kube-dns" err="EndpointSlice informer cache is out of date" I0423 07:45:14.025475 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="11.276158ms" I0423 07:45:14.025641 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="126.083µs" I0423 07:45:14.034851 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled down replica set coredns-7c96b6546b to 0 from 1" I0423 07:45:14.046508 1 event.go:307] "Event occurred" object="kube-system/coredns-7c96b6546b" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulDelete" message="Deleted pod: coredns-7c96b6546b-j8z2r" I0423 07:45:14.057402 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="22.633866ms" I0423 07:45:14.069463 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="12.006526ms" I0423 07:45:14.069570 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="74.061µs" I0423 07:45:14.069620 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="27.621µs" I0423 07:45:19.321116 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="63.011µs" I0423 07:45:19.424627 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="13.244786ms" I0423 07:45:19.424760 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="envoy-gateway-system/envoy-gateway-78446f96c9" duration="76.522µs" I0423 07:45:20.022196 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="50.461µs" I0423 07:45:20.023917 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-7c96b6546b" duration="96.493µs" I0423 07:45:20.929548 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ciliumnodeconfigs.cilium.io" I0423 07:45:20.929629 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ciliumnetworkpolicies.cilium.io" I0423 07:45:20.929658 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ciliumendpoints.cilium.io" I0423 07:45:20.929735 1 shared_informer.go:311] Waiting for caches to sync for resource quota I0423 07:45:20.958551 1 shared_informer.go:311] Waiting for caches to sync for garbage collector I0423 07:45:21.030524 1 shared_informer.go:318] Caches are synced for resource quota I0423 07:45:21.059395 1 shared_informer.go:318] Caches are synced for garbage collector I0423 07:45:22.882743 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="8.187471ms" I0423 07:45:22.882889 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-67659f764b" duration="112.312µs" I0423 07:45:38.026171 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:45:38.061771 1 job_controller.go:562] "enqueueing job" key="envoy-gateway-system/envoy-gateway-gateway-helm-certgen" I0423 07:46:19.071806 1 event.go:307] "Event occurred" object="kube-system/ceph-csi-rbd-nodeplugin" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: ceph-csi-rbd-nodeplugin-fplmk" I0423 07:46:19.073132 1 event.go:307] "Event occurred" object="kube-system/ceph-csi-rbd-provisioner" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set ceph-csi-rbd-provisioner-69cb97fb6 to 1" I0423 07:46:19.082906 1 event.go:307] "Event occurred" object="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: ceph-csi-rbd-provisioner-69cb97fb6-ch972" I0423 07:46:19.097342 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="25.288477ms" I0423 07:46:19.130098 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="32.70505ms" I0423 07:46:19.130192 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="61.222µs" I0423 07:46:19.130240 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="31.371µs" I0423 07:46:49.346753 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="10.283805ms" I0423 07:46:49.346867 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-69cb97fb6" duration="69.092µs" I0423 07:49:52.697092 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0423 07:49:55.269043 1 reconciler.go:352] "attacherDetacher.AttachVolume started" volumeName="kubernetes.io/csi/rbd.csi.ceph.com^0001-0024-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-0000000000000002-2996a54c-a05a-4167-8408-6e5db517c154" nodeName="instance" scheduledPods=["default/test-pvc-pod"] I0423 07:49:55.795586 1 operation_generator.go:400] AttachVolume.Attach succeeded for volume "pvc-7b183739-abdd-4e82-960a-be42b713dafa" (UniqueName: "kubernetes.io/csi/rbd.csi.ceph.com^0001-0024-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-0000000000000002-2996a54c-a05a-4167-8408-6e5db517c154") from node "instance" I0423 07:49:55.795904 1 event.go:307] "Event occurred" object="default/test-pvc-pod" fieldPath="" kind="Pod" apiVersion="v1" type="Normal" reason="SuccessfulAttachVolume" message="AttachVolume.Attach succeeded for volume \"pvc-7b183739-abdd-4e82-960a-be42b713dafa\" " I0423 07:50:08.280126 1 reconciler.go:267] "attacherDetacher.DetachVolume started" node="instance" volumeName="kubernetes.io/csi/rbd.csi.ceph.com^0001-0024-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-0000000000000002-2996a54c-a05a-4167-8408-6e5db517c154" I0423 07:50:08.283078 1 operation_generator.go:1636] Verified volume is safe to detach for volume "pvc-7b183739-abdd-4e82-960a-be42b713dafa" (UniqueName: "kubernetes.io/csi/rbd.csi.ceph.com^0001-0024-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-0000000000000002-2996a54c-a05a-4167-8408-6e5db517c154") on node "instance" I0423 07:50:08.821600 1 operation_generator.go:520] DetachVolume.Detach succeeded for volume "pvc-7b183739-abdd-4e82-960a-be42b713dafa" (UniqueName: "kubernetes.io/csi/rbd.csi.ceph.com^0001-0024-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-0000000000000002-2996a54c-a05a-4167-8408-6e5db517c154") on node "instance"