I0203 10:27:48.853311 1 serving.go:348] Generated self-signed cert in-memory I0203 10:27:49.534415 1 controllermanager.go:189] "Starting" version="v1.28.13" I0203 10:27:49.534438 1 controllermanager.go:191] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" I0203 10:27:49.539111 1 secure_serving.go:213] Serving securely on [::]:10257 I0203 10:27:49.539255 1 leaderelection.go:250] attempting to acquire leader lease kube-system/kube-controller-manager... I0203 10:27:49.539500 1 dynamic_cafile_content.go:157] "Starting controller" name="request-header::/etc/kubernetes/pki/front-proxy-ca.crt" I0203 10:27:49.539575 1 tlsconfig.go:240] "Starting DynamicServingCertificateController" I0203 10:27:49.539618 1 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" E0203 10:27:50.730496 1 leaderelection.go:332] error retrieving resource lock kube-system/kube-controller-manager: leases.coordination.k8s.io "kube-controller-manager" is forbidden: User "system:kube-controller-manager" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-system" I0203 10:27:52.951044 1 leaderelection.go:260] successfully acquired lease kube-system/kube-controller-manager I0203 10:27:52.951507 1 event.go:307] "Event occurred" object="kube-system/kube-controller-manager" fieldPath="" kind="Lease" apiVersion="coordination.k8s.io/v1" type="Normal" reason="LeaderElection" message="instance_1e5fecee-80f9-4c29-b3e4-aa8a43e0a56a became leader" I0203 10:27:52.964512 1 shared_informer.go:311] Waiting for caches to sync for tokens I0203 10:27:52.975431 1 controllermanager.go:642] "Started controller" controller="horizontal-pod-autoscaler-controller" I0203 10:27:52.975597 1 horizontal.go:200] "Starting HPA controller" I0203 10:27:52.975804 1 shared_informer.go:311] Waiting for caches to sync for HPA I0203 10:27:52.981022 1 controllermanager.go:642] "Started controller" controller="cronjob-controller" I0203 10:27:52.981142 1 cronjob_controllerv2.go:139] "Starting cronjob controller v2" I0203 10:27:52.981166 1 shared_informer.go:311] Waiting for caches to sync for cronjob I0203 10:27:52.985478 1 controllermanager.go:642] "Started controller" controller="token-cleaner-controller" I0203 10:27:52.985567 1 tokencleaner.go:112] "Starting token cleaner controller" I0203 10:27:52.985710 1 shared_informer.go:311] Waiting for caches to sync for token_cleaner I0203 10:27:52.985785 1 shared_informer.go:318] Caches are synced for token_cleaner I0203 10:27:52.991753 1 controllermanager.go:642] "Started controller" controller="ephemeral-volume-controller" I0203 10:27:52.991932 1 controller.go:169] "Starting ephemeral volume controller" I0203 10:27:52.991942 1 shared_informer.go:311] Waiting for caches to sync for ephemeral E0203 10:27:52.997909 1 core.go:213] "Failed to start cloud node lifecycle controller" err="no cloud provider provided" I0203 10:27:52.998193 1 controllermanager.go:620] "Warning: skipping controller" controller="cloud-node-lifecycle-controller" I0203 10:27:53.003036 1 controllermanager.go:642] "Started controller" controller="replicationcontroller-controller" I0203 10:27:53.003259 1 replica_set.go:214] "Starting controller" name="replicationcontroller" I0203 10:27:53.003463 1 shared_informer.go:311] Waiting for caches to sync for ReplicationController I0203 10:27:53.015666 1 controllermanager.go:642] "Started controller" controller="pod-garbage-collector-controller" I0203 10:27:53.015839 1 gc_controller.go:101] "Starting GC controller" I0203 10:27:53.015957 1 shared_informer.go:311] Waiting for caches to sync for GC I0203 10:27:53.020916 1 controllermanager.go:642] "Started controller" controller="daemonset-controller" I0203 10:27:53.021149 1 daemon_controller.go:291] "Starting daemon sets controller" I0203 10:27:53.021247 1 shared_informer.go:311] Waiting for caches to sync for daemon sets I0203 10:27:53.026519 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-kubelet-serving" I0203 10:27:53.026641 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-kubelet-serving I0203 10:27:53.026708 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0203 10:27:53.027333 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-kubelet-client" I0203 10:27:53.027406 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-kubelet-client I0203 10:27:53.027488 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0203 10:27:53.028037 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-kube-apiserver-client" I0203 10:27:53.028112 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-kube-apiserver-client I0203 10:27:53.028187 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0203 10:27:53.028883 1 controllermanager.go:642] "Started controller" controller="certificatesigningrequest-signing-controller" I0203 10:27:53.028980 1 certificate_controller.go:115] "Starting certificate controller" name="csrsigning-legacy-unknown" I0203 10:27:53.029120 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrsigning-legacy-unknown I0203 10:27:53.029011 1 dynamic_serving_content.go:132] "Starting controller" name="csr-controller::/etc/kubernetes/pki/ca.crt::/etc/kubernetes/pki/ca.key" I0203 10:27:53.055437 1 controllermanager.go:642] "Started controller" controller="certificatesigningrequest-cleaner-controller" I0203 10:27:53.055517 1 cleaner.go:83] "Starting CSR cleaner controller" I0203 10:27:53.065252 1 shared_informer.go:318] Caches are synced for tokens I0203 10:27:53.208454 1 controllermanager.go:642] "Started controller" controller="serviceaccount-controller" I0203 10:27:53.208543 1 serviceaccounts_controller.go:111] "Starting service account controller" I0203 10:27:53.208554 1 shared_informer.go:311] Waiting for caches to sync for service account I0203 10:27:53.357875 1 controllermanager.go:642] "Started controller" controller="persistentvolume-attach-detach-controller" I0203 10:27:53.358212 1 attach_detach_controller.go:337] "Starting attach detach controller" I0203 10:27:53.358272 1 shared_informer.go:311] Waiting for caches to sync for attach detach I0203 10:27:53.507361 1 controllermanager.go:642] "Started controller" controller="persistentvolume-expander-controller" I0203 10:27:53.507601 1 expand_controller.go:328] "Starting expand controller" I0203 10:27:53.507626 1 shared_informer.go:311] Waiting for caches to sync for expand I0203 10:27:53.659406 1 controllermanager.go:642] "Started controller" controller="persistentvolumeclaim-protection-controller" I0203 10:27:53.659448 1 core.go:224] "Will not configure cloud provider routes for allocate-node-cidrs" CIDRs=false routes=true I0203 10:27:53.659457 1 controllermanager.go:620] "Warning: skipping controller" controller="node-route-controller" I0203 10:27:53.659826 1 pvc_protection_controller.go:102] "Starting PVC protection controller" I0203 10:27:53.659843 1 shared_informer.go:311] Waiting for caches to sync for PVC protection I0203 10:27:53.811336 1 controllermanager.go:642] "Started controller" controller="persistentvolume-protection-controller" I0203 10:27:53.811419 1 pv_protection_controller.go:78] "Starting PV protection controller" I0203 10:27:53.811893 1 shared_informer.go:311] Waiting for caches to sync for PV protection I0203 10:27:53.956589 1 controllermanager.go:642] "Started controller" controller="replicaset-controller" I0203 10:27:53.957071 1 replica_set.go:214] "Starting controller" name="replicaset" I0203 10:27:53.957106 1 shared_informer.go:311] Waiting for caches to sync for ReplicaSet I0203 10:27:54.155780 1 controllermanager.go:642] "Started controller" controller="disruption-controller" I0203 10:27:54.156127 1 disruption.go:433] "Sending events to api server." I0203 10:27:54.156342 1 disruption.go:444] "Starting disruption controller" I0203 10:27:54.156364 1 shared_informer.go:311] Waiting for caches to sync for disruption I0203 10:27:54.318338 1 controllermanager.go:642] "Started controller" controller="ttl-controller" I0203 10:27:54.318593 1 ttl_controller.go:124] "Starting TTL controller" I0203 10:27:54.318610 1 shared_informer.go:311] Waiting for caches to sync for TTL I0203 10:27:54.496910 1 controllermanager.go:642] "Started controller" controller="bootstrap-signer-controller" I0203 10:27:54.496981 1 shared_informer.go:311] Waiting for caches to sync for bootstrap_signer I0203 10:27:54.505780 1 node_lifecycle_controller.go:431] "Controller will reconcile labels" I0203 10:27:54.505826 1 controllermanager.go:642] "Started controller" controller="node-lifecycle-controller" I0203 10:27:54.506100 1 controllermanager.go:620] "Warning: skipping controller" controller="node-ipam-controller" I0203 10:27:54.506254 1 node_lifecycle_controller.go:465] "Sending events to api server" I0203 10:27:54.506395 1 node_lifecycle_controller.go:476] "Starting node controller" I0203 10:27:54.506408 1 shared_informer.go:311] Waiting for caches to sync for taint E0203 10:27:54.656387 1 core.go:92] "Failed to start service controller" err="WARNING: no cloud provider provided, services of type LoadBalancer will fail" I0203 10:27:54.656908 1 controllermanager.go:620] "Warning: skipping controller" controller="service-lb-controller" I0203 10:27:54.805953 1 controllermanager.go:642] "Started controller" controller="persistentvolume-binder-controller" I0203 10:27:54.806195 1 pv_controller_base.go:319] "Starting persistent volume controller" I0203 10:27:54.806230 1 shared_informer.go:311] Waiting for caches to sync for persistent volume I0203 10:27:55.096203 1 garbagecollector.go:155] "Starting controller" controller="garbagecollector" I0203 10:27:55.096239 1 shared_informer.go:311] Waiting for caches to sync for garbage collector I0203 10:27:55.096255 1 graph_builder.go:294] "Running" component="GraphBuilder" I0203 10:27:55.096317 1 controllermanager.go:642] "Started controller" controller="garbage-collector-controller" I0203 10:27:55.221632 1 controllermanager.go:642] "Started controller" controller="certificatesigningrequest-approving-controller" I0203 10:27:55.221698 1 certificate_controller.go:115] "Starting certificate controller" name="csrapproving" I0203 10:27:55.221716 1 shared_informer.go:311] Waiting for caches to sync for certificate-csrapproving I0203 10:27:55.522114 1 controllermanager.go:642] "Started controller" controller="clusterrole-aggregation-controller" I0203 10:27:55.522274 1 clusterroleaggregation_controller.go:189] "Starting ClusterRoleAggregator controller" I0203 10:27:55.522286 1 shared_informer.go:311] Waiting for caches to sync for ClusterRoleAggregator I0203 10:27:55.539433 1 controllermanager.go:642] "Started controller" controller="endpoints-controller" I0203 10:27:55.539556 1 endpoints_controller.go:177] "Starting endpoint controller" I0203 10:27:55.539569 1 shared_informer.go:311] Waiting for caches to sync for endpoint I0203 10:27:55.669143 1 controllermanager.go:642] "Started controller" controller="endpointslice-mirroring-controller" I0203 10:27:55.669220 1 endpointslicemirroring_controller.go:223] "Starting EndpointSliceMirroring controller" I0203 10:27:55.669226 1 shared_informer.go:311] Waiting for caches to sync for endpoint_slice_mirroring I0203 10:27:55.914956 1 controllermanager.go:642] "Started controller" controller="namespace-controller" I0203 10:27:55.915029 1 namespace_controller.go:197] "Starting namespace controller" I0203 10:27:55.915039 1 shared_informer.go:311] Waiting for caches to sync for namespace I0203 10:27:56.135717 1 controllermanager.go:642] "Started controller" controller="deployment-controller" I0203 10:27:56.136008 1 deployment_controller.go:168] "Starting controller" controller="deployment" I0203 10:27:56.136026 1 shared_informer.go:311] Waiting for caches to sync for deployment I0203 10:27:56.220143 1 controllermanager.go:642] "Started controller" controller="statefulset-controller" I0203 10:27:56.220306 1 stateful_set.go:163] "Starting stateful set controller" I0203 10:27:56.220318 1 shared_informer.go:311] Waiting for caches to sync for stateful set I0203 10:27:56.359813 1 controllermanager.go:642] "Started controller" controller="endpointslice-controller" I0203 10:27:56.359952 1 endpointslice_controller.go:264] "Starting endpoint slice controller" I0203 10:27:56.360364 1 shared_informer.go:311] Waiting for caches to sync for endpoint_slice I0203 10:27:56.663376 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="statefulsets.apps" I0203 10:27:56.663410 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="endpoints" I0203 10:27:56.663426 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="limitranges" I0203 10:27:56.663451 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="jobs.batch" I0203 10:27:56.663475 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="rolebindings.rbac.authorization.k8s.io" W0203 10:27:56.663490 1 shared_informer.go:593] resyncPeriod 12h17m40.72612536s is smaller than resyncCheckPeriod 22h59m8.161718808s and the informer has already started. Changing it to 22h59m8.161718808s I0203 10:27:56.663540 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="roles.rbac.authorization.k8s.io" I0203 10:27:56.663569 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="networkpolicies.networking.k8s.io" I0203 10:27:56.663585 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="replicasets.apps" I0203 10:27:56.663597 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ingresses.networking.k8s.io" I0203 10:27:56.663615 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="csistoragecapacities.storage.k8s.io" W0203 10:27:56.663624 1 shared_informer.go:593] resyncPeriod 20h38m17.920075401s is smaller than resyncCheckPeriod 22h59m8.161718808s and the informer has already started. Changing it to 22h59m8.161718808s I0203 10:27:56.663673 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="serviceaccounts" I0203 10:27:56.663707 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="controllerrevisions.apps" I0203 10:27:56.663751 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="poddisruptionbudgets.policy" I0203 10:27:56.663771 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="cronjobs.batch" I0203 10:27:56.663786 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="leases.coordination.k8s.io" I0203 10:27:56.663801 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="podtemplates" I0203 10:27:56.663823 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="deployments.apps" I0203 10:27:56.663840 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="daemonsets.apps" I0203 10:27:56.663856 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="horizontalpodautoscalers.autoscaling" I0203 10:27:56.663876 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="endpointslices.discovery.k8s.io" I0203 10:27:56.663899 1 controllermanager.go:642] "Started controller" controller="resourcequota-controller" I0203 10:27:56.663941 1 resource_quota_controller.go:294] "Starting resource quota controller" I0203 10:27:56.663984 1 shared_informer.go:311] Waiting for caches to sync for resource quota I0203 10:27:56.664015 1 resource_quota_monitor.go:305] "QuotaMonitor running" I0203 10:27:56.807264 1 controllermanager.go:642] "Started controller" controller="job-controller" I0203 10:27:56.807449 1 job_controller.go:226] "Starting job controller" I0203 10:27:56.807490 1 shared_informer.go:311] Waiting for caches to sync for job I0203 10:27:56.957327 1 controllermanager.go:642] "Started controller" controller="ttl-after-finished-controller" I0203 10:27:56.957722 1 ttlafterfinished_controller.go:109] "Starting TTL after finished controller" I0203 10:27:56.957971 1 shared_informer.go:311] Waiting for caches to sync for TTL after finished I0203 10:27:57.106771 1 controllermanager.go:642] "Started controller" controller="root-ca-certificate-publisher-controller" I0203 10:27:57.108730 1 publisher.go:102] "Starting root CA cert publisher controller" I0203 10:27:57.108740 1 shared_informer.go:311] Waiting for caches to sync for crt configmap I0203 10:27:57.116742 1 shared_informer.go:311] Waiting for caches to sync for resource quota I0203 10:27:57.126347 1 shared_informer.go:318] Caches are synced for ClusterRoleAggregator I0203 10:27:57.128826 1 actual_state_of_world.go:547] "Failed to update statusUpdateNeeded field in actual state of world" err="Failed to set statusUpdateNeeded to needed true, because nodeName=\"instance\" does not exist" I0203 10:27:57.139577 1 shared_informer.go:318] Caches are synced for deployment I0203 10:27:57.140916 1 shared_informer.go:318] Caches are synced for endpoint I0203 10:27:57.140974 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-legacy-unknown I0203 10:27:57.150910 1 shared_informer.go:311] Waiting for caches to sync for garbage collector I0203 10:27:57.157131 1 shared_informer.go:318] Caches are synced for ReplicaSet I0203 10:27:57.165111 1 shared_informer.go:318] Caches are synced for endpoint_slice I0203 10:27:57.165123 1 shared_informer.go:318] Caches are synced for TTL after finished I0203 10:27:57.165131 1 shared_informer.go:318] Caches are synced for attach detach I0203 10:27:57.166753 1 shared_informer.go:318] Caches are synced for PVC protection I0203 10:27:57.169367 1 shared_informer.go:318] Caches are synced for endpoint_slice_mirroring I0203 10:27:57.176067 1 shared_informer.go:318] Caches are synced for HPA I0203 10:27:57.181272 1 shared_informer.go:318] Caches are synced for cronjob I0203 10:27:57.192585 1 shared_informer.go:318] Caches are synced for ephemeral I0203 10:27:57.203948 1 shared_informer.go:318] Caches are synced for ReplicationController I0203 10:27:57.207182 1 shared_informer.go:318] Caches are synced for taint I0203 10:27:57.207459 1 node_lifecycle_controller.go:1225] "Initializing eviction metric for zone" zone="" I0203 10:27:57.207527 1 shared_informer.go:318] Caches are synced for job I0203 10:27:57.207301 1 shared_informer.go:318] Caches are synced for persistent volume I0203 10:27:57.207664 1 taint_manager.go:205] "Starting NoExecuteTaintManager" I0203 10:27:57.207805 1 shared_informer.go:318] Caches are synced for expand I0203 10:27:57.208803 1 taint_manager.go:210] "Sending events to api server" I0203 10:27:57.208960 1 shared_informer.go:318] Caches are synced for service account I0203 10:27:57.207974 1 event.go:307] "Event occurred" object="instance" fieldPath="" kind="Node" apiVersion="v1" type="Normal" reason="RegisteredNode" message="Node instance event: Registered Node instance in Controller" I0203 10:27:57.211121 1 node_lifecycle_controller.go:877] "Missing timestamp for Node. Assuming now as a timestamp" node="instance" I0203 10:27:57.211469 1 node_lifecycle_controller.go:1029] "Controller detected that all Nodes are not-Ready. Entering master disruption mode" I0203 10:27:57.212912 1 shared_informer.go:318] Caches are synced for PV protection I0203 10:27:57.215095 1 shared_informer.go:318] Caches are synced for namespace I0203 10:27:57.216514 1 shared_informer.go:318] Caches are synced for GC I0203 10:27:57.218715 1 shared_informer.go:318] Caches are synced for TTL I0203 10:27:57.220705 1 shared_informer.go:318] Caches are synced for stateful set I0203 10:27:57.221907 1 shared_informer.go:318] Caches are synced for daemon sets I0203 10:27:57.226023 1 shared_informer.go:318] Caches are synced for certificate-csrapproving I0203 10:27:57.227246 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-kubelet-serving I0203 10:27:57.227593 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-kubelet-client I0203 10:27:57.228882 1 shared_informer.go:318] Caches are synced for certificate-csrsigning-kube-apiserver-client I0203 10:27:57.257195 1 shared_informer.go:318] Caches are synced for disruption I0203 10:27:57.298189 1 shared_informer.go:318] Caches are synced for bootstrap_signer I0203 10:27:57.309315 1 shared_informer.go:318] Caches are synced for crt configmap I0203 10:27:57.317208 1 shared_informer.go:318] Caches are synced for resource quota I0203 10:27:57.364845 1 shared_informer.go:318] Caches are synced for resource quota I0203 10:27:57.415780 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set coredns-5dd5756b68 to 2" I0203 10:27:57.765829 1 shared_informer.go:318] Caches are synced for garbage collector I0203 10:27:57.797192 1 shared_informer.go:318] Caches are synced for garbage collector I0203 10:27:57.797243 1 garbagecollector.go:166] "All resource monitors have synced. Proceeding to collect garbage" I0203 10:27:58.210645 1 event.go:307] "Event occurred" object="kube-system/kube-proxy" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: kube-proxy-fm2sk" I0203 10:27:58.378166 1 event.go:307] "Event occurred" object="kube-system/coredns-5dd5756b68" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-5dd5756b68-r8ldp" I0203 10:27:58.398280 1 event.go:307] "Event occurred" object="kube-system/coredns-5dd5756b68" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-5dd5756b68-s9dlb" I0203 10:27:58.420574 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="1.004341841s" I0203 10:27:58.432013 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="11.399754ms" I0203 10:27:58.432067 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="26.821µs" I0203 10:28:08.198208 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set coredns-77cccfdc44 to 1" I0203 10:28:08.211161 1 event.go:307] "Event occurred" object="kube-system/coredns-77cccfdc44" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-77cccfdc44-ldbxm" I0203 10:28:08.268382 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="75.388155ms" I0203 10:28:08.275807 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="7.168331ms" I0203 10:28:08.276214 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="91.182µs" I0203 10:28:08.284663 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled down replica set coredns-5dd5756b68 to 1 from 2" I0203 10:28:08.301192 1 event.go:307] "Event occurred" object="kube-system/coredns-5dd5756b68" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulDelete" message="Deleted pod: coredns-5dd5756b68-r8ldp" I0203 10:28:08.319276 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="38.304924ms" I0203 10:28:08.323860 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set coredns-77cccfdc44 to 2 from 1" I0203 10:28:08.351114 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="31.768129ms" I0203 10:28:08.351452 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="267.597µs" I0203 10:28:08.365840 1 event.go:307] "Event occurred" object="kube-system/coredns-77cccfdc44" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: coredns-77cccfdc44-qk7sv" I0203 10:28:08.388604 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="65.330926ms" I0203 10:28:08.407286 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="18.260268ms" I0203 10:28:08.407542 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="144.504µs" I0203 10:28:20.769103 1 event.go:307] "Event occurred" object="kube-system/cilium-operator" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set cilium-operator-7fcd8d6ffd to 1" I0203 10:28:20.778502 1 event.go:307] "Event occurred" object="kube-system/cilium" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: cilium-gnfs4" I0203 10:28:20.781380 1 event.go:307] "Event occurred" object="kube-system/cilium-operator-7fcd8d6ffd" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: cilium-operator-7fcd8d6ffd-vpxpf" I0203 10:28:20.809875 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-7fcd8d6ffd" duration="41.075858ms" I0203 10:28:20.851991 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-7fcd8d6ffd" duration="42.075034ms" I0203 10:28:20.852271 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-7fcd8d6ffd" duration="40.871µs" I0203 10:28:41.226966 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-7fcd8d6ffd" duration="87.092µs" I0203 10:28:41.251248 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-7fcd8d6ffd" duration="8.961139ms" I0203 10:28:41.251432 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/cilium-operator-7fcd8d6ffd" duration="74.002µs" I0203 10:28:43.211868 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="68.851µs" I0203 10:28:43.218262 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="50.922µs" I0203 10:28:43.218380 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="17.561µs" I0203 10:28:43.239760 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="112.893µs" I0203 10:28:43.268390 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="74.142µs" I0203 10:28:43.285055 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="100.753µs" I0203 10:28:47.227575 1 node_lifecycle_controller.go:1048] "Controller detected that some Nodes are Ready. Exiting master disruption mode" I0203 10:28:51.295481 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="61.531µs" I0203 10:28:51.359558 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="23.755571ms" I0203 10:28:51.359666 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="80.192µs" I0203 10:28:51.386760 1 event.go:307] "Event occurred" object="kube-system/coredns" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled down replica set coredns-5dd5756b68 to 0 from 1" I0203 10:28:51.443652 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="53.502283ms" I0203 10:28:51.445941 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-77cccfdc44" duration="44.511µs" I0203 10:28:51.456168 1 event.go:307] "Event occurred" object="kube-system/coredns-5dd5756b68" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulDelete" message="Deleted pod: coredns-5dd5756b68-s9dlb" I0203 10:28:51.473209 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="93.405649ms" I0203 10:28:51.498944 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="25.700332ms" I0203 10:28:51.499731 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="747.65µs" I0203 10:28:57.328501 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ciliumendpoints.cilium.io" I0203 10:28:57.328554 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ciliumnodeconfigs.cilium.io" I0203 10:28:57.328576 1 resource_quota_monitor.go:224] "QuotaMonitor created object count evaluator" resource="ciliumnetworkpolicies.cilium.io" I0203 10:28:57.328640 1 shared_informer.go:311] Waiting for caches to sync for resource quota I0203 10:28:57.429463 1 shared_informer.go:318] Caches are synced for resource quota I0203 10:28:57.782841 1 shared_informer.go:311] Waiting for caches to sync for garbage collector I0203 10:28:57.883729 1 shared_informer.go:318] Caches are synced for garbage collector I0203 10:28:59.942474 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="53.971µs" I0203 10:29:00.330981 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="57.742µs" I0203 10:29:00.333368 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/coredns-5dd5756b68" duration="34.501µs" I0203 10:29:26.983726 1 event.go:307] "Event occurred" object="kube-system/ceph-csi-rbd-provisioner" fieldPath="" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set ceph-csi-rbd-provisioner-795cfbf7fb to 1" I0203 10:29:27.004176 1 event.go:307] "Event occurred" object="kube-system/ceph-csi-rbd-nodeplugin" fieldPath="" kind="DaemonSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: ceph-csi-rbd-nodeplugin-nt5xd" I0203 10:29:27.004783 1 event.go:307] "Event occurred" object="kube-system/ceph-csi-rbd-provisioner-795cfbf7fb" fieldPath="" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: ceph-csi-rbd-provisioner-795cfbf7fb-vznvp" I0203 10:29:27.048591 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-795cfbf7fb" duration="64.335704ms" I0203 10:29:27.108606 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-795cfbf7fb" duration="59.967087ms" I0203 10:29:27.108779 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-795cfbf7fb" duration="147.244µs" I0203 10:30:06.629150 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-795cfbf7fb" duration="10.740978ms" I0203 10:30:06.629217 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/ceph-csi-rbd-provisioner-795cfbf7fb" duration="39.761µs" I0203 10:32:58.490900 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rbd.csi.ceph.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered." I0203 10:33:03.474184 1 reconciler.go:352] "attacherDetacher.AttachVolume started" volumeName="kubernetes.io/csi/rbd.csi.ceph.com^0001-0024-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-0000000000000002-af6f3a9c-52a1-4141-8f1a-7110d43d84ca" nodeName="instance" scheduledPods=["default/test-pvc-pod"] I0203 10:33:04.010728 1 operation_generator.go:400] AttachVolume.Attach succeeded for volume "pvc-832b7551-e41d-4901-9361-b994d8f3bc6c" (UniqueName: "kubernetes.io/csi/rbd.csi.ceph.com^0001-0024-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-0000000000000002-af6f3a9c-52a1-4141-8f1a-7110d43d84ca") from node "instance" I0203 10:33:04.011335 1 event.go:307] "Event occurred" object="default/test-pvc-pod" fieldPath="" kind="Pod" apiVersion="v1" type="Normal" reason="SuccessfulAttachVolume" message="AttachVolume.Attach succeeded for volume \"pvc-832b7551-e41d-4901-9361-b994d8f3bc6c\" " I0203 10:33:22.828914 1 reconciler.go:267] "attacherDetacher.DetachVolume started" node="instance" volumeName="kubernetes.io/csi/rbd.csi.ceph.com^0001-0024-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-0000000000000002-af6f3a9c-52a1-4141-8f1a-7110d43d84ca" I0203 10:33:22.831857 1 operation_generator.go:1636] Verified volume is safe to detach for volume "pvc-832b7551-e41d-4901-9361-b994d8f3bc6c" (UniqueName: "kubernetes.io/csi/rbd.csi.ceph.com^0001-0024-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-0000000000000002-af6f3a9c-52a1-4141-8f1a-7110d43d84ca") on node "instance" I0203 10:33:23.355556 1 operation_generator.go:520] DetachVolume.Detach succeeded for volume "pvc-832b7551-e41d-4901-9361-b994d8f3bc6c" (UniqueName: "kubernetes.io/csi/rbd.csi.ceph.com^0001-0024-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-0000000000000002-af6f3a9c-52a1-4141-8f1a-7110d43d84ca") on node "instance"