I0409 15:57:38.652934 1 options.go:263] external host was not specified, using 199.19.213.86 I0409 15:57:38.658598 1 server.go:150] Version: v1.34.0 I0409 15:57:38.658654 1 server.go:152] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" W0409 15:57:39.031690 1 logging.go:55] [core] [Channel #1 SubChannel #2]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:39.032837 1 logging.go:55] [core] [Channel #4 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" I0409 15:57:39.036727 1 shared_informer.go:349] "Waiting for caches to sync" controller="node_authorizer" I0409 15:57:39.053884 1 shared_informer.go:349] "Waiting for caches to sync" controller="*generic.policySource[*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicy,*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicyBinding,k8s.io/apiserver/pkg/admission/plugin/policy/validating.Validator]" I0409 15:57:39.060593 1 plugins.go:157] Loaded 14 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,RuntimeClass,DefaultIngressClass,PodTopologyLabels,MutatingAdmissionPolicy,MutatingAdmissionWebhook. I0409 15:57:39.060625 1 plugins.go:160] Loaded 13 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,PodSecurity,Priority,PersistentVolumeClaimResize,RuntimeClass,CertificateApproval,CertificateSigning,ClusterTrustBundleAttest,CertificateSubjectRestriction,ValidatingAdmissionPolicy,ValidatingAdmissionWebhook,ResourceQuota. I0409 15:57:39.061012 1 instance.go:239] Using reconciler: lease W0409 15:57:39.062474 1 logging.go:55] [core] [Channel #7 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.091136 1 logging.go:55] [core] [Channel #12 SubChannel #13]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.106996 1 logging.go:55] [core] [Channel #22 SubChannel #23]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" I0409 15:57:46.116857 1 handler.go:285] Adding GroupVersion apiextensions.k8s.io v1 to ResourceManager W0409 15:57:46.116879 1 genericapiserver.go:784] Skipping API apiextensions.k8s.io/v1beta1 because it has no resources. I0409 15:57:46.120700 1 cidrallocator.go:197] starting ServiceCIDR Allocator Controller W0409 15:57:46.121840 1 logging.go:55] [core] [Channel #27 SubChannel #28]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.131360 1 logging.go:55] [core] [Channel #31 SubChannel #32]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.140740 1 logging.go:55] [core] [Channel #35 SubChannel #36]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.152526 1 logging.go:55] [core] [Channel #39 SubChannel #40]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.161532 1 logging.go:55] [core] [Channel #43 SubChannel #44]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.169569 1 logging.go:55] [core] [Channel #47 SubChannel #48]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.179813 1 logging.go:55] [core] [Channel #51 SubChannel #52]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.190640 1 logging.go:55] [core] [Channel #55 SubChannel #56]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.200754 1 logging.go:55] [core] [Channel #59 SubChannel #60]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.210463 1 logging.go:55] [core] [Channel #63 SubChannel #64]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.221642 1 logging.go:55] [core] [Channel #67 SubChannel #68]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.236228 1 logging.go:55] [core] [Channel #71 SubChannel #72]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.246855 1 logging.go:55] [core] [Channel #75 SubChannel #76]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.257668 1 logging.go:55] [core] [Channel #79 SubChannel #80]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.268274 1 logging.go:55] [core] [Channel #83 SubChannel #84]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.279329 1 logging.go:55] [core] [Channel #87 SubChannel #88]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.288666 1 logging.go:55] [core] [Channel #91 SubChannel #92]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" I0409 15:57:46.314310 1 handler.go:285] Adding GroupVersion v1 to ResourceManager I0409 15:57:46.314711 1 apis.go:112] API group "internal.apiserver.k8s.io" is not enabled, skipping. W0409 15:57:46.316289 1 logging.go:55] [core] [Channel #95 SubChannel #96]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.329099 1 logging.go:55] [core] [Channel #99 SubChannel #100]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.340003 1 logging.go:55] [core] [Channel #103 SubChannel #104]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.349895 1 logging.go:55] [core] [Channel #107 SubChannel #108]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.359012 1 logging.go:55] [core] [Channel #111 SubChannel #112]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.368449 1 logging.go:55] [core] [Channel #115 SubChannel #116]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.377600 1 logging.go:55] [core] [Channel #119 SubChannel #120]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.389087 1 logging.go:55] [core] [Channel #123 SubChannel #124]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.401175 1 logging.go:55] [core] [Channel #127 SubChannel #128]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.412392 1 logging.go:55] [core] [Channel #131 SubChannel #132]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.422000 1 logging.go:55] [core] [Channel #135 SubChannel #136]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.432869 1 logging.go:55] [core] [Channel #139 SubChannel #140]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.444262 1 logging.go:55] [core] [Channel #143 SubChannel #144]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.454647 1 logging.go:55] [core] [Channel #147 SubChannel #148]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.466364 1 logging.go:55] [core] [Channel #151 SubChannel #152]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.475761 1 logging.go:55] [core] [Channel #155 SubChannel #156]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.485514 1 logging.go:55] [core] [Channel #159 SubChannel #160]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.526197 1 logging.go:55] [core] [Channel #163 SubChannel #164]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.535828 1 logging.go:55] [core] [Channel #167 SubChannel #168]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.543871 1 logging.go:55] [core] [Channel #171 SubChannel #172]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.553734 1 logging.go:55] [core] [Channel #175 SubChannel #176]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.571257 1 logging.go:55] [core] [Channel #183 SubChannel #184]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.582574 1 logging.go:55] [core] [Channel #187 SubChannel #188]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" I0409 15:57:46.591481 1 apis.go:112] API group "storagemigration.k8s.io" is not enabled, skipping. W0409 15:57:46.592918 1 logging.go:55] [core] [Channel #191 SubChannel #192]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.607594 1 logging.go:55] [core] [Channel #195 SubChannel #196]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.618379 1 logging.go:55] [core] [Channel #199 SubChannel #200]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.626406 1 logging.go:55] [core] [Channel #203 SubChannel #204]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.636587 1 logging.go:55] [core] [Channel #207 SubChannel #208]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.645664 1 logging.go:55] [core] [Channel #211 SubChannel #212]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.655664 1 logging.go:55] [core] [Channel #215 SubChannel #216]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.664407 1 logging.go:55] [core] [Channel #219 SubChannel #220]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.672353 1 logging.go:55] [core] [Channel #223 SubChannel #224]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.683500 1 logging.go:55] [core] [Channel #227 SubChannel #228]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.691893 1 logging.go:55] [core] [Channel #231 SubChannel #232]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.705786 1 logging.go:55] [core] [Channel #235 SubChannel #236]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:57:46.714733 1 logging.go:55] [core] [Channel #239 SubChannel #240]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.752328 1 logging.go:55] [core] [Channel #243 SubChannel #244]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.768879 1 logging.go:55] [core] [Channel #247 SubChannel #248]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:57:46.777823 1 logging.go:55] [core] [Channel #251 SubChannel #252]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" I0409 15:57:46.798476 1 handler.go:285] Adding GroupVersion authentication.k8s.io v1 to ResourceManager W0409 15:57:46.798509 1 genericapiserver.go:784] Skipping API authentication.k8s.io/v1beta1 because it has no resources. W0409 15:57:46.798515 1 genericapiserver.go:784] Skipping API authentication.k8s.io/v1alpha1 because it has no resources. I0409 15:57:46.799045 1 handler.go:285] Adding GroupVersion authorization.k8s.io v1 to ResourceManager W0409 15:57:46.799058 1 genericapiserver.go:784] Skipping API authorization.k8s.io/v1beta1 because it has no resources. I0409 15:57:46.800040 1 handler.go:285] Adding GroupVersion autoscaling v2 to ResourceManager I0409 15:57:46.800807 1 handler.go:285] Adding GroupVersion autoscaling v1 to ResourceManager W0409 15:57:46.800823 1 genericapiserver.go:784] Skipping API autoscaling/v2beta1 because it has no resources. W0409 15:57:46.800828 1 genericapiserver.go:784] Skipping API autoscaling/v2beta2 because it has no resources. I0409 15:57:46.802169 1 handler.go:285] Adding GroupVersion batch v1 to ResourceManager W0409 15:57:46.802183 1 genericapiserver.go:784] Skipping API batch/v1beta1 because it has no resources. I0409 15:57:46.803043 1 handler.go:285] Adding GroupVersion certificates.k8s.io v1 to ResourceManager W0409 15:57:46.803056 1 genericapiserver.go:784] Skipping API certificates.k8s.io/v1beta1 because it has no resources. W0409 15:57:46.803061 1 genericapiserver.go:784] Skipping API certificates.k8s.io/v1alpha1 because it has no resources. I0409 15:57:46.803611 1 handler.go:285] Adding GroupVersion coordination.k8s.io v1 to ResourceManager W0409 15:57:46.803623 1 genericapiserver.go:784] Skipping API coordination.k8s.io/v1beta1 because it has no resources. W0409 15:57:46.803627 1 genericapiserver.go:784] Skipping API coordination.k8s.io/v1alpha2 because it has no resources. I0409 15:57:46.804196 1 handler.go:285] Adding GroupVersion discovery.k8s.io v1 to ResourceManager W0409 15:57:46.804212 1 genericapiserver.go:784] Skipping API discovery.k8s.io/v1beta1 because it has no resources. I0409 15:57:46.806601 1 handler.go:285] Adding GroupVersion networking.k8s.io v1 to ResourceManager W0409 15:57:46.806635 1 genericapiserver.go:784] Skipping API networking.k8s.io/v1beta1 because it has no resources. I0409 15:57:46.807112 1 handler.go:285] Adding GroupVersion node.k8s.io v1 to ResourceManager W0409 15:57:46.807126 1 genericapiserver.go:784] Skipping API node.k8s.io/v1beta1 because it has no resources. W0409 15:57:46.807131 1 genericapiserver.go:784] Skipping API node.k8s.io/v1alpha1 because it has no resources. I0409 15:57:46.807919 1 handler.go:285] Adding GroupVersion policy v1 to ResourceManager W0409 15:57:46.807932 1 genericapiserver.go:784] Skipping API policy/v1beta1 because it has no resources. I0409 15:57:46.809852 1 handler.go:285] Adding GroupVersion rbac.authorization.k8s.io v1 to ResourceManager W0409 15:57:46.809867 1 genericapiserver.go:784] Skipping API rbac.authorization.k8s.io/v1beta1 because it has no resources. W0409 15:57:46.809872 1 genericapiserver.go:784] Skipping API rbac.authorization.k8s.io/v1alpha1 because it has no resources. I0409 15:57:46.810238 1 handler.go:285] Adding GroupVersion scheduling.k8s.io v1 to ResourceManager W0409 15:57:46.810250 1 genericapiserver.go:784] Skipping API scheduling.k8s.io/v1beta1 because it has no resources. W0409 15:57:46.810254 1 genericapiserver.go:784] Skipping API scheduling.k8s.io/v1alpha1 because it has no resources. I0409 15:57:46.812394 1 handler.go:285] Adding GroupVersion storage.k8s.io v1 to ResourceManager W0409 15:57:46.812412 1 genericapiserver.go:784] Skipping API storage.k8s.io/v1beta1 because it has no resources. W0409 15:57:46.812416 1 genericapiserver.go:784] Skipping API storage.k8s.io/v1alpha1 because it has no resources. I0409 15:57:46.813441 1 handler.go:285] Adding GroupVersion flowcontrol.apiserver.k8s.io v1 to ResourceManager W0409 15:57:46.813455 1 genericapiserver.go:784] Skipping API flowcontrol.apiserver.k8s.io/v1beta3 because it has no resources. W0409 15:57:46.813459 1 genericapiserver.go:784] Skipping API flowcontrol.apiserver.k8s.io/v1beta2 because it has no resources. W0409 15:57:46.813462 1 genericapiserver.go:784] Skipping API flowcontrol.apiserver.k8s.io/v1beta1 because it has no resources. I0409 15:57:46.816886 1 handler.go:285] Adding GroupVersion apps v1 to ResourceManager W0409 15:57:46.816904 1 genericapiserver.go:784] Skipping API apps/v1beta2 because it has no resources. W0409 15:57:46.816907 1 genericapiserver.go:784] Skipping API apps/v1beta1 because it has no resources. I0409 15:57:46.818838 1 handler.go:285] Adding GroupVersion admissionregistration.k8s.io v1 to ResourceManager W0409 15:57:46.818851 1 genericapiserver.go:784] Skipping API admissionregistration.k8s.io/v1beta1 because it has no resources. W0409 15:57:46.818855 1 genericapiserver.go:784] Skipping API admissionregistration.k8s.io/v1alpha1 because it has no resources. I0409 15:57:46.819181 1 handler.go:285] Adding GroupVersion events.k8s.io v1 to ResourceManager W0409 15:57:46.819193 1 genericapiserver.go:784] Skipping API events.k8s.io/v1beta1 because it has no resources. W0409 15:57:46.819226 1 genericapiserver.go:784] Skipping API resource.k8s.io/v1beta2 because it has no resources. I0409 15:57:46.820557 1 handler.go:285] Adding GroupVersion resource.k8s.io v1 to ResourceManager W0409 15:57:46.820570 1 genericapiserver.go:784] Skipping API resource.k8s.io/v1beta1 because it has no resources. W0409 15:57:46.820574 1 genericapiserver.go:784] Skipping API resource.k8s.io/v1alpha3 because it has no resources. W0409 15:57:46.823527 1 logging.go:55] [core] [Channel #255 SubChannel #256]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" I0409 15:57:46.836277 1 handler.go:285] Adding GroupVersion apiregistration.k8s.io v1 to ResourceManager W0409 15:57:46.836320 1 genericapiserver.go:784] Skipping API apiregistration.k8s.io/v1beta1 because it has no resources. I0409 15:57:47.258553 1 dynamic_cafile_content.go:161] "Starting controller" name="request-header::/etc/kubernetes/pki/front-proxy-ca.crt" I0409 15:57:47.258584 1 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" I0409 15:57:47.259028 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::/etc/kubernetes/pki/apiserver.crt::/etc/kubernetes/pki/apiserver.key" I0409 15:57:47.259304 1 secure_serving.go:211] Serving securely on [::]:6443 I0409 15:57:47.259390 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" I0409 15:57:47.259484 1 controller.go:78] Starting OpenAPI AggregationController I0409 15:57:47.259560 1 apf_controller.go:377] Starting API Priority and Fairness config controller I0409 15:57:47.259665 1 system_namespaces_controller.go:66] Starting system namespaces controller I0409 15:57:47.259978 1 default_servicecidr_controller.go:111] Starting kubernetes-service-cidr-controller I0409 15:57:47.260027 1 shared_informer.go:349] "Waiting for caches to sync" controller="kubernetes-service-cidr-controller" I0409 15:57:47.260256 1 controller.go:119] Starting legacy_token_tracking_controller I0409 15:57:47.260271 1 cluster_authentication_trust_controller.go:459] Starting cluster_authentication_trust_controller controller I0409 15:57:47.260296 1 shared_informer.go:349] "Waiting for caches to sync" controller="configmaps" I0409 15:57:47.260324 1 shared_informer.go:349] "Waiting for caches to sync" controller="cluster_authentication_trust_controller" I0409 15:57:47.260564 1 customresource_discovery_controller.go:294] Starting DiscoveryController I0409 15:57:47.260600 1 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" I0409 15:57:47.260649 1 controller.go:142] Starting OpenAPI controller I0409 15:57:47.260701 1 controller.go:90] Starting OpenAPI V3 controller I0409 15:57:47.260733 1 naming_controller.go:299] Starting NamingConditionController I0409 15:57:47.260762 1 establishing_controller.go:81] Starting EstablishingController I0409 15:57:47.260802 1 nonstructuralschema_controller.go:195] Starting NonStructuralSchemaConditionController I0409 15:57:47.260912 1 apiapproval_controller.go:189] Starting KubernetesAPIApprovalPolicyConformantConditionController I0409 15:57:47.260941 1 crd_finalizer.go:269] Starting CRDFinalizer I0409 15:57:47.261056 1 dynamic_cafile_content.go:161] "Starting controller" name="request-header::/etc/kubernetes/pki/front-proxy-ca.crt" I0409 15:57:47.262284 1 local_available_controller.go:156] Starting LocalAvailability controller I0409 15:57:47.262302 1 cache.go:32] Waiting for caches to sync for LocalAvailability controller I0409 15:57:47.262333 1 remote_available_controller.go:425] Starting RemoteAvailability controller I0409 15:57:47.262342 1 cache.go:32] Waiting for caches to sync for RemoteAvailability controller I0409 15:57:47.262405 1 aggregator.go:169] waiting for initial CRD sync... I0409 15:57:47.262458 1 dynamic_serving_content.go:135] "Starting controller" name="aggregator-proxy-cert::/etc/kubernetes/pki/front-proxy-client.crt::/etc/kubernetes/pki/front-proxy-client.key" I0409 15:57:47.262555 1 apiservice_controller.go:100] Starting APIServiceRegistrationController I0409 15:57:47.262561 1 cache.go:32] Waiting for caches to sync for APIServiceRegistrationController controller I0409 15:57:47.262578 1 controller.go:80] Starting OpenAPI V3 AggregationController I0409 15:57:47.262637 1 gc_controller.go:78] Starting apiserver lease garbage collector I0409 15:57:47.262704 1 crdregistration_controller.go:114] Starting crd-autoregister controller I0409 15:57:47.262710 1 shared_informer.go:349] "Waiting for caches to sync" controller="crd-autoregister" I0409 15:57:47.262760 1 repairip.go:210] Starting ipallocator-repair-controller I0409 15:57:47.262765 1 shared_informer.go:349] "Waiting for caches to sync" controller="ipallocator-repair-controller" E0409 15:57:47.296365 1 authentication.go:75] "Unable to authenticate the request" err="invalid bearer token" I0409 15:57:47.321099 1 cidrallocator.go:301] created ClusterIP allocator for Service CIDR 10.96.0.0/12 I0409 15:57:47.336974 1 shared_informer.go:356] "Caches are synced" controller="node_authorizer" I0409 15:57:47.354976 1 shared_informer.go:356] "Caches are synced" controller="*generic.policySource[*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicy,*k8s.io/api/admissionregistration/v1.ValidatingAdmissionPolicyBinding,k8s.io/apiserver/pkg/admission/plugin/policy/validating.Validator]" I0409 15:57:47.355004 1 policy_source.go:240] refreshing policies I0409 15:57:47.362742 1 shared_informer.go:356] "Caches are synced" controller="kubernetes-service-cidr-controller" I0409 15:57:47.362801 1 default_servicecidr_controller.go:137] Shutting down kubernetes-service-cidr-controller I0409 15:57:47.362850 1 shared_informer.go:356] "Caches are synced" controller="configmaps" I0409 15:57:47.362902 1 shared_informer.go:356] "Caches are synced" controller="cluster_authentication_trust_controller" I0409 15:57:47.363064 1 apf_controller.go:382] Running API Priority and Fairness config worker I0409 15:57:47.363803 1 apf_controller.go:385] Running API Priority and Fairness periodic rebalancing process I0409 15:57:47.363106 1 cache.go:39] Caches are synced for RemoteAvailability controller I0409 15:57:47.363089 1 shared_informer.go:356] "Caches are synced" controller="ipallocator-repair-controller" I0409 15:57:47.363116 1 shared_informer.go:356] "Caches are synced" controller="crd-autoregister" I0409 15:57:47.363135 1 cache.go:39] Caches are synced for LocalAvailability controller I0409 15:57:47.363971 1 aggregator.go:171] initial CRD sync complete... I0409 15:57:47.363988 1 autoregister_controller.go:144] Starting autoregister controller I0409 15:57:47.364001 1 cache.go:32] Waiting for caches to sync for autoregister controller I0409 15:57:47.364026 1 cache.go:39] Caches are synced for autoregister controller I0409 15:57:47.363149 1 cache.go:39] Caches are synced for APIServiceRegistrationController controller I0409 15:57:47.373158 1 handler_discovery.go:451] Starting ResourceDiscoveryManager I0409 15:57:47.417850 1 controller.go:667] quota admission added evaluator for: leases.coordination.k8s.io I0409 15:57:48.268394 1 storage_scheduling.go:111] all system priority classes are created successfully or already exist. W0409 15:57:48.786410 1 lease.go:265] Resetting endpoints for master service "kubernetes" to [199.19.213.100 199.19.213.86] I0409 15:57:48.789173 1 controller.go:667] quota admission added evaluator for: endpoints I0409 15:57:48.796785 1 controller.go:667] quota admission added evaluator for: endpointslices.discovery.k8s.io I0409 15:58:52.395783 1 controller.go:667] quota admission added evaluator for: serviceaccounts I0409 15:58:56.249164 1 stats.go:136] "Error getting keys" err="empty key: \"\"" I0409 15:59:13.111768 1 handler.go:285] Adding GroupVersion gateway.networking.x-k8s.io v1alpha1 to ResourceManager I0409 15:59:13.154053 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1 to ResourceManager I0409 15:59:13.154106 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1beta1 to ResourceManager I0409 15:59:13.235869 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1alpha2 to ResourceManager I0409 15:59:13.272815 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1alpha2 to ResourceManager I0409 15:59:13.313528 1 handler.go:285] Adding GroupVersion gateway.networking.x-k8s.io v1alpha1 to ResourceManager I0409 15:59:13.335719 1 handler.go:285] Adding GroupVersion gateway.networking.x-k8s.io v1alpha1 to ResourceManager I0409 15:59:13.369531 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1 to ResourceManager I0409 15:59:13.369585 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1alpha3 to ResourceManager I0409 15:59:13.400928 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1alpha2 to ResourceManager I0409 15:59:13.400977 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1alpha3 to ResourceManager I0409 15:59:13.475018 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1 to ResourceManager I0409 15:59:13.475087 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1beta1 to ResourceManager I0409 15:59:13.530244 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1 to ResourceManager I0409 15:59:13.555381 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1beta1 to ResourceManager I0409 15:59:14.094985 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1 to ResourceManager I0409 15:59:14.095085 1 handler.go:285] Adding GroupVersion gateway.networking.k8s.io v1beta1 to ResourceManager I0409 15:59:14.345405 1 handler.go:285] Adding GroupVersion gateway.envoyproxy.io v1alpha1 to ResourceManager I0409 15:59:14.418730 1 handler.go:285] Adding GroupVersion gateway.envoyproxy.io v1alpha1 to ResourceManager I0409 15:59:14.460118 1 handler.go:285] Adding GroupVersion gateway.envoyproxy.io v1alpha1 to ResourceManager I0409 15:59:14.676472 1 handler.go:285] Adding GroupVersion gateway.envoyproxy.io v1alpha1 to ResourceManager I0409 15:59:15.282710 1 stats.go:136] "Error getting keys" err="empty key: \"\"" I0409 15:59:15.530032 1 handler.go:285] Adding GroupVersion gateway.envoyproxy.io v1alpha1 to ResourceManager I0409 15:59:15.658488 1 handler.go:285] Adding GroupVersion gateway.envoyproxy.io v1alpha1 to ResourceManager I0409 15:59:15.706346 1 handler.go:285] Adding GroupVersion gateway.envoyproxy.io v1alpha1 to ResourceManager I0409 15:59:15.799225 1 handler.go:285] Adding GroupVersion gateway.envoyproxy.io v1alpha1 to ResourceManager I0409 15:59:20.039956 1 handler.go:285] Adding GroupVersion cilium.io v2alpha1 to ResourceManager I0409 15:59:20.053251 1 handler.go:285] Adding GroupVersion cilium.io v2alpha1 to ResourceManager I0409 15:59:20.067075 1 handler.go:285] Adding GroupVersion cilium.io v2alpha1 to ResourceManager {"level":"warn","ts":"2026-04-09T15:59:22.199000Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc000071860/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = DeadlineExceeded desc = context deadline exceeded"} {"level":"warn","ts":"2026-04-09T15:59:24.199872Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0003a3860/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = DeadlineExceeded desc = context deadline exceeded"} {"level":"warn","ts":"2026-04-09T15:59:25.284940Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc000071860/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = DeadlineExceeded desc = context deadline exceeded"} {"level":"warn","ts":"2026-04-09T15:59:26.209153Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0003a3860/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = DeadlineExceeded desc = context deadline exceeded"} {"level":"warn","ts":"2026-04-09T15:59:27.621035Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00254f4a0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = DeadlineExceeded desc = context deadline exceeded"} E0409 15:59:27.624913 1 status.go:71] "Unhandled Error" err="apiserver received an error that is not an metav1.Status: context.deadlineExceededError{}: context deadline exceeded" logger="UnhandledError" E0409 15:59:27.625095 1 writers.go:123] "Unhandled Error" err="apiserver was unable to write a JSON response: http: Handler timeout" logger="UnhandledError" E0409 15:59:27.626299 1 status.go:71] "Unhandled Error" err="apiserver received an error that is not an metav1.Status: &errors.errorString{s:\"http: Handler timeout\"}: http: Handler timeout" logger="UnhandledError" E0409 15:59:27.626327 1 writers.go:136] "Unhandled Error" err="apiserver was unable to write a fallback JSON response: http: Handler timeout" logger="UnhandledError" E0409 15:59:27.627655 1 timeout.go:140] "Post-timeout activity" logger="UnhandledError" timeElapsed="5.349538ms" method="GET" path="/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-scheduler" result=null {"level":"warn","ts":"2026-04-09T15:59:28.160769Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00254f4a0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Canceled desc = context canceled"} E0409 15:59:28.160931 1 status.go:71] "Unhandled Error" err="apiserver received an error that is not an metav1.Status: &errors.errorString{s:\"context canceled\"}: context canceled" logger="UnhandledError" E0409 15:59:28.161046 1 writers.go:123] "Unhandled Error" err="apiserver was unable to write a JSON response: http: Handler timeout" logger="UnhandledError" E0409 15:59:28.162191 1 status.go:71] "Unhandled Error" err="apiserver received an error that is not an metav1.Status: &errors.errorString{s:\"http: Handler timeout\"}: http: Handler timeout" logger="UnhandledError" E0409 15:59:28.162240 1 writers.go:136] "Unhandled Error" err="apiserver was unable to write a fallback JSON response: http: Handler timeout" logger="UnhandledError" {"level":"warn","ts":"2026-04-09T15:59:28.162363Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0029f9860/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.162825Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc001a75a40/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.162884Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00254f4a0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.162848Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00254f4a0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.163043Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc001c72d20/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.163222Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0015601e0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.163276Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0015601e0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.163302Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0015601e0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.163401Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0015601e0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.163433Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0015601e0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.163453Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0015601e0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.163475Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0015601e0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:28.163484Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0015601e0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} E0409 15:59:28.163649 1 timeout.go:140] "Post-timeout activity" logger="UnhandledError" timeElapsed="2.497978ms" method="GET" path="/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager" result=null I0409 15:59:28.179903 1 handler.go:285] Adding GroupVersion cilium.io v2 to ResourceManager I0409 15:59:28.206945 1 handler.go:285] Adding GroupVersion cilium.io v2alpha1 to ResourceManager I0409 15:59:28.225196 1 handler.go:285] Adding GroupVersion cilium.io v2 to ResourceManager I0409 15:59:28.238743 1 handler.go:285] Adding GroupVersion cilium.io v2 to ResourceManager I0409 15:59:28.298609 1 handler.go:285] Adding GroupVersion cilium.io v2alpha1 to ResourceManager I0409 15:59:28.319812 1 handler.go:285] Adding GroupVersion cilium.io v2 to ResourceManager W0409 15:59:28.370605 1 logging.go:55] [core] [Channel #265 SubChannel #266]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" I0409 15:59:28.463862 1 handler.go:285] Adding GroupVersion cilium.io v2 to ResourceManager I0409 15:59:28.588287 1 handler.go:285] Adding GroupVersion cilium.io v2 to ResourceManager W0409 15:59:29.305156 1 logging.go:55] [core] [Channel #269 SubChannel #270]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:59:29.318880 1 logging.go:55] [core] [Channel #273 SubChannel #274]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:59:29.419164 1 logging.go:55] [core] [Channel #277 SubChannel #278]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:59:29.439812 1 logging.go:55] [core] [Channel #281 SubChannel #282]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: authentication handshake failed: context canceled" W0409 15:59:29.472041 1 logging.go:55] [core] [Channel #285 SubChannel #286]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" W0409 15:59:29.610988 1 logging.go:55] [core] [Channel #289 SubChannel #290]grpc: addrConn.createTransport failed to connect to {Addr: "127.0.0.1:2379", ServerName: "127.0.0.1:2379", BalancerAttributes: {"<%!p(pickfirstleaf.managedByPickfirstKeyType={})>": "<%!p(bool=true)>" }}. Err: connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:2379: operation was canceled" {"level":"warn","ts":"2026-04-09T15:59:33.058882Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0003a3860/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = DeadlineExceeded desc = context deadline exceeded"} {"level":"warn","ts":"2026-04-09T15:59:33.515546Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00254f4a0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:33.516016Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00254f4a0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:37.543998Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0029f8f00/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:37.544043Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00254f4a0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:43.205242Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00098b2c0/127.0.0.1:2379","method":"/etcdserverpb.KV/Txn","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: request timed out"} E0409 15:59:43.205408 1 status.go:71] "Unhandled Error" err="apiserver received an error that is not an metav1.Status: rpctypes.EtcdError{code:0xe, desc:\"etcdserver: request timed out\"}: etcdserver: request timed out" logger="UnhandledError" {"level":"warn","ts":"2026-04-09T15:59:52.161876Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00254f4a0/127.0.0.1:2379","method":"/etcdserverpb.KV/Range","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: leader changed"} {"level":"warn","ts":"2026-04-09T15:59:58.586368Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00254f4a0/127.0.0.1:2379","method":"/etcdserverpb.KV/Txn","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: request timed out, possibly due to previous leader failure"} E0409 15:59:58.586534 1 status.go:71] "Unhandled Error" err="apiserver received an error that is not an metav1.Status: rpctypes.EtcdError{code:0xe, desc:\"etcdserver: request timed out, possibly due to previous leader failure\"}: etcdserver: request timed out, possibly due to previous leader failure" logger="UnhandledError" {"level":"warn","ts":"2026-04-09T15:59:58.622714Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc001c723c0/127.0.0.1:2379","method":"/etcdserverpb.KV/Txn","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: request timed out, possibly due to previous leader failure"} E0409 15:59:58.622964 1 status.go:71] "Unhandled Error" err="apiserver received an error that is not an metav1.Status: rpctypes.EtcdError{code:0xe, desc:\"etcdserver: request timed out, possibly due to previous leader failure\"}: etcdserver: request timed out, possibly due to previous leader failure" logger="UnhandledError" {"level":"warn","ts":"2026-04-09T15:59:58.649942Z","logger":"etcd-client","caller":"v3@v3.6.4/retry_interceptor.go:65","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc00254f4a0/127.0.0.1:2379","method":"/etcdserverpb.KV/Txn","attempt":0,"error":"rpc error: code = Unavailable desc = etcdserver: request timed out, possibly due to previous leader failure"} E0409 15:59:58.650139 1 status.go:71] "Unhandled Error" err="apiserver received an error that is not an metav1.Status: rpctypes.EtcdError{code:0xe, desc:\"etcdserver: request timed out, possibly due to previous leader failure\"}: etcdserver: request timed out, possibly due to previous leader failure" logger="UnhandledError" E0409 15:59:58.651024 1 controller.go:195] "Failed to update lease" err="etcdserver: request timed out, possibly due to previous leader failure" I0409 15:59:58.982989 1 stats.go:136] "Error getting keys" err="empty key: \"\"" I0409 16:00:20.266963 1 stats.go:136] "Error getting keys" err="empty key: \"\"" I0409 16:01:06.633753 1 stats.go:136] "Error getting keys" err="empty key: \"\"" I0409 16:01:44.145304 1 stats.go:136] "Error getting keys" err="empty key: \"\"" I0409 16:02:10.056792 1 stats.go:136] "Error getting keys" err="empty key: \"\""