Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─617 /lib/systemd/systemd --user │ │ └─618 (sd-pam) │ ├─session-4.scope │ │ ├─1348 sshd: zuul [priv] │ │ └─1361 sshd: zuul@notty │ └─session-1.scope │ ├─ 614 sshd: zuul [priv] │ ├─ 631 sshd: zuul@notty │ ├─ 768 /usr/bin/python3 │ ├─ 1352 ssh: /home/zuul/.ansible/cp/199.19.213.202-22-zuul [mux] │ ├─18383 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-khxuxidfllhgwnlmrbpaurmerbofucxr ; /usr/bin/python3'"'"' && sleep 0' │ ├─18384 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-khxuxidfllhgwnlmrbpaurmerbofucxr ; /usr/bin/python3' && sleep 0 │ ├─18385 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-khxuxidfllhgwnlmrbpaurmerbofucxr ; /usr/bin/python3 │ ├─18386 /bin/sh -c echo BECOME-SUCCESS-khxuxidfllhgwnlmrbpaurmerbofucxr ; /usr/bin/python3 │ ├─18387 /usr/bin/python3 │ ├─18388 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─18390 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 9612 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─11193 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 582e2f60475ce765c571f5fbe74a86dd336d7d475942131e7814a02c309fdfec -address /run/containerd/containerd.sock │ │ ├─11217 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id feec0a331ee6e998cca914417319807f8c6645e6247e9bdca70747ec474ce1b1 -address /run/containerd/containerd.sock │ │ ├─11236 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d046843eade9c067bacebcb73534ab854dd2e3649b7e52cfa21591ab5bc5dfa0 -address /run/containerd/containerd.sock │ │ ├─11248 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4c76d5e800d6058c82cab30036f48b5b95e30bd35599345d8b2d2af59a44c815 -address /run/containerd/containerd.sock │ │ ├─11249 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6592725fdfd4c27a7f6bfeb0ee65641ad32e8dc935802b4e0c9ab85407106737 -address /run/containerd/containerd.sock │ │ ├─11740 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1dd24d5ac143eae75f6eb4649f31ba7d7a1c0a04dd3bc097263b3c9fd87d7e99 -address /run/containerd/containerd.sock │ │ ├─12423 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c33673930a8baa135237082a133774794a02b8e19187ad16f7453aa3b2ac13ac -address /run/containerd/containerd.sock │ │ ├─12437 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 99178d5f904a71fa3c1ff4e133c47d1bbdbc57d3c67d2416b8f979c34cd489c9 -address /run/containerd/containerd.sock │ │ ├─13626 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9c4ccfce4a456750bbb8b0a199b838cadbf8b43cad40043ad74f8c90750a5b8d -address /run/containerd/containerd.sock │ │ ├─13997 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id baf6a12af4f7466d580bfa3a5988b76bea8aaa36b109f5e85cae111a3d968f5b -address /run/containerd/containerd.sock │ │ ├─14036 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4456161b69370eb00a8daf5637cbe57d23d25ae39d8754b24b2ceea703bc7a4f -address /run/containerd/containerd.sock │ │ └─14392 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6316c3c423c195093b89b1991a4f6c737a39d9ccad9c7dad38b3272485c4fac2 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1092 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─439 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ ├─ 328 /lib/systemd/systemd-udevd │ │ ├─18368 /lib/systemd/systemd-udevd │ │ └─18369 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─544 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─559 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─534 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─11625 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.19.213.202 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─294 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─560 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─uuidd.service │ │ └─1281 /usr/sbin/uuidd --socket-activation │ ├─systemd-resolved.service │ │ └─334 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─cloud-init.service │ ├─dbus.service │ │ └─531 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─335 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─540 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─537 /lib/systemd/systemd-logind ├─k8s.io │ ├─6592725fdfd4c27a7f6bfeb0ee65641ad32e8dc935802b4e0c9ab85407106737 │ │ └─11347 /pause │ ├─02f73ef1adfcaa3f4a1f956256eb4fe0900b1de486f0cdba6f4bb7ceecd536a6 │ │ └─11793 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─9c4ccfce4a456750bbb8b0a199b838cadbf8b43cad40043ad74f8c90750a5b8d │ │ └─13651 /pause │ ├─baf6a12af4f7466d580bfa3a5988b76bea8aaa36b109f5e85cae111a3d968f5b │ │ └─14020 /pause │ ├─feec0a331ee6e998cca914417319807f8c6645e6247e9bdca70747ec474ce1b1 │ │ └─11372 /pause │ ├─1dd24d5ac143eae75f6eb4649f31ba7d7a1c0a04dd3bc097263b3c9fd87d7e99 │ │ └─11764 /pause │ ├─c4a29c5f0870cde64f2013f4d477df34bae12859b9b3a62ddaec11076ed83f00 │ │ └─11468 etcd --advertise-client-urls=https://199.19.213.202:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.19.213.202:2380 --initial-cluster=instance=https://199.19.213.202:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.19.213.202:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.19.213.202:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─d046843eade9c067bacebcb73534ab854dd2e3649b7e52cfa21591ab5bc5dfa0 │ │ └─11333 /pause │ ├─53c333305f660f537334c60db8406914720a774703546bb064f2c8927c7afa0e │ │ └─11488 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─4c76d5e800d6058c82cab30036f48b5b95e30bd35599345d8b2d2af59a44c815 │ │ └─11354 /pause │ ├─08ae9845f84e84cde000f3ef8e3169caf46ec3bc1e3ae3dc3895dd569c038b45 │ │ ├─12893 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─13233 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─009841efb8817de312eb763c7137bb5583f0de6097337ed4e7386c75453249b4 │ │ └─11450 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─582e2f60475ce765c571f5fbe74a86dd336d7d475942131e7814a02c309fdfec │ │ └─11308 /pause │ ├─99178d5f904a71fa3c1ff4e133c47d1bbdbc57d3c67d2416b8f979c34cd489c9 │ │ └─12472 /pause │ ├─6316c3c423c195093b89b1991a4f6c737a39d9ccad9c7dad38b3272485c4fac2 │ │ └─14418 /pause │ ├─4456161b69370eb00a8daf5637cbe57d23d25ae39d8754b24b2ceea703bc7a4f │ │ └─14067 /pause │ ├─5593ee9c434ddf4d5b6123aa145779591746fd11792b79e2d7c549c1050db997 │ │ └─14496 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─6ff8b77404fef47a6b1368eddf5a16dd0fa9f85fa4c95d3d0deec54e9f375229 │ │ └─12692 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─c33673930a8baa135237082a133774794a02b8e19187ad16f7453aa3b2ac13ac │ │ └─12493 /pause │ ├─18031d6cdfa590e96f4cdc06f25e8b0b232e6183ef35339c25585f9f0f01710d │ │ └─14086 /coredns -conf /etc/coredns/Corefile │ ├─b20a3ecdf3b0acb72ca653cc4892dd8382fae0821173d9891bbc1ec245c69122 │ │ └─11431 kube-apiserver --advertise-address=199.19.213.202 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─c200e6d8939afbc46c64a315b0c62a3349d8ad95fc1dc5818f988b50d226889b │ │ └─13686 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─3170a45193d010091131d33fa571c61defb9ae6eb1e4efac950baf228e4310bd │ │ └─11574 /kube-vip manager │ └─af0591f7975b771d8c183bac14eaf23dd0935dbbfb3df12b0938b508c3dcc8eb │ └─14127 /coredns -conf /etc/coredns/Corefile ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount