Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─session-9.scope │ │ ├─19515 sshd: zuul [priv] │ │ └─19528 sshd: zuul@notty │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─617 /lib/systemd/systemd --user │ │ └─618 (sd-pam) │ ├─session-1.scope │ │ ├─ 614 sshd: zuul [priv] │ │ ├─ 631 sshd: zuul@notty │ │ ├─ 768 /usr/bin/python3 │ │ ├─ 9505 ssh: /home/zuul/.ansible/cp/3a24a1da63 [mux] │ │ ├─19519 ssh: /home/zuul/.ansible/cp/199.204.45.141-22-zuul [mux] │ │ ├─20119 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-igoxdirtaolqnkybmdhgvfklrrklnsls ; /usr/bin/python3'"'"' && sleep 0' │ │ ├─20120 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-igoxdirtaolqnkybmdhgvfklrrklnsls ; /usr/bin/python3' && sleep 0 │ │ ├─20121 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-igoxdirtaolqnkybmdhgvfklrrklnsls ; /usr/bin/python3 │ │ ├─20122 /bin/sh -c echo BECOME-SUCCESS-igoxdirtaolqnkybmdhgvfklrrklnsls ; /usr/bin/python3 │ │ ├─20123 /usr/bin/python3 │ │ ├─20124 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ │ └─20126 systemd-cgls --full --all --no-pager │ └─session-5.scope │ ├─9501 sshd: zuul [priv] │ └─9514 sshd: zuul@notty ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 40 nofb ├─system.slice │ ├─containerd.service … │ │ ├─12050 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─15305 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ffc7152113856e1b2c5ada71d5f64150ebcbb6f66c0dffee34fe1166687aa671 -address /run/containerd/containerd.sock │ │ ├─15339 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 74adb06f698cf1d0235671d420fe21040b114a25e949176f989d2f2be1bca12a -address /run/containerd/containerd.sock │ │ ├─15343 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 653ca56c5ab39d4fe3f5078bfe08206f40b65aa6dc2913e66de4aa09343e3d59 -address /run/containerd/containerd.sock │ │ ├─15361 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 837a0dda911c31aa23c0a0d1428dc924a70ad51b0480f7b589dfa6625e8ae3ab -address /run/containerd/containerd.sock │ │ ├─15366 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d2d71e99fdc98530707cf7730574d41f72b8bbb8cccb3bc4dcce9c3c3959f421 -address /run/containerd/containerd.sock │ │ ├─16060 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3b25f53093c8b7e33dc5e3c29715077641c7ae8df7c6f4e85eaba421b690f00e -address /run/containerd/containerd.sock │ │ ├─17114 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f0a6a71326e0184be368fa59c27e5ae2fce770a58852cbddc1b6eccae28d0259 -address /run/containerd/containerd.sock │ │ ├─17132 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 20d86afef5339f4820d24baa8157776fd51c889158c1e019cf39d65b0abf2bba -address /run/containerd/containerd.sock │ │ ├─18329 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9f4e5427ffa8e7d492fe27fab60be8b6ba7bd04fed66d9479b7a92bdbc28634b -address /run/containerd/containerd.sock │ │ ├─18744 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5e45076c39cfc24a8dd718d73acedcbc38f69e765cdc796654ba071840b11020 -address /run/containerd/containerd.sock │ │ ├─18783 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 952e1b4660f873338782acc4f71bfbd90073c7924e628de50f3ba1ab0295226b -address /run/containerd/containerd.sock │ │ └─19230 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8fe9705da0b2c9b982891e99cccb78f9d55e60e6d07b9b7aae5b307381e24e71 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1092 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─437 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ ├─ 328 /lib/systemd/systemd-udevd │ │ ├─20108 /lib/systemd/systemd-udevd │ │ └─20109 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─543 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─560 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─533 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─15745 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.141 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─cloud-init-local.service │ ├─systemd-journald.service │ │ └─297 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─561 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1446 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─331 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─cloud-init.service │ ├─dbus.service │ │ └─530 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─332 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─542 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─536 /lib/systemd/systemd-logind ├─k8s.io │ ├─3b25f53093c8b7e33dc5e3c29715077641c7ae8df7c6f4e85eaba421b690f00e │ │ └─16085 /pause │ ├─f41f55878e334f65b09fe1777d7b428d0375b62a539afd832511d6e8a51c198b │ │ └─18869 /coredns -conf /etc/coredns/Corefile │ ├─232da2b98c677791783c56c4b84538f1fc4fbd1066983f9b258d0eefe81cd275 │ │ └─16110 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─ffc7152113856e1b2c5ada71d5f64150ebcbb6f66c0dffee34fe1166687aa671 │ │ └─15418 /pause │ ├─653ca56c5ab39d4fe3f5078bfe08206f40b65aa6dc2913e66de4aa09343e3d59 │ │ └─15433 /pause │ ├─74adb06f698cf1d0235671d420fe21040b114a25e949176f989d2f2be1bca12a │ │ └─15467 /pause │ ├─065d36c3cf1bd21b252a5da26b1e8792b113aa9a27a2fa0ae06c5316eb951422 │ │ └─19377 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─5e45076c39cfc24a8dd718d73acedcbc38f69e765cdc796654ba071840b11020 │ │ └─18773 /pause │ ├─226ef1c3411d704512badf2bad991da17f0296e7031d96b7d34797904b117f3f │ │ └─15571 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─d2d71e99fdc98530707cf7730574d41f72b8bbb8cccb3bc4dcce9c3c3959f421 │ │ └─15479 /pause │ ├─8fe9705da0b2c9b982891e99cccb78f9d55e60e6d07b9b7aae5b307381e24e71 │ │ └─19255 /pause │ ├─6dd2dcd4ac880db7931bc6615c376f6d3db78ce0f91c9365e8fc07dd4441e45f │ │ └─15695 /kube-vip manager │ ├─9f4e5427ffa8e7d492fe27fab60be8b6ba7bd04fed66d9479b7a92bdbc28634b │ │ └─18352 /pause │ ├─837a0dda911c31aa23c0a0d1428dc924a70ad51b0480f7b589dfa6625e8ae3ab │ │ └─15478 /pause │ ├─d289be4912b36e87dcd4f0bcc49374edff9fc07b2c08f54684cc3de83493943a │ │ └─15552 etcd --advertise-client-urls=https://199.204.45.141:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.141:2380 --initial-cluster=instance=https://199.204.45.141:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.141:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.141:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─0ef94a3cff438b3d66ab4600bb95296ff69538401a91dee56f69f49d5da98d78 │ │ └─15607 kube-apiserver --advertise-address=199.204.45.141 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─08913d8125518f413c5def6c96f8fd8cb938795c4d6c93ba0296a9917a456e93 │ │ └─18838 /coredns -conf /etc/coredns/Corefile │ ├─5e781595dd106650c233449d438cc2745d74fac7f20b85838e4f94f35d0db8fa │ │ └─17394 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─375c7125b933af316ccc0f91f94f4a9a97c65233fbf4e4fe64d8d88a3dff4c5d │ │ ├─17605 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─17946 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─c29b1b02da61ce6b7bc9458be0d64cc0b0d511abec429c5fad4adec60663537f │ │ └─15588 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─952e1b4660f873338782acc4f71bfbd90073c7924e628de50f3ba1ab0295226b │ │ └─18806 /pause │ ├─20d86afef5339f4820d24baa8157776fd51c889158c1e019cf39d65b0abf2bba │ │ └─17175 /pause │ ├─856a37ca9375d126ef7c50156f6ff4e6d3e7ee3552bb4c46017b52aa20698e13 │ │ └─18397 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ └─f0a6a71326e0184be368fa59c27e5ae2fce770a58852cbddc1b6eccae28d0259 │ └─17157 /pause ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount