Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─620 /lib/systemd/systemd --user │ │ └─621 (sd-pam) │ ├─session-4.scope │ │ ├─1362 sshd: zuul [priv] │ │ └─1375 sshd: zuul@notty │ └─session-1.scope │ ├─ 617 sshd: zuul [priv] │ ├─ 634 sshd: zuul@notty │ ├─ 771 /usr/bin/python3 │ ├─ 1366 ssh: /home/zuul/.ansible/cp/199.204.45.132-22-zuul [mux] │ ├─18542 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-jmdqbtynprskptccricfzuqypcjdiagw ; /usr/bin/python3'"'"' && sleep 0' │ ├─18543 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-jmdqbtynprskptccricfzuqypcjdiagw ; /usr/bin/python3' && sleep 0 │ ├─18544 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-jmdqbtynprskptccricfzuqypcjdiagw ; /usr/bin/python3 │ ├─18545 /bin/sh -c echo BECOME-SUCCESS-jmdqbtynprskptccricfzuqypcjdiagw ; /usr/bin/python3 │ ├─18546 /usr/bin/python3 │ ├─18547 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─18549 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 9627 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─11222 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 89f49054ca221cf460e7f4c036a8d7e4233e3927ad1708878e20715ae1df35a9 -address /run/containerd/containerd.sock │ │ ├─11227 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a94679530187d0262a18be41a49f9a87b09632b677995822c52b286ee4a1c6f8 -address /run/containerd/containerd.sock │ │ ├─11251 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ae1002b933a6479e4bfff55312d41d42f89a518c6db8f5f544968b846fbdb127 -address /run/containerd/containerd.sock │ │ ├─11254 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c8fe28dd2883cfc7bbb593f5ef4f066844d92cd9ffd61b8df59e9bd591ebd43f -address /run/containerd/containerd.sock │ │ ├─11273 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e28abaf55f9ab931be7a069ca6f045a83b7ee30d1b34f398ade6f0fdf8e0437d -address /run/containerd/containerd.sock │ │ ├─11801 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f07bbb62b43a6de521f28cb74e07e5e4a52afec696a456d10c0d4d067e851f22 -address /run/containerd/containerd.sock │ │ ├─12431 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b1a08debe636ec7c7241f76f0c0f69da73f2226969a498fdacf08b4211dab26a -address /run/containerd/containerd.sock │ │ ├─12557 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id bdbe869aa829e408829f599a82cfd136da7b71b2dcb2664d3c260dafe423bf1c -address /run/containerd/containerd.sock │ │ ├─13882 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b69b24d6eaf07dfaed072e7f20e4334f03a3a58031ec177940255db7b14c454c -address /run/containerd/containerd.sock │ │ ├─14196 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 161cf4652f71d8bf43d0c997b24bb2711c8a9aa40ff58cfd8cf4fbecbee3dfa2 -address /run/containerd/containerd.sock │ │ ├─14242 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7933465538d3f98e6892834f7caeafa5f3a4484521c78e5c541b3eda90af4847 -address /run/containerd/containerd.sock │ │ ├─14616 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7bfbcede2d5a075d001f1f81af174a07f704d5537af1fa7fc994ad5c5cf46dfc -address /run/containerd/containerd.sock │ │ └─18361 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 32e3bb5639766ba22b2cdbf5fb8de1f7113557474e3412db01d52373497293e4 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1095 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─437 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─327 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─540 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─558 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─533 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─11642 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.132 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─cloud-init-local.service │ ├─systemd-journald.service │ │ └─297 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─564 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─uuidd.service │ │ └─1297 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─332 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─dbus.service │ │ └─530 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─333 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─539 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─536 /lib/systemd/systemd-logind ├─k8s.io │ ├─a83b3319cd4096442bdbeb2f336ec62264892175ecfc4a89093155781f33e483 │ │ └─14285 /coredns -conf /etc/coredns/Corefile │ ├─136690ae0676d45999594a69fa3c510e7de1404375b14556f7ab0ef32d2f4876 │ │ └─14763 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─7bfbcede2d5a075d001f1f81af174a07f704d5537af1fa7fc994ad5c5cf46dfc │ │ └─14639 /pause │ ├─c8fe28dd2883cfc7bbb593f5ef4f066844d92cd9ffd61b8df59e9bd591ebd43f │ │ └─11385 /pause │ ├─b69b24d6eaf07dfaed072e7f20e4334f03a3a58031ec177940255db7b14c454c │ │ └─13906 /pause │ ├─32e3bb5639766ba22b2cdbf5fb8de1f7113557474e3412db01d52373497293e4 │ │ └─18384 /pause │ ├─ae1002b933a6479e4bfff55312d41d42f89a518c6db8f5f544968b846fbdb127 │ │ └─11355 /pause │ ├─f07bbb62b43a6de521f28cb74e07e5e4a52afec696a456d10c0d4d067e851f22 │ │ └─11827 /pause │ ├─7a5a9f1f45a80f0c171f5657fe9c2568464c65a7226c016aa99cef8eca8db557 │ │ ├─13090 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─13420 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─a94679530187d0262a18be41a49f9a87b09632b677995822c52b286ee4a1c6f8 │ │ └─11340 /pause │ ├─161cf4652f71d8bf43d0c997b24bb2711c8a9aa40ff58cfd8cf4fbecbee3dfa2 │ │ └─14220 /pause │ ├─89f49054ca221cf460e7f4c036a8d7e4233e3927ad1708878e20715ae1df35a9 │ │ └─11332 /pause │ ├─cdb73668fd0b36d5a1868cd0219d55f93e16e5138e0f60cdbd06e7df8d026fc9 │ │ └─11493 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─67afefa1c916ad613cdd831e4b420d356e61b767ce6dd1be49c93bfb318ecab5 │ │ └─11452 etcd --advertise-client-urls=https://199.204.45.132:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.132:2380 --initial-cluster=instance=https://199.204.45.132:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.132:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.132:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─e28abaf55f9ab931be7a069ca6f045a83b7ee30d1b34f398ade6f0fdf8e0437d │ │ └─11374 /pause │ ├─af129b4fd45a6ba9152b7d0187aa81feca2ef3fa3800d030ae4c96c0bde5a4be │ │ └─11594 /kube-vip manager │ ├─1391d93c0668305095efddf7765d1325a8ab7e4cb49ea21279300710dddc7c76 │ │ └─13936 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─a19313722540440550140e9fc198646697ff454327c4fd8b35eb60366f169b1f │ │ └─11854 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─ebc6fc78783d1a2dc5e30426e9c35450b3eb0b4d6d247fc4f705b762dad786e5 │ │ └─14325 /coredns -conf /etc/coredns/Corefile │ ├─66a37df2b50cfeb99c47d2bf75bfa2a6005a77c41b5108cc8bd0bff4c713ce16 │ │ └─11465 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─b1a08debe636ec7c7241f76f0c0f69da73f2226969a498fdacf08b4211dab26a │ │ └─12459 /pause │ ├─2fea80c4e3880d12654b6d8a3400d613e61ce7f210f25cddb5553c92c5e67ce7 │ │ └─11505 kube-apiserver --advertise-address=199.204.45.132 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─ac08cd9663acca7213411fbe0399301db39061a5e59dc650d7e0bb8f0e144232 │ │ └─12588 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─bdbe869aa829e408829f599a82cfd136da7b71b2dcb2664d3c260dafe423bf1c │ │ └─12600 /pause │ └─7933465538d3f98e6892834f7caeafa5f3a4484521c78e5c541b3eda90af4847 │ └─14273 /pause ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount