Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─544 /lib/systemd/systemd --user │ │ └─545 (sd-pam) │ ├─session-4.scope │ │ ├─1243 sshd: zuul [priv] │ │ └─1256 sshd: zuul@notty │ └─session-1.scope │ ├─ 541 sshd: zuul [priv] │ ├─ 558 sshd: zuul@notty │ ├─ 689 /usr/bin/python3 │ ├─ 1247 ssh: /home/zuul/.ansible/cp/199.19.213.143-22-zuul [mux] │ ├─16900 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-rtdzqkscthpfretrpjuussedjzeryhsl ; /usr/bin/python3'"'"' && sleep 0' │ ├─16901 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-rtdzqkscthpfretrpjuussedjzeryhsl ; /usr/bin/python3' && sleep 0 │ ├─16902 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-rtdzqkscthpfretrpjuussedjzeryhsl ; /usr/bin/python3 │ ├─16903 /bin/sh -c echo BECOME-SUCCESS-rtdzqkscthpfretrpjuussedjzeryhsl ; /usr/bin/python3 │ ├─16904 /usr/bin/python3 │ ├─16905 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─16906 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 38 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 9465 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─11041 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f675c5cfa2c9c438bfcb5db30e644e362919246caef1ec634bd06d58b6ce0ff2 -address /run/containerd/containerd.sock │ │ ├─11080 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 25e6c8635328c4fedcce8755fef27c0a264d216ad448d45778af2b9dc5fc6046 -address /run/containerd/containerd.sock │ │ ├─11084 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6ed7423c363a414bfa452edec87037f3941053924c5fcd0ff4d04d1acd2c683f -address /run/containerd/containerd.sock │ │ ├─11092 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d210c0eb90479355507fdc85fac2ab3ab78810cf857c554cb5c5f84e17c75d2c -address /run/containerd/containerd.sock │ │ ├─11093 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9ba3760c3818db69c0157f3a5328c808e0a1d358cff78df5c1b56d78389383eb -address /run/containerd/containerd.sock │ │ ├─11550 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 465f244884f82bb39be23fe4a759cd12aa10d8cf5dbd1059174e94f504ad2e19 -address /run/containerd/containerd.sock │ │ ├─12185 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7555bc70b71c372fb251423dc0d3d6d26dc2ce6312097fa8f31a4f3098f66460 -address /run/containerd/containerd.sock │ │ ├─12201 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 71bc60d855a6891cca911af1659b967b671ea398ac3ae6ace19fd38d186d4ee1 -address /run/containerd/containerd.sock │ │ ├─14108 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1aa61de3b2bf405ebc8b08f375890d9a522b6bbcc69a2f7ab671a4d65bb23208 -address /run/containerd/containerd.sock │ │ ├─14117 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b46ec90a1334e0119581231faee66df3e03012b70f1f0e8024066c060943e99e -address /run/containerd/containerd.sock │ │ └─14146 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e0543fc9dc7938d7ed3c9c02b6f22efe3147c750c12f3b713053df25c56238c7 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1013 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─363 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ ├─ 267 /lib/systemd/systemd-udevd │ │ ├─16888 /lib/systemd/systemd-udevd │ │ └─16889 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─470 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─487 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─459 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─11443 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.19.213.143 --pod-infra-container-image=registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─236 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─486 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─cloud-final.service │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─270 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─cloud-init.service │ ├─dbus.service │ │ └─455 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─271 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─469 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─461 /lib/systemd/systemd-logind ├─k8s.io │ ├─d210c0eb90479355507fdc85fac2ab3ab78810cf857c554cb5c5f84e17c75d2c │ │ └─11202 /pause │ ├─601abfd4702bb0b912859e21a9c99b74fd23ec1d2bc00b0883950d8be4380022 │ │ └─11395 /kube-vip manager │ ├─864e784db8502f4088d6dcd483d26b8b15083729566c752019d0c38aa6ba69f7 │ │ └─11277 etcd --advertise-client-urls=https://199.19.213.143:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.19.213.143:2380 --initial-cluster=instance=https://199.19.213.143:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.19.213.143:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.19.213.143:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─16c6759e53fa34ad1e78d138be4833d138f8f091265b1aacc6d3893e61cc8c03 │ │ └─11306 kube-apiserver --advertise-address=199.19.213.143 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─1aa61de3b2bf405ebc8b08f375890d9a522b6bbcc69a2f7ab671a4d65bb23208 │ │ └─14200 /pause │ ├─14d1dbe46444815ab47d43a9137b6dc7b7cdc70cfaab07a53f16a3ed9d57aeef │ │ └─14244 /coredns -conf /etc/coredns/Corefile │ ├─f675c5cfa2c9c438bfcb5db30e644e362919246caef1ec634bd06d58b6ce0ff2 │ │ └─11129 /pause │ ├─25e6c8635328c4fedcce8755fef27c0a264d216ad448d45778af2b9dc5fc6046 │ │ └─11203 /pause │ ├─9ba3760c3818db69c0157f3a5328c808e0a1d358cff78df5c1b56d78389383eb │ │ └─11193 /pause │ ├─d3a144f514a51d607b823a2444e94f27ad87a13720a57fbad03a0fd06552f274 │ │ └─12937 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─c0021193bb0b8036e3391b27400bfc310728b7e9e0d7b7c81f1fe4d5210307ac │ │ └─14280 /coredns -conf /etc/coredns/Corefile │ ├─6ed7423c363a414bfa452edec87037f3941053924c5fcd0ff4d04d1acd2c683f │ │ └─11155 /pause │ ├─20031237b87a20d4e03761c9dabeaf578bedc3465eb2c643bce538a10d135818 │ │ └─11284 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─b46ec90a1334e0119581231faee66df3e03012b70f1f0e8024066c060943e99e │ │ └─14175 /pause │ ├─4e6a0f4ec27d88f9a9a914ccee6cda398f6e9d90b826e7ed71733ffbb6fa7646 │ │ └─11597 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─465f244884f82bb39be23fe4a759cd12aa10d8cf5dbd1059174e94f504ad2e19 │ │ └─11573 /pause │ ├─f2e459110b55e6f5b630e7755de99d62d5dc55abd70775d46c7be7030fb56439 │ │ └─14391 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─7555bc70b71c372fb251423dc0d3d6d26dc2ce6312097fa8f31a4f3098f66460 │ │ └─12232 /pause │ ├─e28012cebff011d1dbdaf48b603661db7b994c1f44edd1405eb3c21dab6a3089 │ │ └─11324 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─71bc60d855a6891cca911af1659b967b671ea398ac3ae6ace19fd38d186d4ee1 │ │ └─12248 /pause │ ├─e0543fc9dc7938d7ed3c9c02b6f22efe3147c750c12f3b713053df25c56238c7 │ │ └─14189 /pause │ └─9ca60a0537ea1caaeb0355b17e3e6d67b2e910b3000b29a328139e56368964ac │ ├─13246 cilium-agent --config-dir=/tmp/cilium/config-map │ └─13768 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount