Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─session-9.scope │ │ ├─26343 sshd: zuul [priv] │ │ └─26356 sshd: zuul@notty │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─617 /lib/systemd/systemd --user │ │ └─618 (sd-pam) │ ├─session-1.scope │ │ ├─ 614 sshd: zuul [priv] │ │ ├─ 631 sshd: zuul@notty │ │ ├─ 768 /usr/bin/python3 │ │ ├─16032 ssh: /home/zuul/.ansible/cp/b7ec02faec [mux] │ │ ├─26347 ssh: /home/zuul/.ansible/cp/199.204.45.244-22-zuul [mux] │ │ ├─27001 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-htkuaprxucavczkpgklxlcbauvnbcxvl ; /usr/bin/python3'"'"' && sleep 0' │ │ ├─27002 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-htkuaprxucavczkpgklxlcbauvnbcxvl ; /usr/bin/python3' && sleep 0 │ │ ├─27003 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-htkuaprxucavczkpgklxlcbauvnbcxvl ; /usr/bin/python3 │ │ ├─27004 /bin/sh -c echo BECOME-SUCCESS-htkuaprxucavczkpgklxlcbauvnbcxvl ; /usr/bin/python3 │ │ ├─27005 /usr/bin/python3 │ │ ├─27006 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ │ └─27008 systemd-cgls --full --all --no-pager │ └─session-5.scope │ ├─16028 sshd: zuul [priv] │ └─16041 sshd: zuul@notty ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─18629 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─21787 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 97002b3b797ce490da5cdb62ce265c5afa8c91e147301a0d803495044be4f3ad -address /run/containerd/containerd.sock │ │ ├─21816 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0d69df692ab491e68dc573c2bec4f05fbd481415a8fa519dbaa0193481e94b64 -address /run/containerd/containerd.sock │ │ ├─21831 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1dc9ce3ef66035232a5c8ba5a951d274cee5e19a843dde3651a67ab5e87ed7fe -address /run/containerd/containerd.sock │ │ ├─21845 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a2a42f836b43db36f68b6135d5d1c6bca96664980c31828a4fb6dddb3fb10e0f -address /run/containerd/containerd.sock │ │ ├─22064 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3c63824b97583b7f1414a053be0608e39057bc62763e6b49bd69245f77d84ea9 -address /run/containerd/containerd.sock │ │ ├─22619 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 026f511449d6549619ac74453c7a136ca90111f6631093835cb564c84a2b17d7 -address /run/containerd/containerd.sock │ │ ├─23611 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6a4360544653a45347beb0c77e727fe9941a868eae31f99bc0eb8c26c39a0447 -address /run/containerd/containerd.sock │ │ ├─23742 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b5919b703665e83266e92c63017debfc9c9576bc93abbeb1935ef2e02147e18e -address /run/containerd/containerd.sock │ │ ├─25079 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 86eb0590358e37a4a59cae1d063bd5fcde78f34d96d3863d6ab986f5f644aa59 -address /run/containerd/containerd.sock │ │ ├─25551 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 03deae0d9a4f8318c78a0bf22beac8d9639fd12c3acf07fd628fdcd966eeec79 -address /run/containerd/containerd.sock │ │ ├─25571 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 26a28e3075876c36efa1a0865e25e91569f8fd6517fc58808bbbaa5737081f15 -address /run/containerd/containerd.sock │ │ └─26061 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9da6f75cb49b170feed0b53ebd30b3cbaf05c28b7f6ebce7da7396aef6c12b5d -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1093 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─437 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ ├─ 328 /lib/systemd/systemd-udevd │ │ ├─26980 /lib/systemd/systemd-udevd │ │ └─26981 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─539 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─558 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─532 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─multipathd.service │ │ └─1445 /sbin/multipathd -d -s │ ├─kubelet.service │ │ └─22223 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.244 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─296 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─563 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─uuidd.service │ │ └─7911 /usr/sbin/uuidd --socket-activation │ ├─systemd-resolved.service │ │ └─332 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─cloud-init.service │ ├─dbus.service │ │ └─529 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─334 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─538 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─535 /lib/systemd/systemd-logind ├─k8s.io │ ├─3c63824b97583b7f1414a053be0608e39057bc62763e6b49bd69245f77d84ea9 │ │ └─22102 /pause │ ├─025584978c59f9847664aca34082d1ebefa9d3eb54966ffd65522da72fafd0c5 │ │ └─25133 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─026f511449d6549619ac74453c7a136ca90111f6631093835cb564c84a2b17d7 │ │ └─22648 /pause │ ├─522d3a96e5178b2b8d371ad328929a0940c83ba8918d20e1fdfd084d45e5573e │ │ └─26205 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─0473d055c1d5f060b95799a5cab43c99ee7c766f9751d0d48fc50f59af3909bb │ │ └─22681 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─03deae0d9a4f8318c78a0bf22beac8d9639fd12c3acf07fd628fdcd966eeec79 │ │ └─25596 /pause │ ├─26a28e3075876c36efa1a0865e25e91569f8fd6517fc58808bbbaa5737081f15 │ │ └─25615 /pause │ ├─b5919b703665e83266e92c63017debfc9c9576bc93abbeb1935ef2e02147e18e │ │ └─23766 /pause │ ├─0d69df692ab491e68dc573c2bec4f05fbd481415a8fa519dbaa0193481e94b64 │ │ └─21894 /pause │ ├─86eb0590358e37a4a59cae1d063bd5fcde78f34d96d3863d6ab986f5f644aa59 │ │ └─25105 /pause │ ├─a2a42f836b43db36f68b6135d5d1c6bca96664980c31828a4fb6dddb3fb10e0f │ │ └─21923 /pause │ ├─97002b3b797ce490da5cdb62ce265c5afa8c91e147301a0d803495044be4f3ad │ │ └─21863 /pause │ ├─1dc9ce3ef66035232a5c8ba5a951d274cee5e19a843dde3651a67ab5e87ed7fe │ │ └─21911 /pause │ ├─a737913f14307d9f3e695e63ef3842bdade88ec36bd7236bae8588b999b3e28e │ │ └─24169 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─507dbe5bfc925bc77a5089f08c7722ce650635a6e724640149c2650a1bb30f5f │ │ ├─24371 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─24727 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─5dbf573542729f0a95e5001246cce29c0c8cfaf4f590a17801e839ae6e956ba9 │ │ └─22022 kube-apiserver --advertise-address=199.204.45.244 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─608c41c5cd68635a36aaf9f392e49e6d66031f3cbad48dbc3a847f7e86aaaf26 │ │ └─25653 /coredns -conf /etc/coredns/Corefile │ ├─0d50f470bfb7155b28f3db042e2c0ed2d6ee371add00c61487f1ab3d16c78703 │ │ └─25665 /coredns -conf /etc/coredns/Corefile │ ├─6a4360544653a45347beb0c77e727fe9941a868eae31f99bc0eb8c26c39a0447 │ │ └─23636 /pause │ ├─8848a4cf8c5003ff18a66ab61894890a8ef4b551cc12315fec9d81365431b68e │ │ └─22174 /kube-vip manager │ ├─c5b7cb58964c4d0f550812812988b03e82a0c0acdad35824a237f97284b9a947 │ │ └─21989 etcd --advertise-client-urls=https://199.204.45.244:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.244:2380 --initial-cluster=instance=https://199.204.45.244:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.244:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.244:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─bb488daf2a25403691643f1ea069a1e43b323c425b4bd83c8ee79d5a7deac8ab │ │ └─22002 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─675b3bf3a0072a0e94e1632dd662786c260459e2ee032f6b6eb13e4b0a8d4b8c │ │ └─22134 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ └─9da6f75cb49b170feed0b53ebd30b3cbaf05c28b7f6ebce7da7396aef6c12b5d │ └─26086 /pause ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount