Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─617 /lib/systemd/systemd --user │ │ └─618 (sd-pam) │ ├─session-4.scope │ │ ├─1351 sshd: zuul [priv] │ │ └─1364 sshd: zuul@notty │ └─session-1.scope │ ├─ 614 sshd: zuul [priv] │ ├─ 631 sshd: zuul@notty │ ├─ 768 /usr/bin/python3 │ ├─ 1355 ssh: /home/zuul/.ansible/cp/199.204.45.153-22-zuul [mux] │ ├─18473 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-eyrdyatwfzedjgvayswpkvwuooxtgyzl ; /usr/bin/python3'"'"' && sleep 0' │ ├─18474 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-eyrdyatwfzedjgvayswpkvwuooxtgyzl ; /usr/bin/python3' && sleep 0 │ ├─18475 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-eyrdyatwfzedjgvayswpkvwuooxtgyzl ; /usr/bin/python3 │ ├─18476 /bin/sh -c echo BECOME-SUCCESS-eyrdyatwfzedjgvayswpkvwuooxtgyzl ; /usr/bin/python3 │ ├─18477 /usr/bin/python3 │ ├─18478 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─18480 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 12 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 9614 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─11224 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id bb122f94d040079dafb634c1ab62bd0e2dd2edf1656bee64b936dc7cdf334a69 -address /run/containerd/containerd.sock │ │ ├─11254 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b9439ed60fcdb1e9cc223786abfd85570424066494e35aedc715c5c2ef6a7ee4 -address /run/containerd/containerd.sock │ │ ├─11279 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a1f215b183169326322b6ececf821027d4df03863b59131d25ae2b0853fdf3e4 -address /run/containerd/containerd.sock │ │ ├─11307 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a040dd38fd37cd59437ad2219d251b51c617f21cc6fbf36bbfa5f57a2ea093d0 -address /run/containerd/containerd.sock │ │ ├─11310 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 90973713607f37b077b48a65b492961ade3d894521bae8212e05961cfe485d0e -address /run/containerd/containerd.sock │ │ ├─11766 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9075b74b0d24e43e9e7f0ecd6d54378da6f04d7d1c37aca748db76c46bec036c -address /run/containerd/containerd.sock │ │ ├─12479 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d9763d7565f07b53f2b16fdca8e13f87080e9f70e7f57e52e15d470037d12d86 -address /run/containerd/containerd.sock │ │ ├─12516 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 33f9df45586a567a1d92735beede1c7fde8f52dba09626f9480e22734ad475fa -address /run/containerd/containerd.sock │ │ ├─13689 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 2ef115b8d0013c4dd2b058d72d090930b8d45c5dee0b8937a67f338a203d0b2c -address /run/containerd/containerd.sock │ │ ├─14057 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 41780395bbe10ead8983f9312b7975a4d7e1067cac247655b482f40157812dac -address /run/containerd/containerd.sock │ │ ├─14094 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 21a96a6aa59061502f5724a9b991399aeeb56fc397ee9d89de92baaa9ab573e2 -address /run/containerd/containerd.sock │ │ └─14456 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 75782e6e4b4721c9ce68533d85f2d711367eb7dadc8ac13113954d82476b506c -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1092 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─438 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ ├─ 331 /lib/systemd/systemd-udevd │ │ ├─18459 /lib/systemd/systemd-udevd │ │ └─18460 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─542 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─559 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─534 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─11667 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.153 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─296 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─560 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─uuidd.service │ │ └─1286 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─335 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─dbus.service │ │ └─530 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─336 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─539 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─536 /lib/systemd/systemd-logind ├─k8s.io │ ├─41780395bbe10ead8983f9312b7975a4d7e1067cac247655b482f40157812dac │ │ └─14080 /pause │ ├─7ef69368edcec4fb33185be6f264462e7900cef047a0ded3da691fb77d35ef95 │ │ └─14532 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─f9754bd8eaf3d398cf87c9f604f023cb6241b859f294824c2a5116eae91ff932 │ │ └─11486 kube-apiserver --advertise-address=199.204.45.153 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─75782e6e4b4721c9ce68533d85f2d711367eb7dadc8ac13113954d82476b506c │ │ └─14481 /pause │ ├─9075b74b0d24e43e9e7f0ecd6d54378da6f04d7d1c37aca748db76c46bec036c │ │ └─11791 /pause │ ├─2ef115b8d0013c4dd2b058d72d090930b8d45c5dee0b8937a67f338a203d0b2c │ │ └─13713 /pause │ ├─b9439ed60fcdb1e9cc223786abfd85570424066494e35aedc715c5c2ef6a7ee4 │ │ └─11355 /pause │ ├─9f36da9db66b65d0c9253005f86aa69a13bdb74d40f68863c7cf54aef4fd7470 │ │ └─14152 /coredns -conf /etc/coredns/Corefile │ ├─4038b7a94b71497982fee252a5343aa24b3ba920e3c269c4d0c346529ce88b73 │ │ └─11828 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─21a96a6aa59061502f5724a9b991399aeeb56fc397ee9d89de92baaa9ab573e2 │ │ └─14118 /pause │ ├─a040dd38fd37cd59437ad2219d251b51c617f21cc6fbf36bbfa5f57a2ea093d0 │ │ └─11400 /pause │ ├─33f9df45586a567a1d92735beede1c7fde8f52dba09626f9480e22734ad475fa │ │ └─12552 /pause │ ├─90973713607f37b077b48a65b492961ade3d894521bae8212e05961cfe485d0e │ │ └─11384 /pause │ ├─a1f215b183169326322b6ececf821027d4df03863b59131d25ae2b0853fdf3e4 │ │ └─11364 /pause │ ├─740fe975ef867af51b93ec48481b8d9a3305b17d6a01eef8c4477e5380f94172 │ │ ├─12964 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─13334 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─fc7336622c53fbc37d864630bc395fda9d8d0b676e8de3c03f8166d41e280478 │ │ └─12756 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─21d9576e47f719f69f21dee8fab083e859d39f02e38bed41c68a0e9776224305 │ │ └─13748 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─e03ae54db7d56c9e60adf7f31bccb2a722076796ce77cf421d834ab27d8a0e0a │ │ └─14176 /coredns -conf /etc/coredns/Corefile │ ├─bb122f94d040079dafb634c1ab62bd0e2dd2edf1656bee64b936dc7cdf334a69 │ │ └─11283 /pause │ ├─d9763d7565f07b53f2b16fdca8e13f87080e9f70e7f57e52e15d470037d12d86 │ │ └─12506 /pause │ ├─3d2cc37d0a3af334961e1a057ba4ebd5bf65c400a7840ca433af000fcd88324f │ │ └─11616 /kube-vip manager │ ├─33e3ee9e7e04d30d7b617a424559a968fcf33ed149e3d7b28f02b737f67c620f │ │ └─11465 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─e7c67db3fa96ecd20fc5611d72a09c87e5ea5838fe84592ab2bcf8f31b0158e5 │ │ └─11525 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ └─f0c6185ddda45017337b2763404a76b77000df1bfdc8206f0e62d07bcaa1fcc5 │ └─11504 etcd --advertise-client-urls=https://199.204.45.153:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.153:2380 --initial-cluster=instance=https://199.204.45.153:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.153:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.153:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount