Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─session-9.scope │ │ ├─19916 sshd: zuul [priv] │ │ └─19929 sshd: zuul@notty │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─618 /lib/systemd/systemd --user │ │ └─619 (sd-pam) │ ├─session-1.scope │ │ ├─ 615 sshd: zuul [priv] │ │ ├─ 632 sshd: zuul@notty │ │ ├─ 769 /usr/bin/python3 │ │ ├─ 9504 ssh: /home/zuul/.ansible/cp/40081a297a [mux] │ │ ├─19920 ssh: /home/zuul/.ansible/cp/199.204.45.116-22-zuul [mux] │ │ ├─20534 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-hfhbhkkhmhpduumziptoakvymyalkydi ; /usr/bin/python3'"'"' && sleep 0' │ │ ├─20535 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-hfhbhkkhmhpduumziptoakvymyalkydi ; /usr/bin/python3' && sleep 0 │ │ ├─20536 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-hfhbhkkhmhpduumziptoakvymyalkydi ; /usr/bin/python3 │ │ ├─20537 /bin/sh -c echo BECOME-SUCCESS-hfhbhkkhmhpduumziptoakvymyalkydi ; /usr/bin/python3 │ │ ├─20538 /usr/bin/python3 │ │ ├─20539 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ │ └─20541 systemd-cgls --full --all --no-pager │ └─session-5.scope │ ├─9500 sshd: zuul [priv] │ └─9513 sshd: zuul@notty ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─12049 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─15307 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d404419b6f0f2e02cf62af21bb3a076640e8488923ba48d2315f63e47f7dc39f -address /run/containerd/containerd.sock │ │ ├─15325 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a309e1d73e7e98167ee5bfae2da95775a8377279bb1d8915e4e8d7e18fe15d53 -address /run/containerd/containerd.sock │ │ ├─15368 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id fe969485ea5463a47ce9d7f377732a1bce46973f7fa6ad7f0028d487fc3f3ca4 -address /run/containerd/containerd.sock │ │ ├─15370 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 2e6643eb5da744c2c375e075ac9c701f47fe48a0f75901404c32ea9c41f9375c -address /run/containerd/containerd.sock │ │ ├─15374 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 112d7552def7ef273f5ad2d26e6b0b0eb7a4e0e940546e965ede6f395b8c0cfa -address /run/containerd/containerd.sock │ │ ├─16121 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 28a51c688caf98a6bd08a8c60b75d99a87511aa78945aec6502cd8a60211ecab -address /run/containerd/containerd.sock │ │ ├─17118 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 252f7f273203130e40705937c455fab28d1fee91663a331fffc87ed1992649ac -address /run/containerd/containerd.sock │ │ ├─17127 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ea0c7352e5547fe193fbb11681f7d2d8097d5ea67bad611eeb929346a0e98264 -address /run/containerd/containerd.sock │ │ ├─18738 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5a7163b8a842bdbe7210d7780cc751d385a4c72564b4bea11d9760c44a458b07 -address /run/containerd/containerd.sock │ │ ├─19177 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6a4ee1bbc1933f602f725f60f1e9d31b4200b098511d90b2400a200a31c9b8f6 -address /run/containerd/containerd.sock │ │ ├─19204 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d995211911d1f7729b832570a61b05186268312f932646d304e9c5b0db7516de -address /run/containerd/containerd.sock │ │ └─19740 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id dcdb4a8635546bcb028e0da8e73320ca1453b1fa105ddefd63bc867254e2f1f7 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1094 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─438 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ ├─ 327 /lib/systemd/systemd-udevd │ │ ├─20520 /lib/systemd/systemd-udevd │ │ └─20521 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─543 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─560 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─533 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─15742 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.116 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─cloud-init-local.service │ ├─systemd-journald.service │ │ └─296 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─562 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1448 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─333 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─cloud-init.service │ ├─dbus.service │ │ └─530 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─334 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─539 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─536 /lib/systemd/systemd-logind ├─k8s.io │ ├─cc2fadca466493d74d543c474c84f0b2bea6a478d4e4eb998deb5ff9da366433 │ │ └─19818 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─cfb6022a535c2086a2c3a6665f43c570a9843f18b9617ab573b97e0fb9cd957f │ │ └─19271 /coredns -conf /etc/coredns/Corefile │ ├─6a4ee1bbc1933f602f725f60f1e9d31b4200b098511d90b2400a200a31c9b8f6 │ │ └─19217 /pause │ ├─252f7f273203130e40705937c455fab28d1fee91663a331fffc87ed1992649ac │ │ └─17175 /pause │ ├─d2538f659a507958716ad28cd0e9c442597df0624936baf379b5e9b2b6a93bfb │ │ └─18798 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─edcf2d2e08e61164789174685b440fc546d56d6d5cbb4ae1b5661191924054da │ │ └─16169 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─d995211911d1f7729b832570a61b05186268312f932646d304e9c5b0db7516de │ │ └─19240 /pause │ ├─d55e144e71946d7de0fd544ff0d92766eaf63a1a348118613992834bd82922e3 │ │ └─15548 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─5a7163b8a842bdbe7210d7780cc751d385a4c72564b4bea11d9760c44a458b07 │ │ └─18761 /pause │ ├─2e6643eb5da744c2c375e075ac9c701f47fe48a0f75901404c32ea9c41f9375c │ │ └─15447 /pause │ ├─7f6306eda4808fce4b6ae587fb8c690c11ce6583af7611c6b27850e252a8e0ba │ │ └─15589 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─112d7552def7ef273f5ad2d26e6b0b0eb7a4e0e940546e965ede6f395b8c0cfa │ │ └─15483 /pause │ ├─a309e1d73e7e98167ee5bfae2da95775a8377279bb1d8915e4e8d7e18fe15d53 │ │ └─15431 /pause │ ├─dcdb4a8635546bcb028e0da8e73320ca1453b1fa105ddefd63bc867254e2f1f7 │ │ └─19765 /pause │ ├─7df315bd3ecc43c0f23fde49794efda88c6c844270a076c425e5b5501901fcbb │ │ ├─18038 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─18361 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─ec8e16d772ba97bd428e222fe3142f730bcbd7224f884cd711b20876ec605bdf │ │ └─15582 etcd --advertise-client-urls=https://199.204.45.116:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.116:2380 --initial-cluster=instance=https://199.204.45.116:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.116:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.116:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─a0f1314cfab3f2f28dd716e16f75fb4917e5cdf242a4c8e2446ad94c63352887 │ │ └─15694 /kube-vip manager │ ├─9da0b904c51547c95653c380c3db5628abfb8d61da14a15ab5dd0bca1ed79bba │ │ └─17841 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─28a51c688caf98a6bd08a8c60b75d99a87511aa78945aec6502cd8a60211ecab │ │ └─16145 /pause │ ├─ea0c7352e5547fe193fbb11681f7d2d8097d5ea67bad611eeb929346a0e98264 │ │ └─17165 /pause │ ├─d404419b6f0f2e02cf62af21bb3a076640e8488923ba48d2315f63e47f7dc39f │ │ └─15397 /pause │ ├─fe969485ea5463a47ce9d7f377732a1bce46973f7fa6ad7f0028d487fc3f3ca4 │ │ └─15469 /pause │ ├─09bca13c07d5d6f299f050cb2c08d62f81e35e64042ffa0a1d2809321562cdc2 │ │ └─15607 kube-apiserver --advertise-address=199.204.45.116 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ └─d3a6ec49b42fe6d6eed71de93d93e6f4565b259bd100af64b4d16c0dd99bfd1e │ └─19297 /coredns -conf /etc/coredns/Corefile ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount