Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─session-9.scope │ │ ├─19502 sshd: zuul [priv] │ │ └─19515 sshd: zuul@notty │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─613 /lib/systemd/systemd --user │ │ └─614 (sd-pam) │ ├─session-1.scope │ │ ├─ 610 sshd: zuul [priv] │ │ ├─ 627 sshd: zuul@notty │ │ ├─ 764 /usr/bin/python3 │ │ ├─ 9499 ssh: /home/zuul/.ansible/cp/3ce8f0ffbe [mux] │ │ ├─19506 ssh: /home/zuul/.ansible/cp/199.19.213.137-22-zuul [mux] │ │ ├─20156 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-oztjvhaqblocelpfqrxnphytbkdeopbb ; /usr/bin/python3'"'"' && sleep 0' │ │ ├─20157 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-oztjvhaqblocelpfqrxnphytbkdeopbb ; /usr/bin/python3' && sleep 0 │ │ ├─20158 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-oztjvhaqblocelpfqrxnphytbkdeopbb ; /usr/bin/python3 │ │ ├─20159 /bin/sh -c echo BECOME-SUCCESS-oztjvhaqblocelpfqrxnphytbkdeopbb ; /usr/bin/python3 │ │ ├─20160 /usr/bin/python3 │ │ ├─20161 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ │ └─20163 systemd-cgls --full --all --no-pager │ └─session-5.scope │ ├─9495 sshd: zuul [priv] │ └─9508 sshd: zuul@notty ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─12046 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─15321 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 50c460ec771b9c18837756215aaf0fe5c53c114077e69ebaeee7e390c985c330 -address /run/containerd/containerd.sock │ │ ├─15338 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1e253144a165b0971a8e69da914d6a6aab8e8871ceb2d7d788bf201f6c992fd4 -address /run/containerd/containerd.sock │ │ ├─15414 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4d4dc423a63ee05971a460712e2718d664b974358789fcb3f8d08ecb31e6de2c -address /run/containerd/containerd.sock │ │ ├─15417 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0c0f8d2e90c25d6bc128d0881f124c349eb40780fe37909ac7d1ed6136db73d1 -address /run/containerd/containerd.sock │ │ ├─15422 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 88820ce150fab5f2388ba25540c26ac0433aec75030ca883327341c9af4690d9 -address /run/containerd/containerd.sock │ │ ├─16075 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3317380f33f1b7c80e05a024e2c9e71ae35db660be70efd4b28a8be35a39c083 -address /run/containerd/containerd.sock │ │ ├─17131 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 470871364ff5629c76b190aee46c0f550104027baf6eb065b04f092d9f06e015 -address /run/containerd/containerd.sock │ │ ├─17174 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7e1154452e0b34a59f8ba5c7bcb31665a0c629d8850b784011f32fb298af3396 -address /run/containerd/containerd.sock │ │ ├─18355 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b38cba07084e2802760cb18b4f49a0ee3ba8c78beedef3e350f38b52fe08b1e3 -address /run/containerd/containerd.sock │ │ ├─18770 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 14075980e2f3666caab0666ef3371a833d79a2becdecf48f1df5c95ff85ae3eb -address /run/containerd/containerd.sock │ │ ├─18810 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a34da34168ea5a46f4ab42cbe257996b426865353ac3fbf79e3f5ed242c41dba -address /run/containerd/containerd.sock │ │ └─19262 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8f2d4d603f6a94d5db3ef322b12acd70e4eee7ce70cc3078e95e0e3c8374497f -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1090 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─436 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ ├─ 328 /lib/systemd/systemd-udevd │ │ ├─20136 /lib/systemd/systemd-udevd │ │ └─20137 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─538 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─555 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─531 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─15759 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.19.213.137 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─296 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─557 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1448 /usr/sbin/uuidd --socket-activation │ ├─systemd-resolved.service │ │ └─331 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─dbus.service │ │ └─528 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─333 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─537 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─534 /lib/systemd/systemd-logind ├─k8s.io │ ├─1e50e78608889276878bd6f472e73d3adbc32153563e087b0becd5296b998bde │ │ └─15632 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─0c0f8d2e90c25d6bc128d0881f124c349eb40780fe37909ac7d1ed6136db73d1 │ │ └─15502 /pause │ ├─840da859d2d54d5ba4721bc5009bc7265aad2951951c5b0bfa76fd630b5ed0cd │ │ └─18884 /coredns -conf /etc/coredns/Corefile │ ├─5752b2dfe1cf48a59446679f8b2426a87b828529603be49e0f16c00a3351bcd5 │ │ └─15582 etcd --advertise-client-urls=https://199.19.213.137:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.19.213.137:2380 --initial-cluster=instance=https://199.19.213.137:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.19.213.137:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.19.213.137:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─22cbfe04e0ab34c42208d9462b6b06bb0c784dc5d7637b3b2ffcc868074d851c │ │ ├─17633 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─17968 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─2781de2a974d966550a55460e97a84ff405983519114205ea415e64ea1e7150c │ │ └─15710 /kube-vip manager │ ├─470871364ff5629c76b190aee46c0f550104027baf6eb065b04f092d9f06e015 │ │ └─17158 /pause │ ├─3317380f33f1b7c80e05a024e2c9e71ae35db660be70efd4b28a8be35a39c083 │ │ └─16100 /pause │ ├─5c793373b0151a911cec4738da730d23f11f588b829cb6c02e939b9ae91f18e0 │ │ └─15569 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─50c460ec771b9c18837756215aaf0fe5c53c114077e69ebaeee7e390c985c330 │ │ └─15373 /pause │ ├─e09b68f86fddcedb84f690759699ffd700e1262117e53f83c877279e211499ac │ │ └─16124 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─14075980e2f3666caab0666ef3371a833d79a2becdecf48f1df5c95ff85ae3eb │ │ └─18799 /pause │ ├─8f2d4d603f6a94d5db3ef322b12acd70e4eee7ce70cc3078e95e0e3c8374497f │ │ └─19285 /pause │ ├─3b13a7a7e53199f46d64b210dbec0a2479c6cf55bda5eeccbe16ef444697319c │ │ └─18405 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─d3de1545f2229249fd123595053a8160cb34c7f3bd973fad0be9e3464d80e74e │ │ └─17428 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─7e1154452e0b34a59f8ba5c7bcb31665a0c629d8850b784011f32fb298af3396 │ │ └─17200 /pause │ ├─88820ce150fab5f2388ba25540c26ac0433aec75030ca883327341c9af4690d9 │ │ └─15500 /pause │ ├─b38cba07084e2802760cb18b4f49a0ee3ba8c78beedef3e350f38b52fe08b1e3 │ │ └─18379 /pause │ ├─4d4dc423a63ee05971a460712e2718d664b974358789fcb3f8d08ecb31e6de2c │ │ └─15503 /pause │ ├─a34da34168ea5a46f4ab42cbe257996b426865353ac3fbf79e3f5ed242c41dba │ │ └─18833 /pause │ ├─c56d8e3d5967ad0be4469877c5e9e299155dc325bd9a6513178c75ede66d4028 │ │ └─18865 /coredns -conf /etc/coredns/Corefile │ ├─0593ed49e00d0d62fac0da39fd9eff33b2de1c6fcc2d0af2016874b5ba0c0c49 │ │ └─15600 kube-apiserver --advertise-address=199.19.213.137 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─1e253144a165b0971a8e69da914d6a6aab8e8871ceb2d7d788bf201f6c992fd4 │ │ └─15383 /pause │ └─ae5175b9b58e95da76aef77d3a677d684001ce292e8252d91a432e9ba19a25d1 │ └─19400 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount