Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─619 /lib/systemd/systemd --user │ │ └─620 (sd-pam) │ ├─session-4.scope │ │ ├─1355 sshd: zuul [priv] │ │ └─1368 sshd: zuul@notty │ └─session-1.scope │ ├─ 616 sshd: zuul [priv] │ ├─ 633 sshd: zuul@notty │ ├─ 773 /usr/bin/python3 │ ├─ 1359 ssh: /home/zuul/.ansible/cp/199.204.45.210-22-zuul [mux] │ ├─18343 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-qvpluxliknnmaufzkqdttkftgbgmvnnh ; /usr/bin/python3'"'"' && sleep 0' │ ├─18344 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-qvpluxliknnmaufzkqdttkftgbgmvnnh ; /usr/bin/python3' && sleep 0 │ ├─18345 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-qvpluxliknnmaufzkqdttkftgbgmvnnh ; /usr/bin/python3 │ ├─18346 /bin/sh -c echo BECOME-SUCCESS-qvpluxliknnmaufzkqdttkftgbgmvnnh ; /usr/bin/python3 │ ├─18347 /usr/bin/python3 │ ├─18348 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─18350 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 9617 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─11213 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 00222ad25f8e2733fa82f1944cbc9a8661cdb3493bc6c5a5b5cc6d42e1163168 -address /run/containerd/containerd.sock │ │ ├─11218 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 07f15b23ee1fc47dbd2b6b2bb43f037a4aefcfeb5d47190197a06d78365c6ebd -address /run/containerd/containerd.sock │ │ ├─11236 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id eaa870ac9fb7e1790ad306248ffcb36f4f9e7b18ae0796d7652b7ad83dc194f9 -address /run/containerd/containerd.sock │ │ ├─11250 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4179ff12255782337aaeea5feea2411fdcddbf13c182be532923245ce7c1e668 -address /run/containerd/containerd.sock │ │ ├─11264 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b37aeedc92359f138dd9046dea7ea1a066c70002ed541e78cdf70f525a5706f6 -address /run/containerd/containerd.sock │ │ ├─11797 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d956f483548c735ee2e78ceca691c710a28480bf4dda4558061e77544fe06773 -address /run/containerd/containerd.sock │ │ ├─12442 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b49057088e396a6e44e46154113a03f71ded2fc8ac3c2820f0b14c1f8cf34378 -address /run/containerd/containerd.sock │ │ ├─12446 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a691db8d4215446aeb96914e072b727ed3c850892df3e23a7c675c52700fbec3 -address /run/containerd/containerd.sock │ │ ├─13631 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b364d5ec1a53c4af45ca4943535db4f325d29b34d3b0173cc764c40353a28445 -address /run/containerd/containerd.sock │ │ ├─13977 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a62361482893359a648dec603f0a79ab41037a1c317b7bec74b2d7529fd81757 -address /run/containerd/containerd.sock │ │ ├─14015 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8a4760180be27a742bec824835b86b97c4075a4945a2f77db9899a3894e7feb8 -address /run/containerd/containerd.sock │ │ └─14371 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b7a68fcd5128e82614fe4e32aed854f39185ef71d55577d2a4d936ca0a78ebff -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1094 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─439 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ ├─ 330 /lib/systemd/systemd-udevd │ │ ├─18330 /lib/systemd/systemd-udevd │ │ └─18331 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─543 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─559 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─533 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─11643 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.210 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─299 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─565 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─uuidd.service │ │ └─1288 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─334 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─cloud-init.service │ ├─dbus.service │ │ └─530 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─335 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─540 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─536 /lib/systemd/systemd-logind ├─k8s.io │ ├─00222ad25f8e2733fa82f1944cbc9a8661cdb3493bc6c5a5b5cc6d42e1163168 │ │ └─11324 /pause │ ├─14463883e0c134243df28d407ec0f29e0e26fd92de8f7927c4ab8988f096faa4 │ │ └─12693 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─d956f483548c735ee2e78ceca691c710a28480bf4dda4558061e77544fe06773 │ │ └─11822 /pause │ ├─d888ee470f1d2be5df0841d03b5f538348d4e1e8c49ba0482421a55348c6943d │ │ └─11487 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─d2a9f467951072f986022ca863578c3505a079f1695a32e7170dde4672d9ea0d │ │ └─14515 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─39c6579333a50755eeabb593e37ce8d6351e444d47e464db1b6a8056f905a796 │ │ ├─12905 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─13239 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─d19172f438193e76b47bbf0c7db70fe840bb7c8321d8a053f56fc94ae94abdde │ │ └─13692 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─eaa870ac9fb7e1790ad306248ffcb36f4f9e7b18ae0796d7652b7ad83dc194f9 │ │ └─11343 /pause │ ├─e33378432bad2b9b2519fe79a0300e0c082562299eddbdf218e8edc1d6c2517c │ │ └─14069 /coredns -conf /etc/coredns/Corefile │ ├─777ad0d8adb10e47b113e7ea49775b2f98bdaeda421ac878f8361cba3a937b3d │ │ └─11593 /kube-vip manager │ ├─b7a68fcd5128e82614fe4e32aed854f39185ef71d55577d2a4d936ca0a78ebff │ │ └─14396 /pause │ ├─b37aeedc92359f138dd9046dea7ea1a066c70002ed541e78cdf70f525a5706f6 │ │ └─11382 /pause │ ├─a691db8d4215446aeb96914e072b727ed3c850892df3e23a7c675c52700fbec3 │ │ └─12498 /pause │ ├─a62361482893359a648dec603f0a79ab41037a1c317b7bec74b2d7529fd81757 │ │ └─14005 /pause │ ├─6b1819b35950a95ed65d01ed69af5faff4eba007e775b5edee2ad84dfc96e6c1 │ │ └─11505 kube-apiserver --advertise-address=199.204.45.210 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─07f15b23ee1fc47dbd2b6b2bb43f037a4aefcfeb5d47190197a06d78365c6ebd │ │ └─11362 /pause │ ├─b48b2b234e045a1004fdaece34845a3cdb8a18d1b69b2d7ebde8c42ab9f860cd │ │ └─14090 /coredns -conf /etc/coredns/Corefile │ ├─8a4760180be27a742bec824835b86b97c4075a4945a2f77db9899a3894e7feb8 │ │ └─14038 /pause │ ├─2f4bcd689c4bbafa8f593956bb97ae644477dc08681dcd287a24ce00c2644ae5 │ │ └─11468 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─043d1ea9e908e5f59115707e1ae59c168eca802737a8d8c77422dbfe089efd19 │ │ └─11450 etcd --advertise-client-urls=https://199.204.45.210:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.210:2380 --initial-cluster=instance=https://199.204.45.210:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.210:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.210:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─b49057088e396a6e44e46154113a03f71ded2fc8ac3c2820f0b14c1f8cf34378 │ │ └─12491 /pause │ ├─4179ff12255782337aaeea5feea2411fdcddbf13c182be532923245ce7c1e668 │ │ └─11364 /pause │ ├─b364d5ec1a53c4af45ca4943535db4f325d29b34d3b0173cc764c40353a28445 │ │ └─13656 /pause │ └─efe735bc0bdf229b83d8da086dac9e4688f3c6be5420322860ba5f478af62506 │ └─11851 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount