Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ ├─user-1002.slice │ │ ├─user@1002.service … │ │ │ ├─app.slice │ │ │ └─init.scope │ │ │ ├─13596 /lib/systemd/systemd --user │ │ │ └─13598 (sd-pam) │ │ └─session-21.scope │ │ ├─19833 sshd: cephadm [priv] │ │ └─19894 sshd: cephadm@notty │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─617 /lib/systemd/systemd --user │ │ └─618 (sd-pam) │ ├─session-4.scope │ │ ├─ 1414 sshd: zuul [priv] │ │ ├─ 1427 sshd: zuul@notty │ │ ├─38287 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-zfwuspiirmxtkkeaxfpnfswcwmorwogh ; /usr/bin/python3.10'"'"' && sleep 0' │ │ ├─38288 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-zfwuspiirmxtkkeaxfpnfswcwmorwogh ; /usr/bin/python3.10' && sleep 0 │ │ ├─38289 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-zfwuspiirmxtkkeaxfpnfswcwmorwogh ; /usr/bin/python3.10 │ │ ├─38290 /bin/sh -c echo BECOME-SUCCESS-zfwuspiirmxtkkeaxfpnfswcwmorwogh ; /usr/bin/python3.10 │ │ ├─38291 /usr/bin/python3.10 │ │ └─38311 /usr/bin/helm upgrade -i --reset-values --create-namespace -f=/tmp/tmpxj65m4i4.yml envoy-gateway /usr/local/src/envoy-gateway │ └─session-1.scope │ ├─ 614 sshd: zuul [priv] │ ├─ 631 sshd: zuul@notty │ ├─ 768 /usr/bin/python3 │ ├─ 1259 sh -c /bin/sh -c '/usr/bin/python3 && sleep 0' │ ├─ 1260 /bin/sh -c /usr/bin/python3 && sleep 0 │ ├─ 1261 /usr/bin/python3 │ ├─ 1262 uv run molecule test --destroy never -s csi │ ├─ 1345 /home/zuul/src/github.com/vexxhost/atmosphere/.venv/bin/python /home/zuul/src/github.com/vexxhost/atmosphere/.venv/bin/molecule test --destroy never -s csi │ ├─ 1418 ssh: /home/zuul/.ansible/cp/199.204.45.138-22-zuul [mux] │ ├─ 8468 /home/zuul/src/github.com/vexxhost/atmosphere/.venv/bin/python /home/zuul/src/github.com/vexxhost/atmosphere/.venv/bin/ansible-playbook --inventory /home/zuul/.ansible/tmp/molecule.v9Wo.csi/inventory --skip-tags molecule-notest,notest --inventory=/home/zuul/src/github.com/vexxhost/atmosphere/inventory.yaml /home/zuul/src/github.com/vexxhost/atmosphere/molecule/csi/converge.yml │ ├─38271 /home/zuul/src/github.com/vexxhost/atmosphere/.venv/bin/python /home/zuul/src/github.com/vexxhost/atmosphere/.venv/bin/ansible-playbook --inventory /home/zuul/.ansible/tmp/molecule.v9Wo.csi/inventory --skip-tags molecule-notest,notest --inventory=/home/zuul/src/github.com/vexxhost/atmosphere/inventory.yaml /home/zuul/src/github.com/vexxhost/atmosphere/molecule/csi/converge.yml │ ├─38286 ssh -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User="zuul" -o ConnectTimeout=10 -o ControlPath="/home/zuul/.ansible/cp/%h-%p-%r" 199.204.45.138 /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-zfwuspiirmxtkkeaxfpnfswcwmorwogh ; /usr/bin/python3.10'"'"' && sleep 0' │ ├─38392 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-owwekewewrcfhhfofrkvmmvigeexadfg ; /usr/bin/python3'"'"' && sleep 0' │ ├─38393 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-owwekewewrcfhhfofrkvmmvigeexadfg ; /usr/bin/python3' && sleep 0 │ ├─38394 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-owwekewewrcfhhfofrkvmmvigeexadfg ; /usr/bin/python3 │ ├─38395 /bin/sh -c echo BECOME-SUCCESS-owwekewewrcfhhfofrkvmmvigeexadfg ; /usr/bin/python3 │ ├─38396 /usr/bin/python3 │ ├─38397 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─38399 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 42 nofb ├─system.slice │ ├─docker-f056b8f753618926dafb94a39ede835cddad0daa9cada0235dbecb9b51c8eea3.scope … │ │ ├─11530 /sbin/docker-init -- /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ │ └─11542 /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ ├─containerd.service … │ │ ├─ 9103 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─11501 /usr/bin/containerd-shim-runc-v2 -namespace moby -id f056b8f753618926dafb94a39ede835cddad0daa9cada0235dbecb9b51c8eea3 -address /run/containerd/containerd.sock │ │ ├─11867 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 463d985ea12b24231cb9a571abd9f9eeaf7668a0dc8a179825a1265311bca5d8 -address /run/containerd/containerd.sock │ │ ├─18080 /usr/bin/containerd-shim-runc-v2 -namespace moby -id bbfe9296dabc58243435290bed9d5eb9631c5cd4be761626523e817eb8d7ad5b -address /run/containerd/containerd.sock │ │ ├─23285 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 06ca6707c73c51acbdbac5b54e6c64d99079abcd1833a2a09ab7369208a9e08a -address /run/containerd/containerd.sock │ │ ├─27450 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 2e9a0455c26b501d615e5e0c2ad887a317e1a85abc22f8b500d11a30b3c897aa -address /run/containerd/containerd.sock │ │ ├─31883 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 6a42ef172d605fff875960c34184879d8789016aa63bcf0254ddcc92008156ee -address /run/containerd/containerd.sock │ │ ├─36932 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8d1b2ec029c5518eb2ca7104443a4e9c9480d363143dc9b17a0c2b83c104cf91 -address /run/containerd/containerd.sock │ │ ├─36983 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 2ff7df1f225c9fefd7dd3c52038752336e51bb6ddcae90c7c2b33791f70c7f7b -address /run/containerd/containerd.sock │ │ ├─36993 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 312b550d5627aafe7d12d4768eb419c98570231ecba8e1e63fcdc2cb0ef3df3e -address /run/containerd/containerd.sock │ │ ├─37003 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8bc42f769781a9b9573fac3cf98677a55975596fad203b338412ff5083400f6a -address /run/containerd/containerd.sock │ │ ├─37006 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 39f219514ef3ed0445fde5dff7f5cae4a628d9de4b0c17ccd33b480e43b3166e -address /run/containerd/containerd.sock │ │ ├─37516 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 28638bc7ef9ca3cced6bee55ca567d6c0918f391329d90d651feba67d0ed2f37 -address /run/containerd/containerd.sock │ │ ├─38181 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 984803396a153dc77569b3371eebb787b353f60b06e0a689d829877d8c6102d6 -address /run/containerd/containerd.sock │ │ └─38183 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id fb0d08134ab3d5750bc79a0a24e3ba1d2e88610d4a5f788df8b2eb913686a0c3 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1152 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─439 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─329 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─543 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─docker-bbfe9296dabc58243435290bed9d5eb9631c5cd4be761626523e817eb8d7ad5b.scope … │ │ ├─18104 /sbin/docker-init -- /usr/bin/ceph-crash -n client.crash.instance │ │ └─18117 /usr/libexec/platform-python -s /usr/bin/ceph-crash -n client.crash.instance │ ├─system-ceph\x2d4837cbf8\x2d4f90\x2d4300\x2db3f6\x2d726c9b9f89b4.slice │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.1.service │ │ │ ├─27170 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/unit.run │ │ │ └─27433 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-1 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1:/var/lib/ceph/osd/ceph-1:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@crash.instance.service │ │ │ ├─18033 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/unit.run │ │ │ └─18051 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-crash --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-crash-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/config:/etc/ceph/ceph.conf:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/keyring:/etc/ceph/ceph.client.crash.instance.keyring quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -n client.crash.instance │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mgr.instance.iuipuh.service │ │ │ ├─11781 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.iuipuh/unit.run │ │ │ └─11810 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mgr --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mgr-instance-iuipuh --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.1 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.iuipuh:/var/lib/ceph/mgr/ceph-instance.iuipuh:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.iuipuh/config:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v18.2.1 -n mgr.instance.iuipuh -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.0.service │ │ │ ├─23000 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/unit.run │ │ │ └─23268 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-0 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0:/var/lib/ceph/osd/ceph-0:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.2.service │ │ │ ├─31596 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/unit.run │ │ │ └─31866 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-2 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2:/var/lib/ceph/osd/ceph-2:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mon.instance.service │ │ ├─11418 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/unit.run │ │ └─11449 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mon-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.1 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance:/var/lib/ceph/mon/ceph-instance:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev quay.io/ceph/ceph:v18.2.1 -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ ├─docker.service … │ │ └─9471 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock │ ├─polkit.service │ │ └─557 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─534 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─37368 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.138 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─297 /lib/systemd/systemd-journald │ ├─ssh.service │ │ ├─ 559 sshd: /usr/sbin/sshd -D [listener] 3 of 10-100 startups │ │ ├─35145 sshd: [accepted] │ │ ├─38329 sshd: root [priv] │ │ ├─38330 sshd: root [net] │ │ ├─38331 sshd: unknown [priv] │ │ └─38332 sshd: unknown [net] │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1349 /usr/sbin/uuidd --socket-activation │ ├─docker-6a42ef172d605fff875960c34184879d8789016aa63bcf0254ddcc92008156ee.scope … │ │ ├─31907 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─31921 /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─cloud-config.service │ ├─chrony.service │ │ ├─9968 /usr/sbin/chronyd -F 1 │ │ └─9969 /usr/sbin/chronyd -F 1 │ ├─docker-06ca6707c73c51acbdbac5b54e6c64d99079abcd1833a2a09ab7369208a9e08a.scope … │ │ ├─23310 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─23323 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker.socket │ ├─systemd-resolved.service │ │ └─333 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─dbus.service │ │ └─530 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─docker-463d985ea12b24231cb9a571abd9f9eeaf7668a0dc8a179825a1265311bca5d8.scope … │ │ ├─11892 /sbin/docker-init -- /usr/bin/ceph-mgr -n mgr.instance.iuipuh -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─11911 /usr/bin/ceph-mgr -n mgr.instance.iuipuh -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─542 /sbin/agetty -o -p -- \u --noclear tty1 linux │ ├─systemd-logind.service │ │ └─536 /lib/systemd/systemd-logind │ └─docker-2e9a0455c26b501d615e5e0c2ad887a317e1a85abc22f8b500d11a30b3c897aa.scope … │ ├─27474 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ └─27487 /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug ├─k8s.io │ ├─8f42bc4871df58a15a47d739e057b44b7e46285b9cbbd04bf818b94ab7d8ed40 │ │ └─37321 /kube-vip manager │ ├─28638bc7ef9ca3cced6bee55ca567d6c0918f391329d90d651feba67d0ed2f37 │ │ └─37539 /pause │ ├─d85337393252921721f9bfe6976c57cc9cb50df70471716176eb0d5eda9a2546 │ │ └─37194 etcd --advertise-client-urls=https://199.204.45.138:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.138:2380 --initial-cluster=instance=https://199.204.45.138:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.138:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.138:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─a91ba0079e9104450bd05ed55ed84a9b0248775d83167851324d53691279e27e │ │ └─37567 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─8a1a01dc9b4957b6106c8cff2250cd6a6f6c0864684fbd59ccb28ccb9fd01a15 │ │ └─37232 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─39f219514ef3ed0445fde5dff7f5cae4a628d9de4b0c17ccd33b480e43b3166e │ │ └─37063 /pause │ ├─fb0d08134ab3d5750bc79a0a24e3ba1d2e88610d4a5f788df8b2eb913686a0c3 │ │ └─38237 /pause │ ├─8bc42f769781a9b9573fac3cf98677a55975596fad203b338412ff5083400f6a │ │ └─37086 /pause │ ├─b836b1a9defd42bcf44fc3a8559ebc11def8b96bed0f73cb644e927f71f6c635 │ │ └─37219 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─312b550d5627aafe7d12d4768eb419c98570231ecba8e1e63fcdc2cb0ef3df3e │ │ └─37107 /pause │ ├─984803396a153dc77569b3371eebb787b353f60b06e0a689d829877d8c6102d6 │ │ └─38228 /pause │ ├─caddba7e964e06a1ce64555b539dcbc9f9027aac361c233ad46186127d6eb246 │ │ └─37172 kube-apiserver --advertise-address=199.204.45.138 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─2ff7df1f225c9fefd7dd3c52038752336e51bb6ddcae90c7c2b33791f70c7f7b │ │ └─37076 /pause │ └─8d1b2ec029c5518eb2ca7104443a4e9c9480d363143dc9b17a0c2b83c104cf91 │ └─36957 /pause ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount