Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ ├─user-1002.slice │ │ ├─user@1002.service … │ │ │ ├─app.slice │ │ │ └─init.scope │ │ │ ├─13364 /lib/systemd/systemd --user │ │ │ └─13366 (sd-pam) │ │ └─session-21.scope │ │ ├─19583 sshd: cephadm [priv] │ │ └─19608 sshd: cephadm@notty │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─547 /lib/systemd/systemd --user │ │ └─548 (sd-pam) │ ├─session-4.scope │ │ ├─1264 sshd: zuul [priv] │ │ └─1277 sshd: zuul@notty │ └─session-1.scope │ ├─ 544 sshd: zuul [priv] │ ├─ 561 sshd: zuul@notty │ ├─ 692 /usr/bin/python3 │ ├─ 1268 ssh: /home/zuul/.ansible/cp/199.204.45.244-22-zuul [mux] │ ├─59971 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-gxwiiddfwebmlhrvujaarlnpqjunonqv ; /usr/bin/python3'"'"' && sleep 0' │ ├─59972 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-gxwiiddfwebmlhrvujaarlnpqjunonqv ; /usr/bin/python3' && sleep 0 │ ├─59973 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-gxwiiddfwebmlhrvujaarlnpqjunonqv ; /usr/bin/python3 │ ├─59974 /bin/sh -c echo BECOME-SUCCESS-gxwiiddfwebmlhrvujaarlnpqjunonqv ; /usr/bin/python3 │ ├─59975 /usr/bin/python3 │ ├─59976 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─59978 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 41 nofb ├─system.slice │ ├─docker-b1320e62bcf1fdd372c2bfd245fe8f757b4c6b2ad3e73247b903e64eab2f1733.scope … │ │ ├─27257 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─27271 /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─containerd.service … │ │ ├─ 8908 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─11253 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 50456a793dd132d5bb5195f7c7148899ea4b068d0c3257325592321ac0696c6d -address /run/containerd/containerd.sock │ │ ├─11587 /usr/bin/containerd-shim-runc-v2 -namespace moby -id c29a6b967c3d8722727a87f35e0f4a05413ba025c6c95fd3e996d7af9b3b0176 -address /run/containerd/containerd.sock │ │ ├─17792 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 1ae54833dd13e7fe4287fa8af6d2d12f12bb151f58d01cb5679ada3d477d44ce -address /run/containerd/containerd.sock │ │ ├─22915 /usr/bin/containerd-shim-runc-v2 -namespace moby -id abcd01f340735cc41d837f36672d38bb0cef385d478d77d80d2bfbb0ccbefb49 -address /run/containerd/containerd.sock │ │ ├─27233 /usr/bin/containerd-shim-runc-v2 -namespace moby -id b1320e62bcf1fdd372c2bfd245fe8f757b4c6b2ad3e73247b903e64eab2f1733 -address /run/containerd/containerd.sock │ │ ├─31408 /usr/bin/containerd-shim-runc-v2 -namespace moby -id e58510356aacc2414d54f784dc4e25d303d21861864c6ba17b1027a47cd02055 -address /run/containerd/containerd.sock │ │ ├─36533 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9bd4c1c7346bb715d9493e5bb1fd5ffb34ca4069eb5b9105bbf67564974e846c -address /run/containerd/containerd.sock │ │ ├─36545 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a200df5cdac2af9cd345bd857174725e311599c239c4feb5385dc2174af8ff8a -address /run/containerd/containerd.sock │ │ ├─36600 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a7d89d41b1e953e86fbee47efb5ff24f87fbc391e1f346d42ba9a420e4ffcbe7 -address /run/containerd/containerd.sock │ │ ├─36601 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9243bb306cbf8bd470d8afbdd0f0fa23646ae2a6651879186f92b5cf9deb2ace -address /run/containerd/containerd.sock │ │ ├─36632 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id bb60800e72e8989dd37cbad96686952752c9b69b39c18a6d8ceb6a508ef4f65c -address /run/containerd/containerd.sock │ │ ├─37091 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8f5da01e17798b8144fe8170e7fcdf041e9b7d5af13a5df0efb54f1b93f88c34 -address /run/containerd/containerd.sock │ │ ├─37766 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 83c8d89cad0cb3ff302eb026e488ff853f79e46d4cdc7e6b07cba08c9a255b58 -address /run/containerd/containerd.sock │ │ ├─37768 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4d599131c066fc78c4795ea909baa56b4959274df91d4de005b94a80eee8a2f4 -address /run/containerd/containerd.sock │ │ ├─39019 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id bdddfcbea7ae427aaecee100f37207ee9c9d41cd83c2dff4608036479daf40a6 -address /run/containerd/containerd.sock │ │ ├─39069 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1db20c08f91bd9c4ad415dc5395269e1c5f5dcf19774bde5fa35557292728752 -address /run/containerd/containerd.sock │ │ ├─40119 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id fd2fac9ea5806b5a65f24b02c305a44073ad587ed20fca892e0dda33bf3720da -address /run/containerd/containerd.sock │ │ └─40191 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id cb41d4e9afb7f114164124a6c9e7ec7d97c167d3cca67180acd8ae55ea24eea9 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1013 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─363 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ ├─ 267 /lib/systemd/systemd-udevd │ │ ├─59942 /lib/systemd/systemd-udevd │ │ └─59943 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─470 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─system-ceph\x2d4837cbf8\x2d4f90\x2d4300\x2db3f6\x2d726c9b9f89b4.slice │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.1.service │ │ │ ├─26960 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/unit.run │ │ │ └─27215 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-1 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1:/var/lib/ceph/osd/ceph-1:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mgr.instance.cmuobj.service │ │ │ ├─11547 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.cmuobj/unit.run │ │ │ └─11559 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mgr --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mgr-instance-cmuobj --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.1 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.cmuobj:/var/lib/ceph/mgr/ceph-instance.cmuobj:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.cmuobj/config:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v18.2.1 -n mgr.instance.cmuobj -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@crash.instance.service │ │ │ ├─17760 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/unit.run │ │ │ └─17772 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-crash --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-crash-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/config:/etc/ceph/ceph.conf:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/keyring:/etc/ceph/ceph.client.crash.instance.keyring quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -n client.crash.instance │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.0.service │ │ │ ├─22647 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/unit.run │ │ │ └─22899 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-0 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0:/var/lib/ceph/osd/ceph-0:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.2.service │ │ │ ├─31131 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/unit.run │ │ │ └─31390 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-2 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2:/var/lib/ceph/osd/ceph-2:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:9f35728f6070a596500c0804814a12ab6b98e05067316dc64876fb4b28d04af3 -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mon.instance.service │ │ ├─11186 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/unit.run │ │ └─11206 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mon-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.1 -e NODE_NAME=instance -e CEPH_USE_RANDOM_NONCE=1 -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance:/var/lib/ceph/mon/ceph-instance:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev quay.io/ceph/ceph:v18.2.1 -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ ├─docker.service … │ │ └─9267 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock │ ├─polkit.service │ │ └─488 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─459 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─36959 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.244 --pod-infra-container-image=registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─cloud-init-local.service │ ├─systemd-journald.service │ │ └─236 /lib/systemd/systemd-journald │ ├─docker-e58510356aacc2414d54f784dc4e25d303d21861864c6ba17b1027a47cd02055.scope … │ │ ├─31433 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─31444 /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker-c29a6b967c3d8722727a87f35e0f4a05413ba025c6c95fd3e996d7af9b3b0176.scope … │ │ ├─11611 /sbin/docker-init -- /usr/bin/ceph-mgr -n mgr.instance.cmuobj -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─11650 /usr/bin/ceph-mgr -n mgr.instance.cmuobj -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─ssh.service │ │ └─493 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─cloud-final.service │ ├─docker-1ae54833dd13e7fe4287fa8af6d2d12f12bb151f58d01cb5679ada3d477d44ce.scope … │ │ ├─17818 /sbin/docker-init -- /usr/bin/ceph-crash -n client.crash.instance │ │ └─17832 /usr/libexec/platform-python -s /usr/bin/ceph-crash -n client.crash.instance │ ├─cloud-config.service │ ├─chrony.service │ │ ├─9760 /usr/sbin/chronyd -F 1 │ │ └─9761 /usr/sbin/chronyd -F 1 │ ├─docker-abcd01f340735cc41d837f36672d38bb0cef385d478d77d80d2bfbb0ccbefb49.scope … │ │ ├─22940 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─22952 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker.socket │ ├─systemd-resolved.service │ │ └─271 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─cloud-init.service │ ├─dbus.service │ │ └─455 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─469 /sbin/agetty -o -p -- \u --noclear tty1 linux │ ├─docker-50456a793dd132d5bb5195f7c7148899ea4b068d0c3257325592321ac0696c6d.scope … │ │ ├─11288 /sbin/docker-init -- /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ │ └─11301 /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ └─systemd-logind.service │ └─461 /lib/systemd/systemd-logind ├─k8s.io │ ├─8f5da01e17798b8144fe8170e7fcdf041e9b7d5af13a5df0efb54f1b93f88c34 │ │ └─37114 /pause │ ├─fd2fac9ea5806b5a65f24b02c305a44073ad587ed20fca892e0dda33bf3720da │ │ └─40153 /pause │ ├─b33f442e849a4538a9678925ef00930c4cd8b8f10735ac40ae8fb889c5c84ad7 │ │ └─41779 /usr/local/bin/cephcsi --type=controller --v=5 --drivername=rbd.csi.ceph.com --drivernamespace=kube-system --setmetadata=true │ ├─bb60800e72e8989dd37cbad96686952752c9b69b39c18a6d8ceb6a508ef4f65c │ │ └─36662 /pause │ ├─a200df5cdac2af9cd345bd857174725e311599c239c4feb5385dc2174af8ff8a │ │ └─36615 /pause │ ├─4d599131c066fc78c4795ea909baa56b4959274df91d4de005b94a80eee8a2f4 │ │ └─37804 /pause │ ├─83c8d89cad0cb3ff302eb026e488ff853f79e46d4cdc7e6b07cba08c9a255b58 │ │ └─37830 /pause │ ├─22e8942359636ba970e37590b54c20463c0a8c5e30f0d84276352a3cba37bd1e │ │ └─36788 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─b5a2ed13907d831ccb7cb348cc1e00fd6aa43ee364ccf4749fed385d7bdd46d2 │ │ └─41393 /csi-snapshotter --csi-address=unix:///csi/csi-provisioner.sock --v=1 --timeout=60s --leader-election=true --extra-create-metadata=true --enable-volume-group-snapshots=false │ ├─9243bb306cbf8bd470d8afbdd0f0fa23646ae2a6651879186f92b5cf9deb2ace │ │ └─36694 /pause │ ├─1bf73fb7b277e76a4903e2d1e037214b5389f90cae95055af1ed57ee74373fb6 │ │ └─41194 /csi-resizer --v=1 --csi-address=unix:///csi/csi-provisioner.sock --timeout=60s --leader-election --retry-interval-start=500ms --handle-volume-inuse-error=false --feature-gates=RecoverVolumeExpansionFailure=true │ ├─7afe2bd46a4cf565d857a14cd91431a7a58a50b66a01d2da5af32d435fc191d4 │ │ └─40890 /csi-node-driver-registrar --v=5 --csi-address=/csi/csi.sock --kubelet-registration-path=/var/lib/kubelet/plugins/rbd.csi.ceph.com/csi.sock │ ├─b41c4bf48c3605bf6b44cc2d9d4aba968abf04123fadb6588a2294ec7d014180 │ │ └─39193 /coredns -conf /etc/coredns/Corefile │ ├─9bd4c1c7346bb715d9493e5bb1fd5ffb34ca4069eb5b9105bbf67564974e846c │ │ └─36663 /pause │ ├─500ecaa36f54065e3b11d1577d96705a7f17f87b12338c675b34e2477c09bc59 │ │ └─39115 /coredns -conf /etc/coredns/Corefile │ ├─1104465a92151789285214f7af59a09e61d67b18389cc3943ff78ca5399360ae │ │ ├─38592 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─38912 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─1db20c08f91bd9c4ad415dc5395269e1c5f5dcf19774bde5fa35557292728752 │ │ └─39099 /pause │ ├─6307f2e11e2e1d6b872b294232ea0502aa5cadb006c902b61fc15c5ee2ee21a7 │ │ └─36913 /kube-vip manager │ ├─888abeae6aac310da21a231bf6f3c043503cad3f320eabaab147eaebf03607a5 │ │ └─41816 /usr/local/bin/cephcsi --type=liveness --endpoint=unix:///csi/csi-provisioner.sock --metricsport=8080 --metricspath=/metrics --polltime=60s --timeout=3s │ ├─cdace0c7ce864bb610d2f35d2ad30231be62f1eec15023e61e08b88d6455bdd7 │ │ └─41096 /csi-provisioner --csi-address=unix:///csi/csi-provisioner.sock --v=1 --timeout=60s --leader-election=true --retry-interval-start=500ms --default-fstype=ext4 --extra-create-metadata=true --feature-gates=HonorPVReclaimPolicy=true --prevent-volume-mode-conversion=true │ ├─cb41d4e9afb7f114164124a6c9e7ec7d97c167d3cca67180acd8ae55ea24eea9 │ │ └─40214 /pause │ ├─0bdeb1cba951c75baf4000a2a2f850c70f626a57888acdeef29636c20190a899 │ │ └─36798 etcd --advertise-client-urls=https://199.204.45.244:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.244:2380 --initial-cluster=instance=https://199.204.45.244:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.244:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.244:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─bdddfcbea7ae427aaecee100f37207ee9c9d41cd83c2dff4608036479daf40a6 │ │ └─39047 /pause │ ├─35a8fabffe8e476b7d7e17b741b86c182867ffe3774738aca615a24b9932ce42 │ │ └─36856 kube-apiserver --advertise-address=199.204.45.244 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─a7d89d41b1e953e86fbee47efb5ff24f87fbc391e1f346d42ba9a420e4ffcbe7 │ │ └─36708 /pause │ ├─bfe9d9f09b6717e2119ae61662e60528d2937273b5868162b695be06a06cb7bb │ │ └─40924 /usr/local/bin/cephcsi --type=liveness --endpoint=unix:///csi/csi.sock --metricsport=8081 --metricspath=/metrics --polltime=60s --timeout=3s │ ├─fc90db71f8de020e568313dd13830062a31be1402161159550325d841b2163d9 │ │ └─36830 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─b25de3b547575458f5e518956cf84ecc95317d5f9663d653d4a6163d817613aa │ │ └─40785 /usr/local/bin/cephcsi --nodeid=instance --pluginpath=/var/lib/kubelet/plugins --stagingpath=/var/lib/kubelet/plugins/kubernetes.io/csi/ --type=rbd --nodeserver=true --pidlimit=-1 --endpoint=unix:///csi/csi.sock --csi-addons-endpoint=unix:///csi/csi-addons.sock --v=5 --drivername=rbd.csi.ceph.com --enable-read-affinity=false │ ├─dbb0d1070f1f5cda25a90f4ded1c4d33aeeab7c617bd93f75f7da6f905cb0564 │ │ └─41720 /csi-attacher --v=1 --csi-address=unix:///csi/csi-provisioner.sock --leader-election=true --retry-interval-start=500ms --default-fstype=ext4 │ ├─0f95c1a13a21ec436b308f77d182ce8a8f48c005549302835a91b951461e5d72 │ │ └─37975 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─85878b807d88423628527e668ab053e923eda5311b30dffcf3cc5d3f376f99c0 │ │ └─40837 /usr/local/bin/cephcsi --nodeid=instance --type=rbd --controllerserver=true --pidlimit=-1 --endpoint=unix:///csi/csi-provisioner.sock --csi-addons-endpoint=unix:///csi/csi-addons.sock --v=5 --drivername=rbd.csi.ceph.com --rbdhardmaxclonedepth=8 --rbdsoftmaxclonedepth=4 --maxsnapshotsonimage=450 --minsnapshotsonimage=250 --setmetadata=true │ └─a8c45e7d51e169980244306d67f6ea17f5744f0fc3d73c7ad20eb8c20a7efedc │ └─37139 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount