Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ ├─user-1002.slice │ │ ├─user@1002.service … │ │ │ ├─app.slice │ │ │ └─init.scope │ │ │ ├─21180 /lib/systemd/systemd --user │ │ │ └─21181 (sd-pam) │ │ └─session-22.scope │ │ ├─21177 sshd: cephadm [priv] │ │ └─21198 sshd: cephadm@notty │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─704 /lib/systemd/systemd --user │ │ └─705 (sd-pam) │ ├─session-4.scope │ │ ├─1452 sshd: zuul [priv] │ │ └─1465 sshd: zuul@notty │ └─session-1.scope │ ├─ 701 sshd: zuul [priv] │ ├─ 718 sshd: zuul@notty │ ├─ 865 /usr/bin/python3 │ ├─ 1456 ssh: /home/zuul/.ansible/cp/199.204.45.132-22-zuul [mux] │ ├─101282 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-btbscidycurlesabdwfotqxggddrexlp ; /usr/bin/python3'"'"' && sleep 0' │ ├─101283 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-btbscidycurlesabdwfotqxggddrexlp ; /usr/bin/python3' && sleep 0 │ ├─101284 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-btbscidycurlesabdwfotqxggddrexlp ; /usr/bin/python3 │ ├─101285 /bin/sh -c echo BECOME-SUCCESS-btbscidycurlesabdwfotqxggddrexlp ; /usr/bin/python3 │ ├─101286 /usr/bin/python3 │ ├─101287 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─101289 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 42 nofb ├─system.slice │ ├─docker-a92d46bea09684ec2cba19a2804131b1024ae1ffe3bdc0ad1c96aea0d63f9b0d.scope … │ │ ├─19164 /sbin/docker-init -- /usr/bin/ceph-crash -n client.crash.instance │ │ └─19178 /usr/bin/python3 -s /usr/bin/ceph-crash -n client.crash.instance │ ├─containerd.service … │ │ ├─ 9795 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─12190 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 756db9fe45e896a49d505988b0a0d3e3b5a1e1720f18f7bfb1a6bb7b8312e3e7 -address /run/containerd/containerd.sock │ │ ├─12549 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 17e02dc1d072f2455dac228d02884aec32b179f4d57bf74e7c962279dab6eb7a -address /run/containerd/containerd.sock │ │ ├─19139 /usr/bin/containerd-shim-runc-v2 -namespace moby -id a92d46bea09684ec2cba19a2804131b1024ae1ffe3bdc0ad1c96aea0d63f9b0d -address /run/containerd/containerd.sock │ │ ├─24317 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 4b45b308784497870794992b3e16769aafd7b0da569505bdce28fd9c1b50264c -address /run/containerd/containerd.sock │ │ ├─28743 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 7f8e7386e0bb962c53673a02156e50a819f6c3223b924f62b7b8dc6265e73793 -address /run/containerd/containerd.sock │ │ ├─33408 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 68a04577b9bb5ca27f28c4f4d47fc7026e320163452efb75007f8cd714e54643 -address /run/containerd/containerd.sock │ │ ├─39058 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b6189ef31018bf333620121a617df8a029279c46853ce0f235d6c27ed6bd6b23 -address /run/containerd/containerd.sock │ │ ├─39102 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4e0f4f04d2e5e198a70fdb66fa227afa977f0fb354374b21a19a2d6ac7b1174e -address /run/containerd/containerd.sock │ │ ├─39148 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 89f7925f8b4a69e36996a05c922c79e530a33df9bebe073a5839588784129d3a -address /run/containerd/containerd.sock │ │ ├─39152 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7838d94f7adaaf89fbdd3035147344adbee89e1c4db8db6954a26ae1accb27f2 -address /run/containerd/containerd.sock │ │ ├─39235 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8d4dc96326837a54ca3fe3e159d5a96b54d6c70da945f1e9e1703a817ce56f3f -address /run/containerd/containerd.sock │ │ ├─39684 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8f72a3fbd908fd0175bc8f8c8e43d784ce048cf94d72183e26ae36591d1b7224 -address /run/containerd/containerd.sock │ │ ├─40413 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 99f22971e8a89fb4693f734f03293ac7f5d9d15a1bdde9bf9e008f9f036da972 -address /run/containerd/containerd.sock │ │ ├─40415 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 28d0c39082672409614b1a431e9aa9386d3582b5d92467762eec32d78d06d08f -address /run/containerd/containerd.sock │ │ ├─41754 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 550eb4183aab8e8b0806f814aef6de799bec3873b4381cf9d5eb44d307a21a14 -address /run/containerd/containerd.sock │ │ ├─42159 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b38d6ff2ab55084119365230f16111c03d6102aa286f052fe19558a6f5f92ff4 -address /run/containerd/containerd.sock │ │ ├─42193 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 99fed808629ddd136c08fe3909df987f93ac5ac6a404e8c38a3db94ed3ae58f7 -address /run/containerd/containerd.sock │ │ ├─42630 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9e33e81d7534f5eefffeb9610d0c0b2a380b068682123256b24dca1b8ae27776 -address /run/containerd/containerd.sock │ │ ├─43229 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f3e5bcea2790e883bee080d724a98374201c5d7d45d51559224dec49cbe13b77 -address /run/containerd/containerd.sock │ │ ├─43267 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ce861a13c5d12b51de1167b5404fa13f1a004562341314c0d0eba00ce2d72a6b -address /run/containerd/containerd.sock │ │ ├─43309 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 19ad9358aa3fbab61931e62948342417bb52a754786e2aea4e92d0d749ba78d9 -address /run/containerd/containerd.sock │ │ ├─45323 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id afdcf95d6e679c135b27ada4724d3aa0bca4a5dd17697462a4f8b5e997d58b2d -address /run/containerd/containerd.sock │ │ ├─45433 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 58048116e16c3e774af8484547187c9cb7a44652b1f9cc27fd358e45c4525de4 -address /run/containerd/containerd.sock │ │ ├─46399 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id cc35cff9231b42cdbd5e7d230043dc4f871e464b4619694597bd734daa5aba3a -address /run/containerd/containerd.sock │ │ ├─46512 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 03fa99779f4bf0a40cd107cd79f875a6fab185c9c1e4b0e9c349093b1373ea27 -address /run/containerd/containerd.sock │ │ ├─46902 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e395e79d442ce7dc731027d1e9fd0259d5af420046cd601723dab10e152e2888 -address /run/containerd/containerd.sock │ │ ├─47232 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 232c4063a11e3173d77f3da433497dfb10ac7969091e779437b3effafe431e16 -address /run/containerd/containerd.sock │ │ ├─47616 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5f8ac727dfbd3d2453d5145ce5136f43300bc85bdd7a1a0e198751a34c748c73 -address /run/containerd/containerd.sock │ │ ├─49212 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0348af833ca9c6649cab58da1e69126191d3d7512931fb1997165f4f64181a79 -address /run/containerd/containerd.sock │ │ ├─51316 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6fecd624960a676239f39663d321d8d25cec9af23bc8f94116bc29b04f510b87 -address /run/containerd/containerd.sock │ │ ├─51882 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6a32f7b9a9cc1dff43feccaf65aa7eff3dfaa4f74d20d5f7386678084f384e2e -address /run/containerd/containerd.sock │ │ ├─51930 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b9a83e37691e345cd84c39355a3b15e74fad511f485304411221d670ac5f38d6 -address /run/containerd/containerd.sock │ │ ├─52080 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e0d281ffa001895cc2e76334f26d61bc91d5dab32e5a957ab00fb16e3f2617ed -address /run/containerd/containerd.sock │ │ ├─55039 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b88c93218a2a79df724d53fe8e2dd1a36dd2104bbec0fb3057018ea057c160e7 -address /run/containerd/containerd.sock │ │ ├─55175 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 73102bb945f23f8e9e79ba2e584e1f5debe8bbba7d14023b063eecb8c75e874e -address /run/containerd/containerd.sock │ │ ├─55216 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6c14c90e1fd194a8d3434fedbab25cd1fadb110d326503a5d08d50353cad90e1 -address /run/containerd/containerd.sock │ │ ├─55264 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 43b54bc54b48126952dfd28165b577f7bfe0d7a7f74f393d139706974302a9b0 -address /run/containerd/containerd.sock │ │ ├─56545 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e72376e5cb5e52480bbdfa1cdc6398d2b3b4af07d2cdb916dcc8de0f84abee4f -address /run/containerd/containerd.sock │ │ ├─56879 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 158c63bd939167c0b9c160acd52da5b8ad913ced90a93dc1744344857ce49b27 -address /run/containerd/containerd.sock │ │ ├─56963 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5b661b3f1b5d15162762d09388c99e5de7a21c0378c23a2519b21bb00714a962 -address /run/containerd/containerd.sock │ │ ├─57182 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a579c3fbd3130cb2c3f96e7fe9261777e15d05b46f57d85b1e2f208f35cb2a4e -address /run/containerd/containerd.sock │ │ ├─57462 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0d1d5e34e3245d448bb614480a9689bc42ac6e6074a6e43b01483c5573d2afa5 -address /run/containerd/containerd.sock │ │ ├─57593 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1b0e647d956443d3ff609fb56a4929957ac72dc791a3ba295c6f1c0a8c8117be -address /run/containerd/containerd.sock │ │ ├─57944 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4c64a65a332bf550324fe3324af92f9dd2bf1f136a54506b27e1727b22d2df98 -address /run/containerd/containerd.sock │ │ ├─58772 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b7f358242df675c425d61e7e9af35a4cbe0c2c567591bab0eed57256c113fd85 -address /run/containerd/containerd.sock │ │ ├─60039 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5dcc53f3cfeaa6c374c6bc1c7dac16c3ac328ebfd4bf683f5cd3e894132eb4f7 -address /run/containerd/containerd.sock │ │ ├─61722 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 28b202ebad9e549a597f35c60fed894ef7a9683d0c728afde6b30d7a4e618667 -address /run/containerd/containerd.sock │ │ ├─68919 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5910dc85c7672bb20ab7f457e85a44fdab647c00612ae476a41cc60be76d6289 -address /run/containerd/containerd.sock │ │ ├─70060 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a8d7c201c771b4fe12f9a35824ffc8ec4db7701ad7212a53bc485217df2d86fb -address /run/containerd/containerd.sock │ │ ├─74596 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b6cbfd8a5a04f122e793d94c812101eb87c566a233b7e1470267c00ffbb6877f -address /run/containerd/containerd.sock │ │ ├─79239 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 62043a2cffed41fff85fa2b8ba1d79ce10892f44bc1c19e6925b6655419701e5 -address /run/containerd/containerd.sock │ │ ├─87143 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 321a8cf1df274d5b411d1465d2ee88a7a2a692d3d4b966add1ee542b590a3275 -address /run/containerd/containerd.sock │ │ ├─87182 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id cc0d43de60e822aef20d118958a0591c5a456e01ad896f17796b182bfba37a1e -address /run/containerd/containerd.sock │ │ ├─91039 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9238600412d32961f953609c2f6b261a62ff27c388ea426c59439f850a863fa7 -address /run/containerd/containerd.sock │ │ ├─92224 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4714267e342e7d83b99f502cd9bf9efb06aab83847b50b2e747158c1e53d2395 -address /run/containerd/containerd.sock │ │ ├─92277 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 199e09fd4513d5b1ea7b78924929880e171afabf429b409d3fcec2870fd2a40d -address /run/containerd/containerd.sock │ │ └─93076 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a36e4fa3ca2b80e9fa2276138ef8fdd6200edd873f4e199aa0f505bc6b31789f -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1188 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─522 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─393 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─624 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─system-ceph\x2d4837cbf8\x2d4f90\x2d4300\x2db3f6\x2d726c9b9f89b4.slice │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.1.service │ │ │ ├─28457 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/unit.run │ │ │ └─28725 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-1 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1:/var/lib/ceph/osd/ceph-1:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@crash.instance.service │ │ │ ├─19094 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/unit.run │ │ │ └─19113 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-crash --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-crash-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/config:/etc/ceph/ceph.conf:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/keyring:/etc/ceph/ceph.client.crash.instance.keyring quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n client.crash.instance │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.0.service │ │ │ ├─24021 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/unit.run │ │ │ └─24300 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-0 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0:/var/lib/ceph/osd/ceph-0:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.2.service │ │ │ ├─33117 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/unit.run │ │ │ └─33392 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-2 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2:/var/lib/ceph/osd/ceph-2:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mon.instance.service │ │ │ ├─12102 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/unit.run │ │ │ └─12133 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mon-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance:/var/lib/ceph/mon/ceph-instance:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev quay.io/ceph/ceph:v18.2.7 -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ │ └─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mgr.instance.rfyibd.service │ │ ├─12465 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.rfyibd/unit.run │ │ └─12494 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mgr --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mgr-instance-rfyibd --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.rfyibd:/var/lib/ceph/mgr/ceph-instance.rfyibd:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.rfyibd/config:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v18.2.7 -n mgr.instance.rfyibd -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker.service … │ │ └─10166 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock │ ├─polkit.service │ │ └─643 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─617 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─39523 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.132 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─358 /lib/systemd/systemd-journald │ ├─docker-17e02dc1d072f2455dac228d02884aec32b179f4d57bf74e7c962279dab6eb7a.scope … │ │ ├─12574 /sbin/docker-init -- /usr/bin/ceph-mgr -n mgr.instance.rfyibd -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─12593 /usr/bin/ceph-mgr -n mgr.instance.rfyibd -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─ssh.service │ │ ├─ 651 sshd: /usr/sbin/sshd -D [listener] 1 of 10-100 startups │ │ └─100487 sshd: [accepted] │ ├─cloud-final.service │ ├─docker-756db9fe45e896a49d505988b0a0d3e3b5a1e1720f18f7bfb1a6bb7b8312e3e7.scope … │ │ ├─12214 /sbin/docker-init -- /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ │ └─12234 /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ ├─uuidd.service │ │ └─1387 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─docker-7f8e7386e0bb962c53673a02156e50a819f6c3223b924f62b7b8dc6265e73793.scope … │ │ ├─28766 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─28779 /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─chrony.service │ │ ├─10673 /usr/sbin/chronyd -F 1 │ │ └─10674 /usr/sbin/chronyd -F 1 │ ├─docker.socket │ ├─systemd-resolved.service │ │ └─397 /lib/systemd/systemd-resolved │ ├─docker-68a04577b9bb5ca27f28c4f4d47fc7026e320163452efb75007f8cd714e54643.scope … │ │ ├─33434 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─33446 /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─system-lvm2\x2dpvscan.slice │ ├─dbus.service │ │ └─614 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─623 /sbin/agetty -o -p -- \u --noclear tty1 linux │ ├─docker-4b45b308784497870794992b3e16769aafd7b0da569505bdce28fd9c1b50264c.scope … │ │ ├─24340 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─24353 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ └─systemd-logind.service │ └─620 /lib/systemd/systemd-logind ├─k8s.io │ ├─e7b51b1eb213f22577cb2cf391889ad2cc3d0062114fead385e6eaa048569387 │ │ └─39349 etcd --advertise-client-urls=https://199.204.45.132:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.132:2380 --initial-cluster=instance=https://199.204.45.132:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.132:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.132:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─3a7f0abf08df50f92cf3bb3cb4a6ed4bd65d57a8575e06c9ec231dea7777cff2 │ │ └─39318 kube-apiserver --advertise-address=199.204.45.132 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─4400ccf9136d6b278747b19c7a70b42485d406e1765f12b735991b3f6d6cc1cd │ │ ├─64345 apache2 -DFOREGROUND │ │ ├─64363 (wsgi:k -DFOREGROUND │ │ ├─64364 (wsgi:k -DFOREGROUND │ │ ├─64365 (wsgi:k -DFOREGROUND │ │ ├─64366 (wsgi:k -DFOREGROUND │ │ ├─64367 apache2 -DFOREGROUND │ │ ├─64369 apache2 -DFOREGROUND │ │ ├─64370 apache2 -DFOREGROUND │ │ ├─64373 apache2 -DFOREGROUND │ │ ├─64374 apache2 -DFOREGROUND │ │ ├─64378 apache2 -DFOREGROUND │ │ ├─64379 apache2 -DFOREGROUND │ │ ├─64380 apache2 -DFOREGROUND │ │ ├─64381 apache2 -DFOREGROUND │ │ └─64382 apache2 -DFOREGROUND │ ├─e987261044f77fa369e6e4e550a94894a475b1a2ba9911aacf5457fdd6d6e0ee │ │ ├─45678 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─45692 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─45798 nginx: master process /usr/bin/nginx -c /etc/nginx/nginx.conf │ │ ├─90523 nginx: worker process │ │ ├─90524 nginx: worker process │ │ └─90525 nginx: cache manager process │ ├─321a8cf1df274d5b411d1465d2ee88a7a2a692d3d4b966add1ee542b590a3275 │ │ └─87168 /pause │ ├─a579c3fbd3130cb2c3f96e7fe9261777e15d05b46f57d85b1e2f208f35cb2a4e │ │ └─57209 /pause │ ├─7838d94f7adaaf89fbdd3035147344adbee89e1c4db8db6954a26ae1accb27f2 │ │ └─39201 /pause │ ├─cf2a0129e45f0f7ff036b0b95daa6b955652d952553107cd5419935760f71824 │ │ ├─47458 haproxy -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ │ └─49701 haproxy -sf 263 -x sockpair@4 -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ ├─cc35cff9231b42cdbd5e7d230043dc4f871e464b4619694597bd734daa5aba3a │ │ └─46421 /pause │ ├─7e12df5972e3a9fb67c0b5388d856c657f30af28cffdf7cb61efbb28cb716da3 │ │ └─59007 /goldpinger --static-file-path /static │ ├─199e09fd4513d5b1ea7b78924929880e171afabf429b409d3fcec2870fd2a40d │ │ └─92307 /pause │ ├─13497532eb7ea196ce64b87ac8f2f3d84e3509608bef172ef75b00f5a501baca │ │ └─51499 /kubernetes-entrypoint │ ├─988eda3bc13c7e23d0fbe3f7879342569124651bf6ba2501ad85ebfe8c337f33 │ │ └─39330 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─62043a2cffed41fff85fa2b8ba1d79ce10892f44bc1c19e6925b6655419701e5 │ │ └─79262 /pause │ ├─23b3a0ffefa8cfe8bfc96400f70fd4787fae9c33db98c104fc5420ddab94fe46 │ │ └─43494 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.10 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 │ ├─a36e4fa3ca2b80e9fa2276138ef8fdd6200edd873f4e199aa0f505bc6b31789f │ │ └─93099 /pause │ ├─0d1d5e34e3245d448bb614480a9689bc42ac6e6074a6e43b01483c5573d2afa5 │ │ └─57487 /pause │ ├─99f22971e8a89fb4693f734f03293ac7f5d9d15a1bdde9bf9e008f9f036da972 │ │ └─40461 /pause │ ├─0ac24968cdc598bb6cecbb2ed2a5b4278e731bb0d68938efe7109849a6383a3e │ │ └─43543 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager │ ├─e395e79d442ce7dc731027d1e9fd0259d5af420046cd601723dab10e152e2888 │ │ └─46925 /pause │ ├─6fecd624960a676239f39663d321d8d25cec9af23bc8f94116bc29b04f510b87 │ │ └─51340 /pause │ ├─e0d281ffa001895cc2e76334f26d61bc91d5dab32e5a957ab00fb16e3f2617ed │ │ └─52105 /pause │ ├─d15ae2c2fcfcc2ec9e398ac99d4b8adf9c14ad2b63cb03b315167aa0f76fa26d │ │ └─75379 /usr/local/bin/rook ceph operator │ ├─a8d7c201c771b4fe12f9a35824ffc8ec4db7701ad7212a53bc485217df2d86fb │ │ └─70083 /pause │ ├─2758e4cc411284c507b4924d48e0810cdeb7080c0b458af7abc4dbf8f6dc8573 │ │ └─45718 /server │ ├─56b4f5cfe1551cb7a81b550677c7a50598e3969cd1661e41e91787b9ea0af456 │ │ └─59982 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.199-204-45-132.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml │ ├─f7e874cef995462b71008ccd0e891b13252a344c6ae5e009a58b34d73b024095 │ │ └─41818 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─5b661b3f1b5d15162762d09388c99e5de7a21c0378c23a2519b21bb00714a962 │ │ └─57012 /pause │ ├─78497fb390999f2c79bac42c326dc7c7949b558362f2be3e312f0b89cc9fb0ba │ │ ├─92735 staffeln-conductor: master process [/var/lib/openstack/bin/staffeln-conductor --config-file /etc/staffeln/staffeln.conf] │ │ ├─92916 staffeln-conductor: Staffeln conductor backup controller worker(0) │ │ └─92918 staffeln-conductor: Staffeln conductor rotation controller worker(0) │ ├─67fcc53b82282dad8952fc610d8d6eb1ca0ec124c3dffb0c0d308e3d14dcf438 │ │ └─39465 /kube-vip manager │ ├─6c14c90e1fd194a8d3434fedbab25cd1fadb110d326503a5d08d50353cad90e1 │ │ └─55243 /pause │ ├─b38d6ff2ab55084119365230f16111c03d6102aa286f052fe19558a6f5f92ff4 │ │ └─42188 /pause │ ├─6a4256b3cc5a29fd94064a160a5ca6e92d7069815c18962c7b00fe4e0599371a │ │ ├─55681 python -u /app/sidecar.py │ │ ├─55858 python -u /app/sidecar.py │ │ └─55859 python -u /app/sidecar.py │ ├─e7f4140bd9742af78456d14e4d5696dbe9228e989bf35ea68f67ead768c3bad8 │ │ └─47825 /bin/mysqld_exporter --collect.info_schema.processlist │ ├─bf80e97aaa0ce302d4aacd6ed5f8f0253e93a5b0313876caadaadc4fa390580a │ │ ├─79343 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─79360 /usr/local/lib/erlang/erts-12.3.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /usr/local/lib/erlang -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false │ │ ├─79367 erl_child_setup 1048576 │ │ ├─79447 /usr/local/lib/erlang/erts-12.3.2/bin/epmd -daemon │ │ ├─79500 inet_gethost 4 │ │ └─79501 inet_gethost 4 │ ├─b6189ef31018bf333620121a617df8a029279c46853ce0f235d6c27ed6bd6b23 │ │ └─39081 /pause │ ├─149036ff82b35a1eab92447d8f8fc352537d07741c39b5baae67ecdfa6adbe71 │ │ └─60207 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config │ ├─9e33e81d7534f5eefffeb9610d0c0b2a380b068682123256b24dca1b8ae27776 │ │ └─42654 /pause │ ├─b68b3f6f2c8984662003fee2ed4a29a77b294d93fdb05dd4ebf514c7dea4f65f │ │ └─60495 /bin/oauth2-proxy │ ├─c215f00269c7f2e271ed914c82a8125f55d88fc2ab070b2823bd19f513fc8111 │ │ └─58141 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml │ ├─ce90b5329baf0fd0c5ed1b0693867ba2164b8fe70083c77d2ff5db78ec348fda │ │ ├─57993 nginx: master process nginx -g daemon off; │ │ ├─58014 nginx: worker process │ │ ├─58015 nginx: worker process │ │ ├─58016 nginx: worker process │ │ ├─58017 nginx: worker process │ │ └─58018 nginx: worker process │ ├─ad00814d81ff5e9db051cca0b789c69d46de6e3ea3bcb357e8657ef7d8b9d66f │ │ └─39733 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─f313b9588c5ab190f00aa860873aa1267ab549838458cc5e94c550453680b69d │ │ └─47039 percona-xtradb-cluster-operator │ ├─19ad9358aa3fbab61931e62948342417bb52a754786e2aea4e92d0d749ba78d9 │ │ └─43338 /pause │ ├─502390dd835817601867acc1e5cfe0a88b59dea436d99362a6e1cc1817f10dbb │ │ └─47504 /usr/bin/peer-list -on-change=/usr/bin/add_pxc_nodes.sh -service=percona-xtradb-pxc │ ├─b6cbfd8a5a04f122e793d94c812101eb87c566a233b7e1470267c00ffbb6877f │ │ └─74620 /pause │ ├─8f72a3fbd908fd0175bc8f8c8e43d784ce048cf94d72183e26ae36591d1b7224 │ │ └─39707 /pause │ ├─b051d6b8b9c38f1680f0b66089974f21804a3670fc671955bf61f34a20d2ea92 │ │ └─46711 /manager --metrics-bind-address=:8080 │ ├─4f3109196bdb9970086ff58cad7b7d348ff0fc357c815fc2a5cac010fd5cfe6a │ │ └─60476 /bin/oauth2-proxy │ ├─b88c93218a2a79df724d53fe8e2dd1a36dd2104bbec0fb3057018ea057c160e7 │ │ └─55109 /pause │ ├─9b7eea061659764d8c7246a0462d6f2896ba1b8c876b974617c9689b0f5d736e │ │ └─42911 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─28d0c39082672409614b1a431e9aa9386d3582b5d92467762eec32d78d06d08f │ │ └─40463 /pause │ ├─5910dc85c7672bb20ab7f457e85a44fdab647c00612ae476a41cc60be76d6289 │ │ └─68943 /pause │ ├─281b3ffc26de1f680fc8c3b46172e04a45abd40e181f7b4cde3283dacc50e753 │ │ └─40707 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─99fed808629ddd136c08fe3909df987f93ac5ac6a404e8c38a3db94ed3ae58f7 │ │ └─42219 /pause │ ├─8f9f20d2adee1da4800c87fc019a9c70888b4f321b5aea1b42d30553fd13cedc │ │ └─60164 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.199-204-45-132.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml │ ├─59212c6bcd5995ecf0806b01e18e451f973bdc7a3209d1bea1cbd64cdef627ed │ │ └─46587 /manager --metrics-bind-address=:9782 │ ├─550eb4183aab8e8b0806f814aef6de799bec3873b4381cf9d5eb44d307a21a14 │ │ └─41777 /pause │ ├─eb797e19bc3cb9a3d4deaeaff02a09051205084c828ac2153076783ca82ee4f9 │ │ └─52157 nfd-worker --server=node-feature-discovery-master:8080 │ ├─28b202ebad9e549a597f35c60fed894ef7a9683d0c728afde6b30d7a4e618667 │ │ └─61746 /pause │ ├─e0803bc492f81d36405d1ab914456ca5c8db54118c74d7a867daeae6170127f8 │ │ ├─69025 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─69049 /usr/local/lib/erlang/erts-12.3.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /usr/local/lib/erlang -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false │ │ ├─69056 erl_child_setup 1048576 │ │ ├─69110 /usr/local/lib/erlang/erts-12.3.2/bin/epmd -daemon │ │ ├─69161 inet_gethost 4 │ │ └─69162 inet_gethost 4 │ ├─9238600412d32961f953609c2f6b261a62ff27c388ea426c59439f850a863fa7 │ │ └─91061 /pause │ ├─3d7560beeaa992ffa2dfe478785a216c11da792f6b38bfdc34e3b7dafdde1dfa │ │ └─56033 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments │ ├─1272158cbaac7e9b81f53e660227c6b65032a46506086d8383b7cb6c06a9999c │ │ ├─55705 python -u /app/sidecar.py │ │ ├─55971 python -u /app/sidecar.py │ │ └─55972 python -u /app/sidecar.py │ ├─6b44f4b56b6b711aeda18dcd103e96a9d0116519db9af657e073e69fd9dc6341 │ │ └─57669 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning │ ├─158c63bd939167c0b9c160acd52da5b8ad913ced90a93dc1744344857ce49b27 │ │ └─56914 /pause │ ├─1b0e647d956443d3ff609fb56a4929957ac72dc791a3ba295c6f1c0a8c8117be │ │ └─57618 /pause │ ├─03fa99779f4bf0a40cd107cd79f875a6fab185c9c1e4b0e9c349093b1373ea27 │ │ └─46536 /pause │ ├─89f7925f8b4a69e36996a05c922c79e530a33df9bebe073a5839588784129d3a │ │ └─39207 /pause │ ├─e0189bf64e6e84aedc900cdceeb177926755124922da918f8da66d3c174cc24c │ │ └─59245 /usr/bin/loki -config.file=/etc/loki/config/config.yaml -target=all │ ├─3f8e75b7dbb972303a23d4c8d958b467e13dac98eb3429f816ba0ec03221d1c3 │ │ ├─91415 glance-api: uWSGI master │ │ ├─91428 glance-api: uWSGI worker 1 │ │ └─91429 glance-api: uWSGI worker 2 │ ├─43b54bc54b48126952dfd28165b577f7bfe0d7a7f74f393d139706974302a9b0 │ │ └─55296 /pause │ ├─f3e5bcea2790e883bee080d724a98374201c5d7d45d51559224dec49cbe13b77 │ │ └─43252 /pause │ ├─b7f358242df675c425d61e7e9af35a4cbe0c2c567591bab0eed57256c113fd85 │ │ └─58796 /pause │ ├─62bb9d90e3294cba372e23b6ed19004c792e347ae68275a6ebc9929eefef85ba │ │ └─52190 nfd-master -featurerules-controller=true │ ├─58db95aad142466b70ae135eed005cb3a881ad0f558f3b172d98590e41ab9dc0 │ │ └─87596 /usr/libexec/platform-python -s /usr/bin/ceph-crash │ ├─6a32f7b9a9cc1dff43feccaf65aa7eff3dfaa4f74d20d5f7386678084f384e2e │ │ └─51910 /pause │ ├─e8a798d3b55715207a38539f788ff37412944ddcc0f6963bcc58b50a4c71990a │ │ └─92442 kubernetes-entrypoint │ ├─cead3bbc953a4d4b9a8b7595e71d9438981e0ff341eaecd5c41c1e9d99e5f4ef │ │ └─47728 mysqld --wsrep_start_position=a1d2605b-3246-11f1-9af3-737f2805815e:22 │ ├─5dcc53f3cfeaa6c374c6bc1c7dac16c3ac328ebfd4bf683f5cd3e894132eb4f7 │ │ └─60062 /pause │ ├─f7cfc6e029cbf0a09d8d76cab982da9f5bbdd05f79113724858dfd74a260571d │ │ └─55817 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.34.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 │ ├─4c64a65a332bf550324fe3324af92f9dd2bf1f136a54506b27e1727b22d2df98 │ │ └─57967 /pause │ ├─7b62ac4ee443bcb5296b1d4d0cfcd6267066b40f82550253435e5802a252a55a │ │ ├─40924 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─41305 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─58048116e16c3e774af8484547187c9cb7a44652b1f9cc27fd358e45c4525de4 │ │ └─45457 /pause │ ├─50874e21c9494f5beb36c0264a1e8577cfe3cd9f28b37eac63444f6fdd77094d │ │ └─59454 /bin/pushgateway │ ├─ff169c01d7183b6bc3238249e870a478ee319b1040671363cf47144d06c4be28 │ │ └─43454 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true │ ├─ce861a13c5d12b51de1167b5404fa13f1a004562341314c0d0eba00ce2d72a6b │ │ └─43294 /pause │ ├─afdcf95d6e679c135b27ada4724d3aa0bca4a5dd17697462a4f8b5e997d58b2d │ │ └─45355 /pause │ ├─91bb75d9ba70e7831c45af95d3114e20d59d18303dcbcafbc5a6a0975b8a4f73 │ │ ├─74088 barbiacan-api: uWSGI master │ │ └─74103 barbiacan-api: uWSGI worker 1 │ ├─a91aa82fef77a2fff2844539d8640a2bb485b6ee931cdc1d1536bcc2d6a0721d │ │ ├─87445 /bin/bash -x -e -m -c CEPH_CLIENT_ID=ceph-client.rgw.ceph.a PERIODICITY=daily LOG_ROTATE_CEPH_FILE=/etc/logrotate.d/ceph LOG_MAX_SIZE=500M # edit the logrotate file to only rotate a specific daemon log # otherwise we will logrotate log files without reloading certain daemons # this might happen when multiple daemons run on the same machine sed -i "s|*.log|$CEPH_CLIENT_ID.log|" "$LOG_ROTATE_CEPH_FILE" # replace default daily with given user input sed --in-place "s/daily/$PERIODICITY/g" "$LOG_ROTATE_CEPH_FILE" if [ "$LOG_MAX_SIZE" != "0" ]; then # adding maxsize $LOG_MAX_SIZE at the 4th line of the logrotate config file with 4 spaces to maintain indentation sed --in-place "4i \ \ \ \ maxsize $LOG_MAX_SIZE" "$LOG_ROTATE_CEPH_FILE" fi while true; do # we don't force the logrorate but we let the logrotate binary handle the rotation based on user's input for periodicity and size logrotate --verbose "$LOG_ROTATE_CEPH_FILE" sleep 15m done │ │ └─87494 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 15m │ ├─5f8ac727dfbd3d2453d5145ce5136f43300bc85bdd7a1a0e198751a34c748c73 │ │ └─47639 /pause │ ├─c1bba7714be1d0586db0e2da524ecbc2deea47eef154b37a8a3267b42b47bad4 │ │ └─60285 memcached -v -p 11211 -U 0 -c 8192 -m 1024 │ ├─e72376e5cb5e52480bbdfa1cdc6398d2b3b4af07d2cdb916dcc8de0f84abee4f │ │ └─56572 /pause │ ├─b9a83e37691e345cd84c39355a3b15e74fad511f485304411221d670ac5f38d6 │ │ └─51954 /pause │ ├─73102bb945f23f8e9e79ba2e584e1f5debe8bbba7d14023b063eecb8c75e874e │ │ └─55201 /pause │ ├─0348af833ca9c6649cab58da1e69126191d3d7512931fb1997165f4f64181a79 │ │ └─49234 /pause │ ├─5491e288f867034784c2c72f07a086a5278c3350a78e1728e5d56cf9ff394fd5 │ │ └─60703 /bin/memcached_exporter │ ├─167f2532251dc63c8d8c83e3259728d02b343ec43653954c3b362192c6c7b3a3 │ │ └─87367 radosgw --fsid=4837cbf8-4f90-4300-b3f6-726c9b9f89b4 --keyring=/etc/ceph/keyring-store/keyring --log-to-stderr=true --err-to-stderr=true --mon-cluster-log-to-stderr=true --log-stderr-prefix=debug --default-log-to-file=false --default-mon-cluster-log-to-file=false --mon-host=[v2:10.96.240.200:3300,v1:10.96.240.200:6789] --mon-initial-members=instance --id=rgw.ceph.a --setuser=ceph --setgroup=ceph --foreground --rgw-frontends=beast port=8080 --host=rook-ceph-rgw-ceph-a-67fd8975c6-5bsb8 --rgw-mime-types-file=/etc/ceph/rgw/mime.types --rgw-realm=ceph --rgw-zonegroup=ceph --rgw-zone=ceph │ ├─4e0f4f04d2e5e198a70fdb66fa227afa977f0fb354374b21a19a2d6ac7b1174e │ │ └─39126 /pause │ ├─83f97d32593394b1a00ccbf12ee01a66539277e8cb2f57c84f9bc473037e92e4 │ │ └─39316 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─d9b6288c6ba7fd57444c9db2e665cfc68fd7b33505e555d9d3990fc6b10282dc │ │ └─49356 java -Dkc.config.built=true -Xms64m -Xmx512m -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --auto-build --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true │ ├─f3bfba2dd35e93067961f10e2b1080fad40ce009a3da17b0862da3f170c4e6fe │ │ ├─60520 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─60538 /usr/local/lib/erlang/erts-12.3.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /usr/local/lib/erlang -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false │ │ ├─60545 erl_child_setup 1048576 │ │ ├─60638 /usr/local/lib/erlang/erts-12.3.2/bin/epmd -daemon │ │ ├─60721 inet_gethost 4 │ │ └─60722 inet_gethost 4 │ ├─c9aff580578ef9b08702ae1d523fc3bd8c0ecb220bab9fcaa4f339ba5aaace15 │ │ └─42251 /coredns -conf /etc/coredns/Corefile │ ├─8b5f3674092199dccb4eff79b04d66fe7eda4fca25c8ba05bd7c5ac47bd832a0 │ │ └─52338 /secretgen-controller │ ├─cc0d43de60e822aef20d118958a0591c5a456e01ad896f17796b182bfba37a1e │ │ └─87205 /pause │ ├─8d4dc96326837a54ca3fe3e159d5a96b54d6c70da945f1e9e1703a817ce56f3f │ │ └─39262 /pause │ ├─232c4063a11e3173d77f3da433497dfb10ac7969091e779437b3effafe431e16 │ │ └─47256 /pause │ ├─967cb40cdf4cab2240780021ee5031ba2ff8720dab5a9f10191cce53b09ad44b │ │ └─60114 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─4f7ac6fde29b6dce2a5e73bdbd596e897bb9cfafc22ac77d50145389f4655783 │ │ ├─58910 /usr/bin/vector --config-dir /etc/vector/ │ │ └─58944 journalctl --follow --all --show-cursor --output=json --boot --since=2000-01-01 │ ├─5850f28cd8c4520f0558510e52eb532fc2b452b9fb6b813e1caa7913bd8aa2f1 │ │ └─42269 /coredns -conf /etc/coredns/Corefile │ ├─4714267e342e7d83b99f502cd9bf9efb06aab83847b50b2e747158c1e53d2395 │ │ └─92250 /pause │ ├─8acbf49662fc00fad24a1515ab3fe3f7fe0180dbcb42867bf3e90b68797ab191 │ │ └─56439 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ └─ec74ca21af9d44f5983cffd120074cbdbed28aa5a561356ce3443429a4073146 │ └─60023 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount