Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ ├─user-1002.slice │ │ ├─user@1002.service … │ │ │ ├─app.slice │ │ │ └─init.scope │ │ │ ├─21130 /lib/systemd/systemd --user │ │ │ └─21131 (sd-pam) │ │ └─session-22.scope │ │ ├─21120 sshd: cephadm [priv] │ │ └─21145 sshd: cephadm@notty │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─703 /lib/systemd/systemd --user │ │ └─704 (sd-pam) │ ├─session-4.scope │ │ ├─1453 sshd: zuul [priv] │ │ └─1466 sshd: zuul@notty │ └─session-1.scope │ ├─ 700 sshd: zuul [priv] │ ├─ 717 sshd: zuul@notty │ ├─ 865 /usr/bin/python3 │ ├─ 1457 ssh: /home/zuul/.ansible/cp/199.204.45.227-22-zuul [mux] │ ├─100486 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-yebvdpjtpsgpaoqzdducrlornihtbedk ; /usr/bin/python3'"'"' && sleep 0' │ ├─100487 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-yebvdpjtpsgpaoqzdducrlornihtbedk ; /usr/bin/python3' && sleep 0 │ ├─100488 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-yebvdpjtpsgpaoqzdducrlornihtbedk ; /usr/bin/python3 │ ├─100489 /bin/sh -c echo BECOME-SUCCESS-yebvdpjtpsgpaoqzdducrlornihtbedk ; /usr/bin/python3 │ ├─100490 /usr/bin/python3 │ ├─100491 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─100493 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 42 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 9801 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─12167 /usr/bin/containerd-shim-runc-v2 -namespace moby -id f4475af1d2ccae9abf95e58708152974d141eea9091215f79c57b7553ab01a93 -address /run/containerd/containerd.sock │ │ ├─12569 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 0d3d1115348c501dbf86b1fa894f23710a0de29097dea5aa4841082a9df8c950 -address /run/containerd/containerd.sock │ │ ├─19259 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 2244186db148a75d07d9078654232ea0157a047e71842e8e5033caaf9abcb8f5 -address /run/containerd/containerd.sock │ │ ├─24444 /usr/bin/containerd-shim-runc-v2 -namespace moby -id a6cc2f088cd0186cbbb29660d294ee7371054aa5d80bd9c6e06ac8b55a0b5221 -address /run/containerd/containerd.sock │ │ ├─28878 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 48f25281b49d8be9a92fb42fbc65e0eb19fc5e3cac68e99cfc70106477c50e36 -address /run/containerd/containerd.sock │ │ ├─33381 /usr/bin/containerd-shim-runc-v2 -namespace moby -id b6af329ea02d353d74fdff30bb385a276fd239cfac9853d5eac0611876e76cb3 -address /run/containerd/containerd.sock │ │ ├─39071 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 53c66716132ca6418a0266d44fc6c29943554a8be6e3ee3a59ffa0985ee12a18 -address /run/containerd/containerd.sock │ │ ├─39104 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1a90d59393b94dd06fba541a401c6ff690683f322a69db49f2565f67cc149f6f -address /run/containerd/containerd.sock │ │ ├─39130 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5a9347f12e9e6f365a924a86ad7d1d8f596762205d79fa97ff0721e232bb58cf -address /run/containerd/containerd.sock │ │ ├─39134 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0d5b4b27af360676dc32cec0354ab5ea20333ec88e27ce9196a4823b3439b161 -address /run/containerd/containerd.sock │ │ ├─39135 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ece4ff736166a3607562ffc297f78b51fe405a9f96c753753b47f182ad603747 -address /run/containerd/containerd.sock │ │ ├─39717 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5c977b1d0954def6ad10ba2e026f2a8f87c189110ab7690011a9dd93b373c65f -address /run/containerd/containerd.sock │ │ ├─40414 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f71b31184b2a677272ada05fcc974410c0ebcf420089133301a3c2cf37f4aece -address /run/containerd/containerd.sock │ │ ├─40417 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c9c030343c93db3249ea89f590319542fb414dd83dfc0a636e71e7d94e1bcb0d -address /run/containerd/containerd.sock │ │ ├─42157 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 220b54a7bc3bc212dcf0476c98acb4e8304617174597afe07fe04bb7b4a80cab -address /run/containerd/containerd.sock │ │ ├─42194 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3880cbf4b1c623580b91cf69b8ca3ae8da83cc7bc80154523fa18837b51fa596 -address /run/containerd/containerd.sock │ │ ├─42246 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8f101f00e1d87c5deedfc68b4dbff055e7c7a893cbf900c257294fc680dff79e -address /run/containerd/containerd.sock │ │ ├─42324 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 02f8e05e894edb6a0167475b3917a52329c954309810765366bbee4d2bf9ed8d -address /run/containerd/containerd.sock │ │ ├─42373 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b4393bd76465cba5ffae2e7c1230e22cde0b086f12356ca3f14c5ad3d1cbc8c4 -address /run/containerd/containerd.sock │ │ ├─42547 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 899243e06641674bf90dd06b184d8f1552f290d2580fbee3d1153bcaac70e1c7 -address /run/containerd/containerd.sock │ │ ├─44775 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0c2a87b56d759bf028c0650c56d61297d6938336bc245af441830cb3860dd04f -address /run/containerd/containerd.sock │ │ ├─44816 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 35019a0ee3f98d5b88fd360631957f8fb4916bb2bfa525f3f1d5c1311b00ac73 -address /run/containerd/containerd.sock │ │ ├─45545 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e8cd9a4957e6f5340aa48a045203d843908e00c2fef20713b28e241ef78c9f72 -address /run/containerd/containerd.sock │ │ ├─45661 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f0be05dffda046509b9a63b501ef8f20089a8b321feb02a9bad7e9759d0bf710 -address /run/containerd/containerd.sock │ │ ├─46027 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5018608fa5cfacd3038fa327d4d0463af3dd0c57877f10acd3f73a60c8cc74eb -address /run/containerd/containerd.sock │ │ ├─46303 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 02b8145f0a2f15ac1bdb61e29758972114b5046ab494048b3e2d9d6105f783ad -address /run/containerd/containerd.sock │ │ ├─46712 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id cde7adc862fe2465e49319ba1c3af79179aef30beb4ffaa86e66891b4c2ceb5f -address /run/containerd/containerd.sock │ │ ├─48438 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ef217634f35bd8972845c6376c40f04113125aa4d15fc39e79ac8aa9142f7a1d -address /run/containerd/containerd.sock │ │ ├─50389 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 224bcf206d63a3d1fe1cddd4a70347f399513d52785fae4fc4af2d91f742ae49 -address /run/containerd/containerd.sock │ │ ├─51013 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 66bf1739fa574f22bdc03e3e83b5825853db70f290dbb81946a351acbc8daa48 -address /run/containerd/containerd.sock │ │ ├─51059 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 67c2132d395ba248fb41abfe1b247e12afc3493d063af09edf346848fab83f61 -address /run/containerd/containerd.sock │ │ ├─51201 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8c8648e084e6d5c8258d7bc731c196decbeeddd7363d4c9f9c151adb59252a75 -address /run/containerd/containerd.sock │ │ ├─54214 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id defd05762b9357d35e562130a36460aa097d65f1b758f3e5c90100a40cdb1fff -address /run/containerd/containerd.sock │ │ ├─54354 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d5b55d320e765212ec632fbf81dbc40603a0e696f1df493333938d4c806f3211 -address /run/containerd/containerd.sock │ │ ├─54392 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6446a1d84b4dff18532c0e85f39f3625e4e71f4dd3d0f2513565a6fce4e55abd -address /run/containerd/containerd.sock │ │ ├─54445 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9dc6018d38037ae0e7b084ad3e4448af996c7b717deb24f1ffa143f7b2386816 -address /run/containerd/containerd.sock │ │ ├─55912 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id eab853b7064c087dd552c1d698757638ce181d2398360781e04f35af82db6a71 -address /run/containerd/containerd.sock │ │ ├─56096 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 2f76ca53cdb762b25c1e05b5062aabb6203aa8fe1b5f683e5c7a0af335e0d73e -address /run/containerd/containerd.sock │ │ ├─56143 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0a9dbdd72c0de43b7dacb2cde5921c07b5e19717bd91e76758b0e48398893cc0 -address /run/containerd/containerd.sock │ │ ├─56392 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 2f97868f3abcfccd9acae402516b13506b50d2a127e1bfdacf4a2938d5d4fb50 -address /run/containerd/containerd.sock │ │ ├─56745 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4ecbfb9a007e75a383c0fb14942a5ac2a2f379885ea8b416e6b1c477411340f5 -address /run/containerd/containerd.sock │ │ ├─57047 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 2b11d1fbf89b412c903c5ab8ee554a43b48ad96ba9f0ba610e3deff5708b2567 -address /run/containerd/containerd.sock │ │ ├─57316 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ba2c574b5dc852f0a5ac2249e886eff5dede2124fcaa051a057229dc3f3bc5d7 -address /run/containerd/containerd.sock │ │ ├─58051 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 05e3f72954e490808ef718e02f12d3bb37fc9b1d68a058df53f1b83dc6ea8402 -address /run/containerd/containerd.sock │ │ ├─59054 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b4fadf61e9fa31e79efd13be87e6541d1ae9a5c53f2c397e69cfad6f11be166f -address /run/containerd/containerd.sock │ │ ├─60977 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 597df630b1d82d9134ab4ee0379e5927ec2ea67806a94d2ebbecbbcec44579e2 -address /run/containerd/containerd.sock │ │ ├─68093 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0da426a526520b2342fed33fa76c42109053e69aa143011c0d28985f98b85c64 -address /run/containerd/containerd.sock │ │ ├─69272 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id dbd137a9c2d75062803ad1bd23c53676a227231683cf55daf6200a9a436aacd3 -address /run/containerd/containerd.sock │ │ ├─73928 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ae378228cecfdb89609b3306077ed8f3c8d5430cfec104328f6c7c23dbc9f447 -address /run/containerd/containerd.sock │ │ ├─78607 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ce5465b944285ee3fd429a242e466fcd0c8a5dd2b8d666c2e5d4b500c045ecec -address /run/containerd/containerd.sock │ │ ├─86491 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 805bbaa89894ce01bf03e3f2b166e81df082effb08798803f77c03882f17332e -address /run/containerd/containerd.sock │ │ ├─86529 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8824be327000c4c9d23078f79f50af517719c7568621278982b9f3b3d8d68246 -address /run/containerd/containerd.sock │ │ ├─90370 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 22cca5756b0dfcdcdad38220d7f3494319c58c1afe66d0f85f9daf873259f163 -address /run/containerd/containerd.sock │ │ ├─91475 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f15d892c07ea6c2c4b97eb89b5b204eb07dd26fb6e1f2815002ab634b88b03aa -address /run/containerd/containerd.sock │ │ ├─91532 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9236d326ec10bf9fb225852a7045c740ae2ac0ab38f4e1105e20a207e9d6150d -address /run/containerd/containerd.sock │ │ └─92284 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1f1f77acd838cd5815ed8d8016983e7d897afbbbc5aee69893ad09cd8b6ce6ee -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1187 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─520 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─392 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─626 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─system-ceph\x2d4837cbf8\x2d4f90\x2d4300\x2db3f6\x2d726c9b9f89b4.slice │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.1.service │ │ │ ├─28591 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/unit.run │ │ │ └─28861 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-1 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1:/var/lib/ceph/osd/ceph-1:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@crash.instance.service │ │ │ ├─19197 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/unit.run │ │ │ └─19217 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-crash --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-crash-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/config:/etc/ceph/ceph.conf:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/keyring:/etc/ceph/ceph.client.crash.instance.keyring quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n client.crash.instance │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.0.service │ │ │ ├─24150 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/unit.run │ │ │ └─24427 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-0 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0:/var/lib/ceph/osd/ceph-0:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.2.service │ │ │ ├─33086 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/unit.run │ │ │ └─33363 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-2 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2:/var/lib/ceph/osd/ceph-2:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mgr.instance.jugeix.service │ │ │ ├─12482 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.jugeix/unit.run │ │ │ └─12513 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mgr --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mgr-instance-jugeix --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.jugeix:/var/lib/ceph/mgr/ceph-instance.jugeix:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.jugeix/config:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v18.2.7 -n mgr.instance.jugeix -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mon.instance.service │ │ ├─12119 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/unit.run │ │ └─12150 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mon-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance:/var/lib/ceph/mon/ceph-instance:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev quay.io/ceph/ceph:v18.2.7 -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ ├─docker.service … │ │ └─10172 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock │ ├─polkit.service │ │ └─646 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─617 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─docker-b6af329ea02d353d74fdff30bb385a276fd239cfac9853d5eac0611876e76cb3.scope … │ │ ├─33405 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─33418 /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─kubelet.service │ │ └─39544 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.227 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─355 /lib/systemd/systemd-journald │ ├─docker-a6cc2f088cd0186cbbb29660d294ee7371054aa5d80bd9c6e06ac8b55a0b5221.scope … │ │ ├─24468 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─24482 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─ssh.service │ │ └─647 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─docker-48f25281b49d8be9a92fb42fbc65e0eb19fc5e3cac68e99cfc70106477c50e36.scope … │ │ ├─28901 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─28914 /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker-0d3d1115348c501dbf86b1fa894f23710a0de29097dea5aa4841082a9df8c950.scope … │ │ ├─12599 /sbin/docker-init -- /usr/bin/ceph-mgr -n mgr.instance.jugeix -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─12613 /usr/bin/ceph-mgr -n mgr.instance.jugeix -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1388 /usr/sbin/uuidd --socket-activation │ ├─docker-f4475af1d2ccae9abf95e58708152974d141eea9091215f79c57b7553ab01a93.scope … │ │ ├─12222 /sbin/docker-init -- /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ │ └─12242 /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ ├─cloud-config.service │ ├─chrony.service │ │ ├─10679 /usr/sbin/chronyd -F 1 │ │ └─10680 /usr/sbin/chronyd -F 1 │ ├─docker-2244186db148a75d07d9078654232ea0157a047e71842e8e5033caaf9abcb8f5.scope … │ │ ├─19285 /sbin/docker-init -- /usr/bin/ceph-crash -n client.crash.instance │ │ └─19302 /usr/bin/python3 -s /usr/bin/ceph-crash -n client.crash.instance │ ├─docker.socket │ ├─systemd-resolved.service │ │ └─397 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─cloud-init.service │ ├─dbus.service │ │ └─613 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─625 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─619 /lib/systemd/systemd-logind ├─k8s.io │ ├─2f97868f3abcfccd9acae402516b13506b50d2a127e1bfdacf4a2938d5d4fb50 │ │ └─56417 /pause │ ├─d50a7f90743e549990de7e5cac8a99e5f95db0d296e3f846ee75ff85a5b704ba │ │ └─46858 mysqld --wsrep_start_position=043ddfda-1f32-11f1-9672-0e0e4c597a89:22 │ ├─ab96675205e07e65b02b8a08e9222ced39dd7ba06a56f14121cc3acdae430653 │ │ └─58785 /bin/pushgateway │ ├─5c977b1d0954def6ad10ba2e026f2a8f87c189110ab7690011a9dd93b373c65f │ │ └─39742 /pause │ ├─fb3219b78875ff1740854f7da5ec2138ba54956702c7470bfc1dd7279bbbf25a │ │ ├─73308 barbiacan-api: uWSGI master │ │ └─73327 barbiacan-api: uWSGI worker 1 │ ├─e34dc5ebcf9ef5a38203edf7ed47b063522a10bd313b63148a4e87061de898b7 │ │ └─50626 /kubernetes-entrypoint │ ├─92aa6f00ec2d82277b8de71c432fca8c5bc415b399c8e7de064b80356d3e145c │ │ └─42484 /coredns -conf /etc/coredns/Corefile │ ├─9dc6018d38037ae0e7b084ad3e4448af996c7b717deb24f1ffa143f7b2386816 │ │ └─54469 /pause │ ├─00ad4c6eebcae8101c0da8ebdb2bebc897301c8ec1bbe04dc0d1e783d8b7f4f6 │ │ └─45862 /manager --metrics-bind-address=:8080 │ ├─597df630b1d82d9134ab4ee0379e5927ec2ea67806a94d2ebbecbbcec44579e2 │ │ └─61002 /pause │ ├─ece4ff736166a3607562ffc297f78b51fe405a9f96c753753b47f182ad603747 │ │ └─39248 /pause │ ├─beced5e1550d5a449845ff009723251b9dd496167b0698367941dcfb0c4c53b1 │ │ └─59280 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config │ ├─defd05762b9357d35e562130a36460aa097d65f1b758f3e5c90100a40cdb1fff │ │ └─54279 /pause │ ├─7c661aa9301006d18301880558d7b0418d7108e0d725c362ef7e9d1134e8a42c │ │ └─46761 /usr/bin/peer-list -on-change=/usr/bin/add_pxc_nodes.sh -service=percona-xtradb-pxc │ ├─0da426a526520b2342fed33fa76c42109053e69aa143011c0d28985f98b85c64 │ │ └─68116 /pause │ ├─3880cbf4b1c623580b91cf69b8ca3ae8da83cc7bc80154523fa18837b51fa596 │ │ └─42218 /pause │ ├─ce5465b944285ee3fd429a242e466fcd0c8a5dd2b8d666c2e5d4b500c045ecec │ │ └─78629 /pause │ ├─e8b281a40ae261e488811f4600a745b79612e8e7413066c17988825a63c8a646 │ │ ├─90642 glance-api: uWSGI master │ │ ├─90655 glance-api: uWSGI worker 1 │ │ └─90656 glance-api: uWSGI worker 2 │ ├─ffe5fabcce36804e74b83c174f1415256b0dcf32bc379ee59af4a8a54caaf366 │ │ └─59243 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.199-204-45-227.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml │ ├─6446a1d84b4dff18532c0e85f39f3625e4e71f4dd3d0f2513565a6fce4e55abd │ │ └─54417 /pause │ ├─7517438b8b5e64101471a993f442e22e70427f5b8826d872bcb4984da6d9220a │ │ ├─46605 haproxy -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ │ └─47979 haproxy -sf 11 -x sockpair@3 -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ ├─5923bb206ba8ddd03d8ad9ab88da8300a5350e628ff66d5f1c8522ffb3842ba2 │ │ └─42417 /coredns -conf /etc/coredns/Corefile │ ├─5896c020b3c02726260fe64e4633c92410741e0714cf9a7476c3bf65385ba470 │ │ └─59106 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.199-204-45-227.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml │ ├─eab853b7064c087dd552c1d698757638ce181d2398360781e04f35af82db6a71 │ │ └─55936 /pause │ ├─ba2c574b5dc852f0a5ac2249e886eff5dede2124fcaa051a057229dc3f3bc5d7 │ │ └─57337 /pause │ ├─f5d349017e8e9ed9c558401670e348bc564e550b740ea0d05929cabc219f7788 │ │ ├─91810 staffeln-conductor: master process [/var/lib/openstack/bin/staffeln-conductor --config-file /etc/staffeln/staffeln.conf] │ │ ├─92085 staffeln-conductor: Staffeln conductor backup controller worker(0) │ │ └─92089 staffeln-conductor: Staffeln conductor rotation controller worker(0) │ ├─05e3f72954e490808ef718e02f12d3bb37fc9b1d68a058df53f1b83dc6ea8402 │ │ └─58075 /pause │ ├─c9c030343c93db3249ea89f590319542fb414dd83dfc0a636e71e7d94e1bcb0d │ │ └─40468 /pause │ ├─774b2fcbd2b4b9aac7014c4330b761e897ca17ad7475a7c1803efe2f578f1ed4 │ │ └─44979 /server │ ├─71626486da9ba15c29752c14b920ead80f4eb5c399b3bd764f33f02015ab8bf4 │ │ └─46956 /bin/mysqld_exporter --collect.info_schema.processlist │ ├─b6dce40b865bef18aa3550405aea370fe08a5f637889b594fe6db7c17dcadb80 │ │ └─42702 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─45a2ba9c1f01793888646bdde7931488b9673a03fa1214fbbb44fb0817e7c5fd │ │ └─74727 /usr/local/bin/rook ceph operator │ ├─2f76ca53cdb762b25c1e05b5062aabb6203aa8fe1b5f683e5c7a0af335e0d73e │ │ └─56119 /pause │ ├─cef440929d683e2a1d7b146af52b5c241eb739fccea3e12614c5dac8ed86e62e │ │ └─58466 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml │ ├─f3de7734238c74e6dcc1f6c6143cf2d13a4b55f1e8f2dfecda9c2b96e8add980 │ │ └─42741 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager │ ├─9c82ed080a335ca0f663a021ffea4d301b7627a815b2ebe6569903085912acb5 │ │ └─91615 kubernetes-entrypoint │ ├─4ecbfb9a007e75a383c0fb14942a5ac2a2f379885ea8b416e6b1c477411340f5 │ │ └─56779 /pause │ ├─f71b31184b2a677272ada05fcc974410c0ebcf420089133301a3c2cf37f4aece │ │ └─40467 /pause │ ├─8ff1879359cc3427b463015d2a226d9a84244be1573f1c41868bf90209ea091d │ │ └─45738 /manager --metrics-bind-address=:9782 │ ├─9ce049abc126e388ab8ca7bde925083d5585cb0c3d1f456e05f94422c882092b │ │ └─48544 java -Dkc.config.built=true -Xms64m -Xmx512m -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --auto-build --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true │ ├─3b2bf4435ca64b581ec25b25f1631183d4e10b22bcafd0c16408edfb527131d7 │ │ └─59742 /bin/oauth2-proxy │ ├─ef217634f35bd8972845c6376c40f04113125aa4d15fc39e79ac8aa9142f7a1d │ │ └─48461 /pause │ ├─44fba82ddbbe588f3177b8315a71cd20de86e5b263bc8c6266ee4ae510ab8d68 │ │ ├─58171 /usr/bin/vector --config-dir /etc/vector/ │ │ └─58230 journalctl --follow --all --show-cursor --output=json --boot --since=2000-01-01 │ ├─fa21380e477c5d032b70f5a4e3bc899513be32a8381a500d49216ecbd3db929b │ │ ├─41444 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─42018 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─53c66716132ca6418a0266d44fc6c29943554a8be6e3ee3a59ffa0985ee12a18 │ │ └─39157 /pause │ ├─0c2a87b56d759bf028c0650c56d61297d6938336bc245af441830cb3860dd04f │ │ └─44802 /pause │ ├─372505654a456c7663d452e694bf0e62b72420c7f96b7cfe6b3c9ffe1a08a206 │ │ ├─63696 apache2 -DFOREGROUND │ │ ├─63711 (wsgi:k -DFOREGROUND │ │ ├─63712 (wsgi:k -DFOREGROUND │ │ ├─63713 (wsgi:k -DFOREGROUND │ │ ├─63714 (wsgi:k -DFOREGROUND │ │ ├─63718 apache2 -DFOREGROUND │ │ ├─63719 apache2 -DFOREGROUND │ │ ├─63720 apache2 -DFOREGROUND │ │ ├─63721 apache2 -DFOREGROUND │ │ ├─63725 apache2 -DFOREGROUND │ │ ├─63726 apache2 -DFOREGROUND │ │ ├─63727 apache2 -DFOREGROUND │ │ ├─63728 apache2 -DFOREGROUND │ │ ├─63729 apache2 -DFOREGROUND │ │ └─63730 apache2 -DFOREGROUND │ ├─0d5b4b27af360676dc32cec0354ab5ea20333ec88e27ce9196a4823b3439b161 │ │ └─39232 /pause │ ├─e89697d020363fed6dfc3c03cbc6fe65aa67e14c0217679e790b7fbebf529dc1 │ │ ├─57367 nginx: master process nginx -g daemon off; │ │ ├─57384 nginx: worker process │ │ ├─57385 nginx: worker process │ │ ├─57386 nginx: worker process │ │ ├─57387 nginx: worker process │ │ └─57388 nginx: worker process │ ├─549a6d6f44a5324eeade87376060cdde76b6c16935e415e74b070113b32f2627 │ │ └─59364 memcached -v -p 11211 -U 0 -c 8192 -m 1024 │ ├─66bf1739fa574f22bdc03e3e83b5825853db70f290dbb81946a351acbc8daa48 │ │ └─51038 /pause │ ├─0a9dbdd72c0de43b7dacb2cde5921c07b5e19717bd91e76758b0e48398893cc0 │ │ └─56167 /pause │ ├─8824be327000c4c9d23078f79f50af517719c7568621278982b9f3b3d8d68246 │ │ └─86559 /pause │ ├─e8cd9a4957e6f5340aa48a045203d843908e00c2fef20713b28e241ef78c9f72 │ │ └─45568 /pause │ ├─9236d326ec10bf9fb225852a7045c740ae2ac0ab38f4e1105e20a207e9d6150d │ │ └─91556 /pause │ ├─1a90d59393b94dd06fba541a401c6ff690683f322a69db49f2565f67cc149f6f │ │ └─39198 /pause │ ├─f0be05dffda046509b9a63b501ef8f20089a8b321feb02a9bad7e9759d0bf710 │ │ └─45683 /pause │ ├─805bbaa89894ce01bf03e3f2b166e81df082effb08798803f77c03882f17332e │ │ └─86514 /pause │ ├─5a9347f12e9e6f365a924a86ad7d1d8f596762205d79fa97ff0721e232bb58cf │ │ └─39223 /pause │ ├─41223e659ffdea8caf5426d5c36efee144c67d30976913b4bb3abbeb5cd3b59e │ │ └─86691 radosgw --fsid=4837cbf8-4f90-4300-b3f6-726c9b9f89b4 --keyring=/etc/ceph/keyring-store/keyring --log-to-stderr=true --err-to-stderr=true --mon-cluster-log-to-stderr=true --log-stderr-prefix=debug --default-log-to-file=false --default-mon-cluster-log-to-file=false --mon-host=[v2:10.96.240.200:3300,v1:10.96.240.200:6789] --mon-initial-members=instance --id=rgw.ceph.a --setuser=ceph --setgroup=ceph --foreground --rgw-frontends=beast port=8080 --host=rook-ceph-rgw-ceph-a-67fd8975c6-jn9kg --rgw-mime-types-file=/etc/ceph/rgw/mime.types --rgw-realm=ceph --rgw-zonegroup=ceph --rgw-zone=ceph │ ├─8c8648e084e6d5c8258d7bc731c196decbeeddd7363d4c9f9c151adb59252a75 │ │ └─51226 /pause │ ├─0cc1c6f08e76902f04f915b38c15717d274267f035b25c8f228c878ab1a8fac6 │ │ └─56928 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning │ ├─1f1f77acd838cd5815ed8d8016983e7d897afbbbc5aee69893ad09cd8b6ce6ee │ │ └─92309 /pause │ ├─8f101f00e1d87c5deedfc68b4dbff055e7c7a893cbf900c257294fc680dff79e │ │ └─42270 /pause │ ├─224bcf206d63a3d1fe1cddd4a70347f399513d52785fae4fc4af2d91f742ae49 │ │ └─50412 /pause │ ├─9acc68b424546a056f964d74629871c0c8fe4f7cefcaaf847512343ef1483efe │ │ └─58338 /goldpinger --static-file-path /static │ ├─f2710b5d8000851652b67515fda6845736d80fbe6d065c7f2562f5582b9adbd6 │ │ └─54826 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments │ ├─521c0ab8f46f14fa7702c33124aa07a4f7919000bef1b29b9ff698a819a00f2b │ │ ├─45178 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─45192 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─45270 nginx: master process /usr/bin/nginx -c /etc/nginx/nginx.conf │ │ ├─89737 nginx: worker process │ │ ├─89738 nginx: worker process │ │ └─89739 nginx: cache manager process │ ├─c6ebd880982728b8a3a44c979b733fde837dbf43bc6fc10b8867625dd42c4b48 │ │ └─39489 /kube-vip manager │ ├─5418e84bf695bd54a230ceee427244235da34024d8b99fba74f9d97f53563f78 │ │ └─59640 /bin/oauth2-proxy │ ├─48d8d407ce435defb6709d76b2aa81996390b2e3139b91b219ac7db4f03a4b1e │ │ └─59160 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 │ ├─d5b55d320e765212ec632fbf81dbc40603a0e696f1df493333938d4c806f3211 │ │ └─54377 /pause │ ├─35019a0ee3f98d5b88fd360631957f8fb4916bb2bfa525f3f1d5c1311b00ac73 │ │ └─44838 /pause │ ├─94b7256b99d5c7e05c14eda21b8e49f119a5620a4abb788d0dc67baf92303913 │ │ └─39316 kube-apiserver --advertise-address=199.204.45.227 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─b4fadf61e9fa31e79efd13be87e6541d1ae9a5c53f2c397e69cfad6f11be166f │ │ └─59076 /pause │ ├─7ec575ae3da60d3b5fe5f2055b4a655a566f812c1bea12e6a8df035437c4f059 │ │ └─54927 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.34.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 │ ├─899243e06641674bf90dd06b184d8f1552f290d2580fbee3d1153bcaac70e1c7 │ │ └─42571 /pause │ ├─dbd137a9c2d75062803ad1bd23c53676a227231683cf55daf6200a9a436aacd3 │ │ └─69295 /pause │ ├─67cbddbc8a9ccd323d2ae24250af4ec5ed2bd8a88cfb210bd37798ddb1c55629 │ │ └─58527 /usr/bin/loki -config.file=/etc/loki/config/config.yaml -target=all │ ├─f15d892c07ea6c2c4b97eb89b5b204eb07dd26fb6e1f2815002ab634b88b03aa │ │ └─91502 /pause │ ├─02b8145f0a2f15ac1bdb61e29758972114b5046ab494048b3e2d9d6105f783ad │ │ └─46331 /pause │ ├─42106bc94ead69aabad25df20bf0d2ddaa3070f7e9f5a33c7356e591130aa702 │ │ └─39766 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─17e1a712991adf599c6276ee8aa4baf7b2f8f6eabb36631e86cff677681166ef │ │ └─39350 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─b4393bd76465cba5ffae2e7c1230e22cde0b086f12356ca3f14c5ad3d1cbc8c4 │ │ └─42404 /pause │ ├─7d1ee1b81cfda0a0a653864de4bf63ac2be19299678d29fcfe0dce79a4a2ee84 │ │ ├─55305 python -u /app/sidecar.py │ │ ├─55597 python -u /app/sidecar.py │ │ └─55598 python -u /app/sidecar.py │ ├─e5e3482b2243572f329c1577664438a53dabc5edc524c2bc9cc17e0e5fea70b9 │ │ └─59203 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─bf7a0bac1e485bedba37e9909f454fad81be010fc72884640bea86a3556e66b0 │ │ └─51463 /secretgen-controller │ ├─22cca5756b0dfcdcdad38220d7f3494319c58c1afe66d0f85f9daf873259f163 │ │ └─90393 /pause │ ├─d016881f711b75473beb77eaa363854f1cbdef4941b7d33028573fb2fe97205f │ │ └─46171 percona-xtradb-cluster-operator │ ├─8ef8f831a13bd45d0595cc56896da1f43e67262970f8bbbbeedcbd41e8245d40 │ │ └─59854 /bin/memcached_exporter │ ├─3c04d0474da4b21b00e9792e355b353ca089699ea3d826785fe1d8591e445463 │ │ └─39402 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─bfa3263c0d318089aa119c66b7110ac40da85fdfe7b5687631b6584cbb5889c0 │ │ └─42602 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true │ ├─02f8e05e894edb6a0167475b3917a52329c954309810765366bbee4d2bf9ed8d │ │ └─42352 /pause │ ├─ecf0f1dd0db1be6f98e7d79294cc3267f045d3aa2caedc2f9817b7d8affb3abb │ │ └─51286 nfd-master -featurerules-controller=true │ ├─51e6f513573d5bd1657c87071132ce0fba57ba33ca629c4164fc219077f742d0 │ │ └─55615 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─c44736da7ce582aeae495bba436a5698b2b42db3dec5cb0e735a92f5d117a763 │ │ └─87497 /usr/libexec/platform-python -s /usr/bin/ceph-crash │ ├─2b11d1fbf89b412c903c5ab8ee554a43b48ad96ba9f0ba610e3deff5708b2567 │ │ └─57071 /pause │ ├─9cd3cf9586ba4eb3f00d284ddc461cfdaeced82d78d62f7a7f24da8ab2bce5e0 │ │ ├─68197 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─68216 /usr/local/lib/erlang/erts-12.3.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /usr/local/lib/erlang -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false │ │ ├─68223 erl_child_setup 1048576 │ │ ├─68277 /usr/local/lib/erlang/erts-12.3.2/bin/epmd -daemon │ │ ├─68328 inet_gethost 4 │ │ └─68329 inet_gethost 4 │ ├─5bb9316c0fa8e76abace77f5db78088492319383ef9fff28e46cfc643caa0840 │ │ └─39428 etcd --advertise-client-urls=https://199.204.45.227:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.227:2380 --initial-cluster=instance=https://199.204.45.227:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.227:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.227:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─f7f62207d0303c977b80754353b9b298c57918d834fc7d6184585e0dda016ae2 │ │ ├─55331 python -u /app/sidecar.py │ │ ├─55601 python -u /app/sidecar.py │ │ └─55602 python -u /app/sidecar.py │ ├─75be5afbd6ecfc313d4a8b820cbee6a340939da41d4f81a63cc32ae3edd2dc6d │ │ └─41352 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─ae378228cecfdb89609b3306077ed8f3c8d5430cfec104328f6c7c23dbc9f447 │ │ └─73953 /pause │ ├─220b54a7bc3bc212dcf0476c98acb4e8304617174597afe07fe04bb7b4a80cab │ │ └─42180 /pause │ ├─cde7adc862fe2465e49319ba1c3af79179aef30beb4ffaa86e66891b4c2ceb5f │ │ └─46735 /pause │ ├─bb1ef409daed0720829212f0b489d2f139e49ab8bbc759b733492da9673feae2 │ │ └─42789 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.10 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 │ ├─67c2132d395ba248fb41abfe1b247e12afc3493d063af09edf346848fab83f61 │ │ └─51094 /pause │ ├─c06656088e19470dfa146380a6aa43f4825d32d2cb198537d64211674c657d4f │ │ ├─78732 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─78749 /usr/local/lib/erlang/erts-12.3.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /usr/local/lib/erlang -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false │ │ ├─78756 erl_child_setup 1048576 │ │ ├─78810 /usr/local/lib/erlang/erts-12.3.2/bin/epmd -daemon │ │ ├─78861 inet_gethost 4 │ │ └─78862 inet_gethost 4 │ ├─545bc4f0fe8ce4b7a791a8098a781fad5e506db86c5f69e084dafee4ab4c99f3 │ │ └─51326 nfd-worker --server=node-feature-discovery-master:8080 │ ├─3e21eb172137e90f95685b7d13692fdae143929faf23be3d84458291bc86c55f │ │ ├─59660 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─59676 /usr/local/lib/erlang/erts-12.3.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /usr/local/lib/erlang -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false │ │ ├─59683 erl_child_setup 1048576 │ │ ├─59781 /usr/local/lib/erlang/erts-12.3.2/bin/epmd -daemon │ │ ├─59873 inet_gethost 4 │ │ └─59874 inet_gethost 4 │ ├─5018608fa5cfacd3038fa327d4d0463af3dd0c57877f10acd3f73a60c8cc74eb │ │ └─46050 /pause │ └─f8116c242fdb4ad3495d87e2b9f4fdd779bfa3b163559f89bc9c37e39adbf0f7 │ ├─86764 /bin/bash -x -e -m -c CEPH_CLIENT_ID=ceph-client.rgw.ceph.a PERIODICITY=daily LOG_ROTATE_CEPH_FILE=/etc/logrotate.d/ceph LOG_MAX_SIZE=500M # edit the logrotate file to only rotate a specific daemon log # otherwise we will logrotate log files without reloading certain daemons # this might happen when multiple daemons run on the same machine sed -i "s|*.log|$CEPH_CLIENT_ID.log|" "$LOG_ROTATE_CEPH_FILE" # replace default daily with given user input sed --in-place "s/daily/$PERIODICITY/g" "$LOG_ROTATE_CEPH_FILE" if [ "$LOG_MAX_SIZE" != "0" ]; then # adding maxsize $LOG_MAX_SIZE at the 4th line of the logrotate config file with 4 spaces to maintain indentation sed --in-place "4i \ \ \ \ maxsize $LOG_MAX_SIZE" "$LOG_ROTATE_CEPH_FILE" fi while true; do # we don't force the logrorate but we let the logrotate binary handle the rotation based on user's input for periodicity and size logrotate --verbose "$LOG_ROTATE_CEPH_FILE" sleep 15m done │ └─86831 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 15m ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount