Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ ├─user-1002.slice │ │ ├─user@1002.service … │ │ │ ├─app.slice │ │ │ └─init.scope │ │ │ ├─21707 /lib/systemd/systemd --user │ │ │ └─21708 (sd-pam) │ │ └─session-22.scope │ │ ├─21670 sshd: cephadm [priv] │ │ └─21724 sshd: cephadm@notty │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─706 /lib/systemd/systemd --user │ │ └─707 (sd-pam) │ ├─session-4.scope │ │ ├─1457 sshd: zuul [priv] │ │ └─1470 sshd: zuul@notty │ └─session-1.scope │ ├─ 703 sshd: zuul [priv] │ ├─ 720 sshd: zuul@notty │ ├─ 865 /usr/bin/python3 │ ├─ 1461 ssh: /home/zuul/.ansible/cp/199.204.45.216-22-zuul [mux] │ ├─122868 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-axvhldlkbbbemmgtarqjdekxustxqque ; /usr/bin/python3'"'"' && sleep 0' │ ├─122869 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-axvhldlkbbbemmgtarqjdekxustxqque ; /usr/bin/python3' && sleep 0 │ ├─122870 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-axvhldlkbbbemmgtarqjdekxustxqque ; /usr/bin/python3 │ ├─122871 /bin/sh -c echo BECOME-SUCCESS-axvhldlkbbbemmgtarqjdekxustxqque ; /usr/bin/python3 │ ├─122872 /usr/bin/python3 │ ├─122873 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─122875 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 42 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 9797 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─ 12437 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 13f45870b4729cce67dd4e93cad579b1489ef0558c44d621db5748b1de08ee42 -address /run/containerd/containerd.sock │ │ ├─ 12817 /usr/bin/containerd-shim-runc-v2 -namespace moby -id d9be62e0722cfa7b924cab98c0cd16db8320ea2dbd1bce43f284ab3c7657b656 -address /run/containerd/containerd.sock │ │ ├─ 19780 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 5083d812b9f4f137ba9ce8fdc4ae4a675f7f7ba1fc5e27becb9a3faeb4b27f26 -address /run/containerd/containerd.sock │ │ ├─ 24964 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 1c74b49499f9019f621cf6cac361da10e2a8da48d77f8134b99b2cc647d49abb -address /run/containerd/containerd.sock │ │ ├─ 29506 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 36418c76bd6230bc5221679d7d0e5c5444647239f38455471a4b84ccd5d7dd1e -address /run/containerd/containerd.sock │ │ ├─ 33987 /usr/bin/containerd-shim-runc-v2 -namespace moby -id e448f89b61631cd0d0de832b8e29ff1e441088e57bad9e91782adde38c915f0b -address /run/containerd/containerd.sock │ │ ├─ 39375 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1d23dd045ab0b2c6e33643bfa65eb2ecf456a9db53b27111806684a0e5c32b49 -address /run/containerd/containerd.sock │ │ ├─ 39379 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7e8dab1e2dcbf6355926895dcee17daa44a9cb94280f31d914233e57cfb37cc4 -address /run/containerd/containerd.sock │ │ ├─ 39417 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 414cc2e2b5e0b86dbab15b38512a67198f6430fdec8b28c153a144edc8e6a9d4 -address /run/containerd/containerd.sock │ │ ├─ 39419 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id cefe5bf49c770aba139abe0907bf8920d98f98c40f74b3b4d5168fb0260f140b -address /run/containerd/containerd.sock │ │ ├─ 39429 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d853093d7bb2cdfbd4e55ac6ba017cef3e0714644b32f1faa73cf59273866b8f -address /run/containerd/containerd.sock │ │ ├─ 39978 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9e106cd8d81d217ca17fece9534836802dd4ad4999d1f8c28d2ca0e6f2fcfb9a -address /run/containerd/containerd.sock │ │ ├─ 40705 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 231e269bf0a2eaef8c70b6a6b728af99f6c03e244abd58cf9af3aa4ab5c4651b -address /run/containerd/containerd.sock │ │ ├─ 40722 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id eac459e43d3219fa9516d3be4d6fb4bef5040fb2944cea92dedc4d115ab309e2 -address /run/containerd/containerd.sock │ │ ├─ 42079 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id aa4088e2cd2cdb4fd0f0fc52dc89efac5e050746c1797da8684e6363dffda375 -address /run/containerd/containerd.sock │ │ ├─ 42522 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f97db45ea14b21c1a4da0fffa2d92ddb95164619ecc454ed184e767595ddf17e -address /run/containerd/containerd.sock │ │ ├─ 42559 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 27ab1599535086cf08b2645dc225bc12635f6fe547af47fe95e627ba8cfe75fd -address /run/containerd/containerd.sock │ │ ├─ 43017 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c024a6d6db5c1d6dd073a9b87a5a6a7c99afe14ac4172f42dd3bf30660cb1e88 -address /run/containerd/containerd.sock │ │ ├─ 43519 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6b7d6f2b7566bfd542e2136ea349a3d514bd509271a031697780aa50d5c28d43 -address /run/containerd/containerd.sock │ │ ├─ 43541 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 406f274685ba98a48c714a003a48feffceaaeb61254f116442c5833ecefda26a -address /run/containerd/containerd.sock │ │ ├─ 43554 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ad4c0e36a7b870ec84d6f2e39ba2e5a7fc054a6b0e0893c092d37f62a79ca4ae -address /run/containerd/containerd.sock │ │ ├─ 45944 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 86dadf64ac0eb5bb1f72feb892e73a1bbaa5ee28c4736c32c5012357a5c630e1 -address /run/containerd/containerd.sock │ │ ├─ 46036 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 35058656a054783c06fc73f0b08c3513bf1ef71ad71d5b61205c724c890da419 -address /run/containerd/containerd.sock │ │ ├─ 46978 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d1e56664addcb79543dca9c152e6309489468122346e39bf0386077e1ea6a13d -address /run/containerd/containerd.sock │ │ ├─ 47076 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ded44c587c002ef3545a35ffe29c71c012bef88d91c129f3f62d503e9a6a3e3a -address /run/containerd/containerd.sock │ │ ├─ 47448 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c0d0f81e849b69ea22aed5767533f03a4b6c9f1914719fa3168c7c58b379370d -address /run/containerd/containerd.sock │ │ ├─ 47745 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id fd8ff134d741bc74ed33ecf0c7dfe669268e263adb2b18851562741ac41442ba -address /run/containerd/containerd.sock │ │ ├─ 48178 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 54a4086485ff8d91f40d0ba3eacfb1cc8ebfc403e6c6c94a35adc4acca640820 -address /run/containerd/containerd.sock │ │ ├─ 50377 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 48004063b6fc2787bc2e7f6d16275750d5e5c29ae8d5f8f0cd3a03291f4a3d17 -address /run/containerd/containerd.sock │ │ ├─ 50580 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6c57dfcb1d5369e2dc812549081761a7340aa1e4f82e4d1c4ee1c22c515015c0 -address /run/containerd/containerd.sock │ │ ├─ 55091 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f7c7233aad30f7e98ef34ff1ef4d238cfa0e8e16111f1e1e7314c3e88b944509 -address /run/containerd/containerd.sock │ │ ├─ 55769 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8c5e2fa4dab318b2474a9a7e92d8f2fa1d793c33bc0b67c413439095dc29089e -address /run/containerd/containerd.sock │ │ ├─ 55810 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c7230897dc41fe46adbc65cdab5022a8a87aea8cb17ace1336ef9f390ed57702 -address /run/containerd/containerd.sock │ │ ├─ 55862 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 56fce6bfe381c0cd99b64e2937a1b9680abf895f51d4047dbbe5d0dc665f90f3 -address /run/containerd/containerd.sock │ │ ├─ 56016 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id caa54bf05e00677f7898b37317309156cf99aee1758657271beef60d0c8ec2bd -address /run/containerd/containerd.sock │ │ ├─ 60586 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 553406028d57cea7d7bba496aa59fa456a47c9950fc54d6286800a6ef13ac494 -address /run/containerd/containerd.sock │ │ ├─ 60677 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f958dcab20d56dba76932b7a0115e83c03ba5b99e82994be633e22b2ae497483 -address /run/containerd/containerd.sock │ │ ├─ 60715 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0c88b36757405c9b9f42dc34422053f7d6c92149c810679839a1000b2b622f8a -address /run/containerd/containerd.sock │ │ ├─ 60760 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4cb2f8ad242492b53ff9e7c63f475ff17063e7c385a312d091979390d13b8a2b -address /run/containerd/containerd.sock │ │ ├─ 62035 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 02d2a47ec4c167a414478082294de618404612034db6c7778bb76c8a3e7488cf -address /run/containerd/containerd.sock │ │ ├─ 62471 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 317b8b108b6d4fdfcbd2a65ffe04cea62f012e42e612cea6d481e19df2366081 -address /run/containerd/containerd.sock │ │ ├─ 62705 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 97c0203b6564d0a224d1201427ce68f00a9d7b379ef421bb665c9b0741959e3e -address /run/containerd/containerd.sock │ │ ├─ 62744 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 04addbc0bcc37da5e3e92e5ac9b93849b2541c941a09078b4f251be1122f3b4b -address /run/containerd/containerd.sock │ │ ├─ 62790 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 35ceaa791e619dc5afabb400c57ccb8b98a6d04b92b1c9cf3aa2c8c9d5ee369d -address /run/containerd/containerd.sock │ │ ├─ 63444 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d5262ced7c52b2b616143069cac215193d285772b9345e3768fe7e354c8f2d66 -address /run/containerd/containerd.sock │ │ ├─ 63931 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ff1af535527830bcfbb16353f9fc4ce8b347c69ffe46ce920cba8a7fe5188431 -address /run/containerd/containerd.sock │ │ ├─ 64098 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f721cc9fb84ffb78c3d200fe28f0bbf9e58852a215cf4710be19999acc1712ea -address /run/containerd/containerd.sock │ │ ├─ 64421 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3f9acef249a8a0e90af7e147e6eb3eb461548cdae22cc1962cff83fa81074f7a -address /run/containerd/containerd.sock │ │ ├─ 65110 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id bac33a8f6006a1914be61d836599459f465b9efdb898da4c7a74c782a50dc1e1 -address /run/containerd/containerd.sock │ │ ├─ 66723 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id da4d2f9ea3a9bf4663f642a937cecc6ead4c2388ae891a7153b1ca91f662574f -address /run/containerd/containerd.sock │ │ ├─ 68669 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b541a595301a6542bbfef0bca2738c5f9f857de67c44557e576b9195d2d83889 -address /run/containerd/containerd.sock │ │ ├─ 78004 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a430e8f26d4e500c262ecf4bbc9e0402c27e4ed2afe982fb7ea39ee3a8500988 -address /run/containerd/containerd.sock │ │ ├─ 79529 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 34496d4eaca6076343562ed325a678174e915469c6b8a22b0c03daba7b58b9a5 -address /run/containerd/containerd.sock │ │ ├─ 85788 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 03d39e5e2fcb6a7e7d2f03ebe8481d1f162fcd3143d177bbbaecf7237d2fd369 -address /run/containerd/containerd.sock │ │ ├─ 92128 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d449da1c4b71d38bb64659220a1fc81ad53f172ab2afa9bcc804a0a725852ae7 -address /run/containerd/containerd.sock │ │ ├─ 95893 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7d5401742d79f37708aaee48fe102894021fe7748db76caa0b5a7b2613cf73c1 -address /run/containerd/containerd.sock │ │ ├─ 95931 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1de1cba409169276bd958944cb995c52093206a56f7340f1328fa33d9e777b85 -address /run/containerd/containerd.sock │ │ ├─109306 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id af1408b4f0308aa97bbbcddc5c738c1403d0fe25b869b77dc79271359ddd1aee -address /run/containerd/containerd.sock │ │ ├─110833 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 34842dc80fb988b39fdd7fc08597e30354ad103da2a496173fba887e8bab2fab -address /run/containerd/containerd.sock │ │ ├─110910 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 38563adf6a64a72beb2d847fef811cc244877ad647f37f1308718b9204968e34 -address /run/containerd/containerd.sock │ │ └─111658 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e19c9006bfe032dfc03a0170a0311f8ed0650940b53b0130ff0b9ece72abc8f1 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1191 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─523 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─392 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─629 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─system-ceph\x2d4837cbf8\x2d4f90\x2d4300\x2db3f6\x2d726c9b9f89b4.slice │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.1.service │ │ │ ├─29215 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/unit.run │ │ │ └─29487 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-1 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1:/var/lib/ceph/osd/ceph-1:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@crash.instance.service │ │ │ ├─19740 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/unit.run │ │ │ └─19759 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-crash --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-crash-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/config:/etc/ceph/ceph.conf:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/keyring:/etc/ceph/ceph.client.crash.instance.keyring quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n client.crash.instance │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.0.service │ │ │ ├─24685 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/unit.run │ │ │ └─24947 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-0 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0:/var/lib/ceph/osd/ceph-0:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.2.service │ │ │ ├─33690 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/unit.run │ │ │ └─33970 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-2 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2:/var/lib/ceph/osd/ceph-2:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mon.instance.service │ │ │ ├─12372 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/unit.run │ │ │ └─12400 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mon-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance:/var/lib/ceph/mon/ceph-instance:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev quay.io/ceph/ceph:v18.2.7 -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ │ └─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mgr.instance.bmbdjr.service │ │ ├─12733 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.bmbdjr/unit.run │ │ └─12763 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mgr --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mgr-instance-bmbdjr --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.bmbdjr:/var/lib/ceph/mgr/ceph-instance.bmbdjr:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.bmbdjr/config:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v18.2.7 -n mgr.instance.bmbdjr -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker.service … │ │ └─10168 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock │ ├─polkit.service │ │ └─648 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─619 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─39831 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.216 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─docker-13f45870b4729cce67dd4e93cad579b1489ef0558c44d621db5748b1de08ee42.scope … │ │ ├─12479 /sbin/docker-init -- /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ │ └─12492 /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ ├─cloud-init-local.service │ ├─systemd-journald.service │ │ └─358 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─651 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─cloud-final.service │ ├─docker-5083d812b9f4f137ba9ce8fdc4ae4a675f7f7ba1fc5e27becb9a3faeb4b27f26.scope … │ │ ├─19806 /sbin/docker-init -- /usr/bin/ceph-crash -n client.crash.instance │ │ └─19824 /usr/bin/python3 -s /usr/bin/ceph-crash -n client.crash.instance │ ├─uuidd.service │ │ └─1392 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─chrony.service │ │ ├─10929 /usr/sbin/chronyd -F 1 │ │ └─10930 /usr/sbin/chronyd -F 1 │ ├─docker-36418c76bd6230bc5221679d7d0e5c5444647239f38455471a4b84ccd5d7dd1e.scope … │ │ ├─29530 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─29543 /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker-e448f89b61631cd0d0de832b8e29ff1e441088e57bad9e91782adde38c915f0b.scope … │ │ ├─34011 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─34024 /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker-1c74b49499f9019f621cf6cac361da10e2a8da48d77f8134b99b2cc647d49abb.scope … │ │ ├─24988 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─25000 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker.socket │ ├─docker-d9be62e0722cfa7b924cab98c0cd16db8320ea2dbd1bce43f284ab3c7657b656.scope … │ │ ├─12841 /sbin/docker-init -- /usr/bin/ceph-mgr -n mgr.instance.bmbdjr -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─12860 /usr/bin/ceph-mgr -n mgr.instance.bmbdjr -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─systemd-resolved.service │ │ └─398 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─dbus.service │ │ └─616 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─628 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─622 /lib/systemd/systemd-logind ├─k8s.io │ ├─469fe7d6bc664ef128510fd921ac0e1046ac356a2414f8ca81fc76948c65c90e │ │ └─48541 /bin/mysqld_exporter --mysqld.username=monitor --collect.info_schema.processlist │ ├─e19c9006bfe032dfc03a0170a0311f8ed0650940b53b0130ff0b9ece72abc8f1 │ │ └─111681 /pause │ ├─948318d15654ee5c6370035f783ae09be45ad5cd4b3224790a4b7e8f243908c6 │ │ └─96223 radosgw --fsid=4837cbf8-4f90-4300-b3f6-726c9b9f89b4 --keyring=/etc/ceph/keyring-store/keyring --default-log-to-stderr=true --default-err-to-stderr=true --default-mon-cluster-log-to-stderr=true --default-log-stderr-prefix=debug --default-log-to-file=false --default-mon-cluster-log-to-file=false --mon-host=[v2:10.96.240.200:3300,v1:10.96.240.200:6789] --mon-initial-members=instance --id=rgw.ceph.a --setuser=ceph --setgroup=ceph --foreground --rgw-frontends=beast port=8080 --host=rook-ceph-rgw-ceph-a-699b8bdb59-cbwqb --rgw-mime-types-file=/etc/ceph/rgw/mime.types --rgw-realm=ceph --rgw-zonegroup=ceph --rgw-zone=ceph │ ├─6db06a31c3848be3f85dde971e20cc43e40754622281fba7f1c7c3cfea96f5ce │ │ └─65498 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 │ ├─10277bf06f6a1ab99be1125454dbb3d0fce764abbef2bd802a577f4bfa59eebf │ │ └─66426 /bin/memcached_exporter │ ├─1c4b51f261f7ca5d82ccc454c0020d6c3740c1d24414d09fcacd06a8594eca73 │ │ └─110897 kubernetes-entrypoint │ ├─fee3e3e809ef65f029fdc31c304731d6d5cb84520898a65814e0876daf1d92fb │ │ ├─111324 staffeln-conductor: master process [/var/lib/openstack/bin/staffeln-conductor --config-file /etc/staffeln/staffeln.conf] │ │ ├─111714 staffeln-conductor: Staffeln conductor backup controller worker(0) │ │ └─111723 staffeln-conductor: Staffeln conductor rotation controller worker(0) │ ├─067aa966b4a373d15998f6a192dbd0cd603502cd41e9c7cd9b09706cbd3046d6 │ │ └─63336 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning │ ├─ff551f0ca51d2f784ebeb13060116638477a5df2b8cce6f22a206f0ba2ef9e8b │ │ └─55297 /kubernetes-entrypoint │ ├─c31367f6d34fbcbbba57aab3353f0bdb8907c975d73e4daea311c01446087d4b │ │ ├─78166 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─78183 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─78189 erl_child_setup 1048576 │ │ ├─78236 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─78237 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─78302 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon │ │ └─78403 /bin/sh -s rabbit_disk_monitor │ ├─2ac94512472511925215ca491733e54d3a8c8444e7503f96daf453e8b7698bdc │ │ └─66784 /bin/oauth2-proxy │ ├─0c88b36757405c9b9f42dc34422053f7d6c92149c810679839a1000b2b622f8a │ │ └─60739 /pause │ ├─da4d2f9ea3a9bf4663f642a937cecc6ead4c2388ae891a7153b1ca91f662574f │ │ └─66746 /pause │ ├─35c4fa3ab2e46004e920ebcaf1cb548323ffdc6cc202da4b26534015e9815711 │ │ └─47150 /manager --metrics-bind-address=:9782 │ ├─48004063b6fc2787bc2e7f6d16275750d5e5c29ae8d5f8f0cd3a03291f4a3d17 │ │ └─50400 /pause │ ├─10b7cfa565f5724bf0fb0b31d39789e90866fd69eb57101968638668a7c2193a │ │ └─65540 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─61a2589e5a09e3bec06a6ecdb164a6a7dd393b1fede5e2e4f0b6f43122095c20 │ │ ├─61858 python -u /app/sidecar.py │ │ ├─62107 python -u /app/sidecar.py │ │ └─62108 python -u /app/sidecar.py │ ├─d7639c1fee165d51fd988f0ecadedd437ebef5af5fcba363c35fe93c5e99ccef │ │ ├─48444 mysqld --wsrep_start_position=245b2361-3dc7-11f1-be98-9701925ad5a1:20 │ │ └─48566 /var/lib/mysql/mysql-state-monitor │ ├─6ff838e2e8b254bb2424d6977d12376b4dc0c356e68facd19de1426be75763d5 │ │ └─43870 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager │ ├─778c111e555946a88373c1103630235ba92a2446dd3e778e9681d403d639ea23 │ │ └─42143 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─2fc8dbea9be7e8072007fce013f8da18e091539bb6fbd70cc41ac6fad14e0229 │ │ ├─41219 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─41597 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─110a44ffc680559742a0f2af531017622720494b46c078bd0b9e43196bced691 │ │ ├─72247 apache2 -DFOREGROUND │ │ ├─72261 (wsgi:k -DFOREGROUND │ │ ├─72262 (wsgi:k -DFOREGROUND │ │ ├─72263 (wsgi:k -DFOREGROUND │ │ ├─72264 (wsgi:k -DFOREGROUND │ │ ├─72265 apache2 -DFOREGROUND │ │ ├─72266 apache2 -DFOREGROUND │ │ ├─72267 apache2 -DFOREGROUND │ │ ├─72273 apache2 -DFOREGROUND │ │ ├─72275 apache2 -DFOREGROUND │ │ ├─72276 apache2 -DFOREGROUND │ │ ├─72277 apache2 -DFOREGROUND │ │ ├─72278 apache2 -DFOREGROUND │ │ ├─72279 apache2 -DFOREGROUND │ │ └─72280 apache2 -DFOREGROUND │ ├─3f9acef249a8a0e90af7e147e6eb3eb461548cdae22cc1962cff83fa81074f7a │ │ └─64443 /pause │ ├─1c4c57760159c78a67b0a08cd121c56bcf9ac820a505b67d0488965ec5d4d79e │ │ └─65455 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.199-204-45-216.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml │ ├─bb9dcdd582eac3b4c918d56f5ae640d73b89378a579840f91e1e5fb2435f4897 │ │ └─56291 /secretgen-controller │ ├─85bbb97a5eee7b8a134f7f7a87b46e7788e9c8f0a40cefc5c36d8c5852afd508 │ │ └─39772 /kube-vip manager │ ├─b8443930975eb73b32ddf0c27cdd818889eb5fc4bcac7703ac1db67dec86b164 │ │ └─46331 /server │ ├─56fce6bfe381c0cd99b64e2937a1b9680abf895f51d4047dbbe5d0dc665f90f3 │ │ └─55885 /pause │ ├─a384712c02829fed18698a543c2deec4f76126638f6982627cf2b5a14dae9aa7 │ │ └─51044 valkey-server *:26379 [sentinel] │ ├─1f015525358d9c7b84756888099bfb0d67354ccb556696add684b0b17caf8ff4 │ │ └─64012 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml │ ├─cd7cd6686c5c0a1e91be1b6d612b6e5119cecbcc9cd6a62206be96c0e2fc0870 │ │ ├─48230 haproxy -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ │ └─50787 haproxy -sf 153 -x sockpair@5 -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ ├─317b8b108b6d4fdfcbd2a65ffe04cea62f012e42e612cea6d481e19df2366081 │ │ └─62493 /pause │ ├─1e8eee85d7e7b4a34f098725a3e61a8567da4b07e3ec3f3964732bf1f3796494 │ │ └─64145 memcached -m 8192 --extended=modern,track_sizes -I 5m -c 16384 -v -u 11211 │ ├─bf4dc8cf0687a48b2c91af7ea1d1d166cf0048e0b5e102c49c26384435e04898 │ │ └─51134 redis_exporter │ ├─caa54bf05e00677f7898b37317309156cf99aee1758657271beef60d0c8ec2bd │ │ └─56039 /pause │ ├─fdaa2ce9cf0f9f1c9519656d2f1fe799d2baec93c91f18851bba36483d53f9c0 │ │ └─66062 /bin/memcached_exporter --memcached.address=localhost:11211 --web.listen-address=0.0.0.0:9150 │ ├─cefe5bf49c770aba139abe0907bf8920d98f98c40f74b3b4d5168fb0260f140b │ │ └─39518 /pause │ ├─4b10e44365c8a1050c21b85130f929bff8f8304d5668319a21191af9da682922 │ │ └─61286 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments │ ├─278263cfdf9ac665e73cab3a5ef27455436099ece023734381aff5db26b278d9 │ │ └─65592 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.199-204-45-216.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml │ ├─f5ff2c6a7899292771d894e47c1802f84466688ac9197d716c2e3095f5bbeffa │ │ ├─92409 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─92435 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─92450 erl_child_setup 1048576 │ │ ├─92623 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─92624 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─92634 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon │ │ └─92835 /bin/sh -s rabbit_disk_monitor │ ├─d449da1c4b71d38bb64659220a1fc81ad53f172ab2afa9bcc804a0a725852ae7 │ │ └─92152 /pause │ ├─c024a6d6db5c1d6dd073a9b87a5a6a7c99afe14ac4172f42dd3bf30660cb1e88 │ │ └─43041 /pause │ ├─4cb2f8ad242492b53ff9e7c63f475ff17063e7c385a312d091979390d13b8a2b │ │ └─60783 /pause │ ├─38563adf6a64a72beb2d847fef811cc244877ad647f37f1308718b9204968e34 │ │ └─110984 /pause │ ├─162ae027642c9b5b0e2ea7eaf073e6ffad3a3b8ce22af2d66cff37ef1b1ee5f0 │ │ └─39659 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─1ade7dcf3e6c3223cc87650bde58e822e37fbd4ea15094deb4df598896dc49a0 │ │ └─40027 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─aa4088e2cd2cdb4fd0f0fc52dc89efac5e050746c1797da8684e6363dffda375 │ │ └─42102 /pause │ ├─8c5e2fa4dab318b2474a9a7e92d8f2fa1d793c33bc0b67c413439095dc29089e │ │ └─55793 /pause │ ├─34496d4eaca6076343562ed325a678174e915469c6b8a22b0c03daba7b58b9a5 │ │ └─79553 /pause │ ├─34842dc80fb988b39fdd7fc08597e30354ad103da2a496173fba887e8bab2fab │ │ └─110861 /pause │ ├─6c57dfcb1d5369e2dc812549081761a7340aa1e4f82e4d1c4ee1c22c515015c0 │ │ └─50604 /pause │ ├─e63da7700c980a81edf41067fc060cf8233d7fc3706da3c69ba77a577096aaff │ │ └─50646 valkey-server *:6379 │ ├─c0c0498f3b5959b27b063d043e89a3f5028f7fb57bebe3c650e0f2e5c455049b │ │ └─66392 /bin/pushgateway │ ├─af1408b4f0308aa97bbbcddc5c738c1403d0fe25b869b77dc79271359ddd1aee │ │ └─109328 /pause │ ├─04addbc0bcc37da5e3e92e5ac9b93849b2541c941a09078b4f251be1122f3b4b │ │ └─62771 /pause │ ├─a05147fd1d5ef8c7142095c3434c739012092708870e004b1f5dbac8fb41c436 │ │ └─41001 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─ac4c15a8c58ee814c2066de34fa9da69353af984bdc63818b35475f1ecd37252 │ │ └─65363 memcached -v -p 11211 -U 0 -c 8192 -m 1024 │ ├─5a42dca3edbb6a010ab271c56a2ed6a7078d67b8cf1dc0accf677e68f5812cc7 │ │ └─47597 percona-xtradb-cluster-operator │ ├─35ceaa791e619dc5afabb400c57ccb8b98a6d04b92b1c9cf3aa2c8c9d5ee369d │ │ └─62815 /pause │ ├─02d2a47ec4c167a414478082294de618404612034db6c7778bb76c8a3e7488cf │ │ └─62058 /pause │ ├─4016ebaa80e47e8cbbdca572f96135416f4c636acee19e12b8f49944a8d670c9 │ │ └─62077 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─57f6ef1fda4b3dedfb0a5e8fcfd5eb928e2bdd269659ca09bca7af5eba987170 │ │ ├─61829 python -u /app/sidecar.py │ │ ├─62024 python -u /app/sidecar.py │ │ └─62025 python -u /app/sidecar.py │ ├─d5557ec072e8c0ff19e074a3db8cdc9b36f6ae2c17ea8d2cafe250d164b92e84 │ │ └─66155 /bin/memcached_exporter --memcached.address=localhost:11211 --web.listen-address=0.0.0.0:9150 │ ├─414cc2e2b5e0b86dbab15b38512a67198f6430fdec8b28c153a144edc8e6a9d4 │ │ └─39536 /pause │ ├─d983f073ee033c4363dadd218324f737b9ca9ff5a2cba15a9e8e4c8517b27a1d │ │ └─65801 /goldpinger --static-file-path /static │ ├─5c5e6092dc3de7b0681a1709d10be5be70a4efab347133e37cdf06ddc3466672 │ │ └─56147 nfd-worker -metrics=8081 │ ├─48b20a3e54e0e91e0436bf974515d054797569d85d0086d25d13efb7b280c5b7 │ │ └─39609 kube-apiserver --advertise-address=199.204.45.216 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─97c0203b6564d0a224d1201427ce68f00a9d7b379ef421bb665c9b0741959e3e │ │ └─62730 /pause │ ├─d853093d7bb2cdfbd4e55ac6ba017cef3e0714644b32f1faa73cf59273866b8f │ │ └─39537 /pause │ ├─f721cc9fb84ffb78c3d200fe28f0bbf9e58852a215cf4710be19999acc1712ea │ │ └─64120 /pause │ ├─42c1fafe2255e890198fe5590078904383c5e91233dbb1677bd4a2774f69de87 │ │ └─56101 nfd-master -crd-controller=true -metrics=8081 │ ├─0b2e86c9b4e1834226a7f176c14572153c87f11a70c2c18ac7029740cc35937a │ │ └─43779 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.17 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 │ ├─91c13c2cfd013f395c11ee6b10eaba136e870e5f8bb88699a0a8a61b16081c2c │ │ └─65633 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config │ ├─ff1af535527830bcfbb16353f9fc4ce8b347c69ffe46ce920cba8a7fe5188431 │ │ └─63955 /pause │ ├─406f274685ba98a48c714a003a48feffceaaeb61254f116442c5833ecefda26a │ │ └─43626 /pause │ ├─86dadf64ac0eb5bb1f72feb892e73a1bbaa5ee28c4736c32c5012357a5c630e1 │ │ └─45976 /pause │ ├─fd8ff134d741bc74ed33ecf0c7dfe669268e263adb2b18851562741ac41442ba │ │ └─47772 /pause │ ├─7440da161905f43cbda20c5246a27a00873b88309733686cc365c5dcf282bf9d │ │ └─43825 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true │ ├─a430e8f26d4e500c262ecf4bbc9e0402c27e4ed2afe982fb7ea39ee3a8500988 │ │ └─78028 /pause │ ├─eac459e43d3219fa9516d3be4d6fb4bef5040fb2944cea92dedc4d115ab309e2 │ │ └─40767 /pause │ ├─ef2266bc599ca05e1b21b88cbdd3f76f5c7ee562b50bea5973a77ba9c57e8816 │ │ └─56184 nfd-gc -gc-interval=1h │ ├─1de1cba409169276bd958944cb995c52093206a56f7340f1328fa33d9e777b85 │ │ └─95955 /pause │ ├─c7230897dc41fe46adbc65cdab5022a8a87aea8cb17ace1336ef9f390ed57702 │ │ └─55845 /pause │ ├─e8b648361e34289e39e52c8ebee0cd12b4e2122dc2ee8546e32897eedd0c30e6 │ │ ├─64905 /usr/bin/vector --config-dir /etc/vector/ │ │ └─64955 journalctl --follow --all --show-cursor --output=json --boot --since=2000-01-01 │ ├─64603238c1ae683ab9d770968525600909cb96bc7f475a6402b24ccd7523f950 │ │ ├─85001 barbiacan-api: uWSGI master │ │ └─85015 barbiacan-api: uWSGI worker 1 │ ├─1d74fee6f4881c34bdcd3d81bf6042d0e45bcc85df28d8db00c28e102e08c0d8 │ │ └─61058 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.35.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 │ ├─f7c7233aad30f7e98ef34ff1ef4d238cfa0e8e16111f1e1e7314c3e88b944509 │ │ └─55115 /pause │ ├─8dba4cb37bdc40367c6186d54f7ebaebfc7207a49428eb650f89f2091425d015 │ │ ├─64521 nginx: master process nginx -g daemon off; │ │ ├─64542 nginx: worker process │ │ ├─64543 nginx: worker process │ │ ├─64544 nginx: worker process │ │ ├─64545 nginx: worker process │ │ └─64546 nginx: worker process │ ├─231e269bf0a2eaef8c70b6a6b728af99f6c03e244abd58cf9af3aa4ab5c4651b │ │ └─40749 /pause │ ├─ad4c0e36a7b870ec84d6f2e39ba2e5a7fc054a6b0e0893c092d37f62a79ca4ae │ │ └─43630 /pause │ ├─f4b92b2650650c0e8b17a549583a8a7fa53a0dcaf9687442f8b0d3a6f33c4ad5 │ │ ├─ 46282 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─ 46294 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─ 46353 nginx: master process /usr/bin/nginx -c /etc/nginx/nginx.conf │ │ ├─105086 nginx: worker process │ │ ├─105087 nginx: worker process │ │ └─105088 nginx: cache manager process │ ├─7e8dab1e2dcbf6355926895dcee17daa44a9cb94280f31d914233e57cfb37cc4 │ │ └─39496 /pause │ ├─1d23dd045ab0b2c6e33643bfa65eb2ecf456a9db53b27111806684a0e5c32b49 │ │ └─39480 /pause │ ├─d5bae91f9173105e2ed28eaf53e71f25b4e9c15b3e1ff8d36501c24dc3a6394b │ │ ├─67144 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─67165 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─67175 erl_child_setup 1048576 │ │ ├─67236 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─67237 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─67247 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon │ │ └─67337 /bin/sh -s rabbit_disk_monitor │ ├─1ec6adf7b83b59ecc6da4d129d30951b6c7f4af9f519db394f806bb91fa862fd │ │ └─66522 /bin/oauth2-proxy │ ├─d5262ced7c52b2b616143069cac215193d285772b9345e3768fe7e354c8f2d66 │ │ └─63470 /pause │ ├─27ab1599535086cf08b2645dc225bc12635f6fe547af47fe95e627ba8cfe75fd │ │ └─42584 /pause │ ├─98acbe9314d83e746ee7832c498cac558ce7c9708f43ead94cb3090c8af0d593 │ │ └─42646 /coredns -conf /etc/coredns/Corefile │ ├─4fba262d0ea3484e17342bb5fdcaee315f5f7898867eed6a7bf6dc4e182efa62 │ │ └─42615 /coredns -conf /etc/coredns/Corefile │ ├─a9dd803268cf1e0e1ef35c193d5c6dcc20a25a99c597bd525e647d468c6c0773 │ │ └─39640 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─35058656a054783c06fc73f0b08c3513bf1ef71ad71d5b61205c724c890da419 │ │ └─46060 /pause │ ├─6b7d6f2b7566bfd542e2136ea349a3d514bd509271a031697780aa50d5c28d43 │ │ └─43600 /pause │ ├─6b5ff78c96ef72ab1133f83b2ffae075fdebafde907b85c6a6cd625edfde50b0 │ │ └─64231 memcached -m 1024 --extended=modern,track_sizes -I 5m -c 16384 -v -u 11211 │ ├─553406028d57cea7d7bba496aa59fa456a47c9950fc54d6286800a6ef13ac494 │ │ └─60623 /pause │ ├─ded44c587c002ef3545a35ffe29c71c012bef88d91c129f3f62d503e9a6a3e3a │ │ └─47100 /pause │ ├─b541a595301a6542bbfef0bca2738c5f9f857de67c44557e576b9195d2d83889 │ │ └─68718 /pause │ ├─f958dcab20d56dba76932b7a0115e83c03ba5b99e82994be633e22b2ae497483 │ │ └─60701 /pause │ ├─08c41a9a8de3e7021e8faff306ce41ed149787d147f91e81ef60f97fc7d88e57 │ │ └─86818 /usr/local/bin/rook ceph operator │ ├─c0d0f81e849b69ea22aed5767533f03a4b6c9f1914719fa3168c7c58b379370d │ │ └─47472 /pause │ ├─c91c425880a66ad63133874d43a315d4f0d860646ded4849b7665136a3fb2055 │ │ └─39622 etcd --advertise-client-urls=https://199.204.45.216:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.216:2380 --initial-cluster=instance=https://199.204.45.216:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.216:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.216:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─d1e56664addcb79543dca9c152e6309489468122346e39bf0386077e1ea6a13d │ │ └─47002 /pause │ ├─03d39e5e2fcb6a7e7d2f03ebe8481d1f162fcd3143d177bbbaecf7237d2fd369 │ │ └─85811 /pause │ ├─72d6dda50ceb4b3597ba0ae66930a84d41c2f6740521ad257f7c3d776fe8fe3f │ │ └─48309 /opt/percona/peer-list -on-change=/opt/percona/haproxy_add_pxc_nodes.sh -service=percona-xtradb-pxc -protocol=$(PEER_LIST_SRV_PROTOCOL) │ ├─bac33a8f6006a1914be61d836599459f465b9efdb898da4c7a74c782a50dc1e1 │ │ └─65132 /pause │ ├─7d5401742d79f37708aaee48fe102894021fe7748db76caa0b5a7b2613cf73c1 │ │ └─95917 /pause │ ├─a9ee1994d378c68c5a684b41a25fea2a074277d0338a2283e2ffcdca7550a231 │ │ └─65949 /usr/bin/loki -config.file=/etc/loki/config/config.yaml -target=all │ ├─e8f3cb55b546609ac891112c66ebeb6b39d43b077bd9c4b8ae50ab8f58d81ab1 │ │ ├─109749 glance-api: uWSGI master │ │ ├─109769 glance-api: uWSGI worker 1 │ │ └─109770 glance-api: uWSGI worker 2 │ ├─9e106cd8d81d217ca17fece9534836802dd4ad4999d1f8c28d2ca0e6f2fcfb9a │ │ └─40002 /pause │ ├─cc53bff3a48e3c6b501def1f18a151bbeed71e61fa21a8ded79fd2a607d0289a │ │ └─47277 /manager --metrics-bind-address=:8080 │ ├─216929f4d256732d9f246608b31ebbdd65410051a538716c900b3397c9edbed2 │ │ └─50853 java -Dkc.config.built=true -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 -XX:FlightRecorderOptions=stackdepth=512 -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:MaxRAMPercentage=70 -XX:MinRAMPercentage=70 -XX:InitialRAMPercentage=50 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dpicocli.disable.closures=true -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true │ ├─59c91901134d8268ac0a28b0fa97cd15645f1d1a03d53bc9085fe5fc88a4c89d │ │ └─43739 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─3dce6017cfbeaef34668131bd8354a5cd947b44cc0e16756c2655a670eca52ff │ │ └─97054 /usr/bin/python3 -s /usr/bin/ceph-crash │ ├─f97db45ea14b21c1a4da0fffa2d92ddb95164619ecc454ed184e767595ddf17e │ │ └─42549 /pause │ ├─54a4086485ff8d91f40d0ba3eacfb1cc8ebfc403e6c6c94a35adc4acca640820 │ │ └─48201 /pause │ └─7830d9f502783f5af331556d2cc5d511f95a354652e00af37a0ac15f6075f255 │ ├─96360 /bin/bash -x -e -m -c CEPH_CLIENT_ID=ceph-client.rgw.ceph.a PERIODICITY=daily LOG_ROTATE_CEPH_FILE=/etc/logrotate.d/ceph LOG_MAX_SIZE=500M ROTATE=7 # edit the logrotate file to only rotate a specific daemon log # otherwise we will logrotate log files without reloading certain daemons # this might happen when multiple daemons run on the same machine sed -i "s|*.log|$CEPH_CLIENT_ID.log|" "$LOG_ROTATE_CEPH_FILE" # replace default daily with given user input sed --in-place "s/daily/$PERIODICITY/g" "$LOG_ROTATE_CEPH_FILE" # replace rotate count, default 7 for all ceph daemons other than rbd-mirror sed --in-place "s/rotate 7/rotate $ROTATE/g" "$LOG_ROTATE_CEPH_FILE" if [ "$LOG_MAX_SIZE" != "0" ]; then # adding maxsize $LOG_MAX_SIZE at the 4th line of the logrotate config file with 4 spaces to maintain indentation sed --in-place "4i \ \ \ \ maxsize $LOG_MAX_SIZE" "$LOG_ROTATE_CEPH_FILE" fi while true; do # we don't force the logrorate but we let the logrotate binary handle the rotation based on user's input for periodicity and size logrotate --verbose "$LOG_ROTATE_CEPH_FILE" sleep 15m done │ └─96417 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 15m ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount