Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ ├─user-1002.slice │ │ ├─user@1002.service … │ │ │ ├─app.slice │ │ │ └─init.scope │ │ │ ├─21647 /lib/systemd/systemd --user │ │ │ └─21648 (sd-pam) │ │ └─session-22.scope │ │ ├─21643 sshd: cephadm [priv] │ │ └─21662 sshd: cephadm@notty │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─701 /lib/systemd/systemd --user │ │ └─702 (sd-pam) │ ├─session-4.scope │ │ ├─1455 sshd: zuul [priv] │ │ └─1468 sshd: zuul@notty │ └─session-1.scope │ ├─ 698 sshd: zuul [priv] │ ├─ 715 sshd: zuul@notty │ ├─ 863 /usr/bin/python3 │ ├─ 1459 ssh: /home/zuul/.ansible/cp/162.253.55.12-22-zuul [mux] │ ├─116811 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-gskjqhxblvurpkqbiblhrcookdhcdulc ; /usr/bin/python3'"'"' && sleep 0' │ ├─116812 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-gskjqhxblvurpkqbiblhrcookdhcdulc ; /usr/bin/python3' && sleep 0 │ ├─116813 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-gskjqhxblvurpkqbiblhrcookdhcdulc ; /usr/bin/python3 │ ├─116814 /bin/sh -c echo BECOME-SUCCESS-gskjqhxblvurpkqbiblhrcookdhcdulc ; /usr/bin/python3 │ ├─116815 /usr/bin/python3 │ ├─116827 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─116829 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 42 nofb ├─system.slice │ ├─docker-7b45b8a4ec6edb4d054b291ca13ae452871d752509f076afb272c1369066d034.scope … │ │ ├─19807 /sbin/docker-init -- /usr/bin/ceph-crash -n client.crash.instance │ │ └─19825 /usr/bin/python3 -s /usr/bin/ceph-crash -n client.crash.instance │ ├─containerd.service … │ │ ├─ 9800 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─ 12432 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 1b15b716c111bb70286ad478e968c002914d447a133b4b1644258bba659dab4c -address /run/containerd/containerd.sock │ │ ├─ 12821 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 66800e98d63a52b778bd6dd7ad630f5a01fdbf490321b3d225ae33310aac9e5d -address /run/containerd/containerd.sock │ │ ├─ 19779 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 7b45b8a4ec6edb4d054b291ca13ae452871d752509f076afb272c1369066d034 -address /run/containerd/containerd.sock │ │ ├─ 24948 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 1c843ddc28f1aa55df87de3c1ae739697fd287cc15a935c50396964571067f57 -address /run/containerd/containerd.sock │ │ ├─ 29677 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 6c233ef8ed2731811e29d8facd583a19b911d979d348f10808b33dd5f1c034df -address /run/containerd/containerd.sock │ │ ├─ 34344 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 0f42e7b6d9127f66213da47bd04acb007505245831ba36c156406a69ccbbba07 -address /run/containerd/containerd.sock │ │ ├─ 39716 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4d70e7b903f232dcbe6677f427fd2df3170921760366e846cb60cfe1a11ef0c2 -address /run/containerd/containerd.sock │ │ ├─ 39771 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5df1125c788b6bf861d54605454d2b046a807b35618cd2de91686438dcf175cc -address /run/containerd/containerd.sock │ │ ├─ 39773 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e2e09b5c5b6f865adb4e53b4212a1dadbf86ebbbfb214377c7b476d4d08aeec7 -address /run/containerd/containerd.sock │ │ ├─ 39778 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a4a957744e74e09a4b8061882ff98da8dcf512aab38381d0844ed6e8f7299aab -address /run/containerd/containerd.sock │ │ ├─ 39789 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 2be49b5ce2cbe75b813442c7d6198b91250cce983e0b4cc2a5fb65f838ddd642 -address /run/containerd/containerd.sock │ │ ├─ 40314 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c555971bae7cc4059591e559035b9761b60b9b98bcef83576001eb6c97f29cc5 -address /run/containerd/containerd.sock │ │ ├─ 41064 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e10076f59d0718f8321ea1983117423984e1c48e08386e0e0ea8be025c7d2d84 -address /run/containerd/containerd.sock │ │ ├─ 41073 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5103fb43e8756ea490965dd582a4b00a86be04b74af202449f4fe45518297c92 -address /run/containerd/containerd.sock │ │ ├─ 42760 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 82c35fc2d216b9f6ea3d3042f28435c77fa9e86688e0f84b2f17d03a4d91a38e -address /run/containerd/containerd.sock │ │ ├─ 42799 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id dc9482564c54e29157d98ac178ed2339db1f5651661a17f0533768f409c348ca -address /run/containerd/containerd.sock │ │ ├─ 42842 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0ae3b78fa032746179735bf719ba52b9e1b5a1d052ec5ad0b624c7f8ae3b6585 -address /run/containerd/containerd.sock │ │ ├─ 42878 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ba74543bffd478efa5b24518958a7c92b40ad25efda11e3e17e57c9899a27bfc -address /run/containerd/containerd.sock │ │ ├─ 43113 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 438a6c2e0cd6f4dd7fccc61251eaf771c4a297434dd38d43db5a90034818be23 -address /run/containerd/containerd.sock │ │ ├─ 43186 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 16920a135e9c98172fc45c3a3900642a4494144cdb053eb16806e8d144c1fd2b -address /run/containerd/containerd.sock │ │ ├─ 45561 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b5e4ffd73e76b9f2ebc1dd96d43144f55c5d7f4e212e643893fc16eca0fb6f99 -address /run/containerd/containerd.sock │ │ ├─ 45660 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 16a01ae464ce67bb94b81e718ad3793b2046a623898a7db6724fc446dc4d263d -address /run/containerd/containerd.sock │ │ ├─ 46383 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 79b7da39a5d3703aab5c07f548d0607c99168a44d3a4806f1a9c74e3ec553742 -address /run/containerd/containerd.sock │ │ ├─ 46486 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 75393f846a92d79b6622292c665e3ad073099023186191116747f7d7e77fc3f6 -address /run/containerd/containerd.sock │ │ ├─ 46874 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8bfd41b676875cee8280d88ca0a9c415a6bfc1f998f067dc7f7ed83682d5f8cf -address /run/containerd/containerd.sock │ │ ├─ 47181 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id dea27d4517cf05d0c2a313bdf66459912074c3287b85af4aafff52fc0556ec52 -address /run/containerd/containerd.sock │ │ ├─ 47595 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 43ec1f98df03547a7d66a28224d2ad60343c1be2ebffb67328826c3e45a21790 -address /run/containerd/containerd.sock │ │ ├─ 50259 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ba0a6fb9008dd9a72a4f42674913e14b3ff4d3bbeb9f71b1cf5a97699f0cb28c -address /run/containerd/containerd.sock │ │ ├─ 50348 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1a5f6db11338e8a2cc66b3987e5f69f3422b7cca897dcff80ed64d3ca1c9dda3 -address /run/containerd/containerd.sock │ │ ├─ 54527 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8d1e5cc52670225e0c682d27365f0a3fefe8851f1f5dcbbe6d428ac207c36977 -address /run/containerd/containerd.sock │ │ ├─ 55185 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id caed602b8fea5243ace5b5788ab0ca76f3e352ec0f93416e2bb42e6c1c9fdd25 -address /run/containerd/containerd.sock │ │ ├─ 55224 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id cde1a348184b2044f0fd98c2251db859a0989db484df58334f5bc93276e9f692 -address /run/containerd/containerd.sock │ │ ├─ 55276 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 58782a9bdaa7df348b9441eaedf9b3e445052cf5e43ee31c1d22dd70540b332d -address /run/containerd/containerd.sock │ │ ├─ 55441 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id abc72f159b50194d0a397b4ad53cfdf1e5a18f7ee68bb30cf98d27af3cc8b4e3 -address /run/containerd/containerd.sock │ │ ├─ 59303 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e64e25daebf84309e6dde01e906c27a96fe01a89e07fe9ff4fefbcc81ed4a33d -address /run/containerd/containerd.sock │ │ ├─ 59432 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id cd42c13d0be44b58d66c65ea94b5fde8c48f16b3c36f36a5d731e2921f3bab70 -address /run/containerd/containerd.sock │ │ ├─ 59486 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 77327fe00e765e089ef46d76ffdbe9b9d4baefc07969b83654c8cbf3397da8d4 -address /run/containerd/containerd.sock │ │ ├─ 59559 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9d81d4ad04462a031afae97fe2f272fb225e518280fad07a4be6516096fdd6bb -address /run/containerd/containerd.sock │ │ ├─ 61128 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 05eedf22c6a4b995c4918d8577ff1ebbf2fb93f86f5523dddf9da8310db90e00 -address /run/containerd/containerd.sock │ │ ├─ 61187 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ab3b868bac3ea3121c9c6ebea55daac6a9db441c92d8c16aee007001334eb1c3 -address /run/containerd/containerd.sock │ │ ├─ 61356 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d99b6d86f7a4172fbc714c30b386941e182b78bc36497b8e8f83f40289e594fc -address /run/containerd/containerd.sock │ │ ├─ 61664 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id dc05fb6edce2d94d14a082ccdf55e4b8422c9f2790aca6778efa4b0abc7d4674 -address /run/containerd/containerd.sock │ │ ├─ 61755 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 00182cedc01af68615bbef6558dec6d692b9d6be1be3fc6e81f100e645016ac4 -address /run/containerd/containerd.sock │ │ ├─ 61902 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 12067e2090837d9ff4b8983b0b75fc4991d1f2e83ccec88bdb2872f9dbef1bb4 -address /run/containerd/containerd.sock │ │ ├─ 62273 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 72f67ad5653f7c1cd237d2af26a0f78296c0f6c8156f42c4c1661aae907ec4b4 -address /run/containerd/containerd.sock │ │ ├─ 62576 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ae39312692ce4302596a5b09768b10c611bcfe9579a04a0124ab851c1dea8018 -address /run/containerd/containerd.sock │ │ ├─ 62684 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c2deec5032b30d0ad185eb8e85d4663ba802c27ee01d78309b08bc023cbc8a94 -address /run/containerd/containerd.sock │ │ ├─ 63563 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 40540810eb27d736791926aad529eb7c2dcf53cd50e3bff84b25618e19d2805d -address /run/containerd/containerd.sock │ │ ├─ 64971 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4ef0e61c1fd3bd33641c9b6fa74d6f43679fccfd7defe252059b414ec0156627 -address /run/containerd/containerd.sock │ │ ├─ 67619 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 58ad8a9c94061201d1dee39e2855982afe3e9e6a916fe0a0ad15674e45855e61 -address /run/containerd/containerd.sock │ │ ├─ 76692 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c0ca97b810097fa083ad574cf6bef829df41ba7512a4097faf39ebe25fea34b3 -address /run/containerd/containerd.sock │ │ ├─ 78224 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7b5b639aa7e4cc14caedb01c9c640c8ab19fd7b4bb878553fa59969cd49d3f30 -address /run/containerd/containerd.sock │ │ ├─ 84150 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3af186198d8a6845539e1676191c8bbe880aa4065492864f9acc8a575fb3b648 -address /run/containerd/containerd.sock │ │ ├─ 90476 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id cc0e1cabd6ea59b7c920853375ec690d0272be9f1b4b7e01efeeb8a2e665f3a5 -address /run/containerd/containerd.sock │ │ ├─ 94305 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 404e81771fc3ebabefbc57511eb8557e48ed87e5130a4e5c3866c2bf2ace705c -address /run/containerd/containerd.sock │ │ ├─ 94352 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 36f107147ea36f9b1b8f14df3411b1ab54632133148cc44605537343369d7379 -address /run/containerd/containerd.sock │ │ ├─103321 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id caf5b5d5dcb1545c0dd13588f3ddf7b02b25e8be5bd062398c451df437fb02cb -address /run/containerd/containerd.sock │ │ ├─104739 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d69d6da050bf27a2215ff6c5f939f065e89fa6ace2c1f1d6e094976d9b227151 -address /run/containerd/containerd.sock │ │ ├─104859 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a48ecb3ed532919cc13bc94b8a981a471e020ebe0019cb894d5feabe06ddbc33 -address /run/containerd/containerd.sock │ │ └─105541 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 32475549b13f39856b845ff3630d7dabae4bc62d401fe7ae3ee4ba43375a6488 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1186 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─519 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─390 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─624 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─system-ceph\x2d4837cbf8\x2d4f90\x2d4300\x2db3f6\x2d726c9b9f89b4.slice │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.1.service │ │ │ ├─29381 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/unit.run │ │ │ └─29658 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-1 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1:/var/lib/ceph/osd/ceph-1:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mgr.instance.ilfgad.service │ │ │ ├─12736 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.ilfgad/unit.run │ │ │ └─12767 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mgr --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mgr-instance-ilfgad --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.ilfgad:/var/lib/ceph/mgr/ceph-instance.ilfgad:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.ilfgad/config:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v18.2.7 -n mgr.instance.ilfgad -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@crash.instance.service │ │ │ ├─19738 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/unit.run │ │ │ └─19758 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-crash --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-crash-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/config:/etc/ceph/ceph.conf:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/keyring:/etc/ceph/ceph.client.crash.instance.keyring quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n client.crash.instance │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.0.service │ │ │ ├─24669 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/unit.run │ │ │ └─24930 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-0 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0:/var/lib/ceph/osd/ceph-0:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.2.service │ │ │ ├─34053 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/unit.run │ │ │ └─34327 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-2 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2:/var/lib/ceph/osd/ceph-2:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mon.instance.service │ │ ├─12376 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/unit.run │ │ └─12407 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mon-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance:/var/lib/ceph/mon/ceph-instance:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev quay.io/ceph/ceph:v18.2.7 -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ ├─docker.service … │ │ └─10171 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock │ ├─polkit.service │ │ └─639 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─614 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─docker-1b15b716c111bb70286ad478e968c002914d447a133b4b1644258bba659dab4c.scope … │ │ ├─12478 /sbin/docker-init -- /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ │ └─12498 /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ ├─kubelet.service │ │ └─40171 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=162.253.55.12 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─357 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─646 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─docker-1c843ddc28f1aa55df87de3c1ae739697fd287cc15a935c50396964571067f57.scope … │ │ ├─24971 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─24984 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1390 /usr/sbin/uuidd --socket-activation │ ├─docker-6c233ef8ed2731811e29d8facd583a19b911d979d348f10808b33dd5f1c034df.scope … │ │ ├─29701 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─29713 /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker-66800e98d63a52b778bd6dd7ad630f5a01fdbf490321b3d225ae33310aac9e5d.scope … │ │ ├─12844 /sbin/docker-init -- /usr/bin/ceph-mgr -n mgr.instance.ilfgad -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─12866 /usr/bin/ceph-mgr -n mgr.instance.ilfgad -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─chrony.service │ │ ├─10931 /usr/sbin/chronyd -F 1 │ │ └─10932 /usr/sbin/chronyd -F 1 │ ├─docker-0f42e7b6d9127f66213da47bd04acb007505245831ba36c156406a69ccbbba07.scope … │ │ ├─34369 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─34381 /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker.socket │ ├─systemd-resolved.service │ │ └─393 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─dbus.service │ │ └─611 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─620 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─617 /lib/systemd/systemd-logind ├─k8s.io │ ├─ebe31f4cf0829218f222318869d0fec9e19f9a0dcc1e4fb7facc9ca69cbdfb86 │ │ └─65811 /bin/oauth2-proxy │ ├─ae39312692ce4302596a5b09768b10c611bcfe9579a04a0124ab851c1dea8018 │ │ └─62599 /pause │ ├─be92a2b8cb919c910ef41ccf3a549eaf5d85967c23786d7128c72abbfbcd8413 │ │ └─54735 /kubernetes-entrypoint │ ├─79b7da39a5d3703aab5c07f548d0607c99168a44d3a4806f1a9c74e3ec553742 │ │ └─46406 /pause │ ├─fd26b913036c363f34862a38c28a5077b840894052b31f9ca596828b018c3f06 │ │ └─63676 memcached -v -p 11211 -U 0 -c 8192 -m 1024 │ ├─ece16841cf01090c3f87e2c3732a6607cd68984a84f51f3dd4f0dea0b4250d3e │ │ ├─42035 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─42592 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─e095b773ec8ecbdd07c24ac05405b2a1321dfbc68b6f2c3fe33380e09e3a7e80 │ │ └─50777 valkey-server *:26379 [sentinel] │ ├─0ec4e78fb6654d013807eb2bae9812399e0cb2df13f3dd7971867c85673fffda │ │ ├─94931 /bin/bash -x -e -m -c CEPH_CLIENT_ID=ceph-client.rgw.ceph.a PERIODICITY=daily LOG_ROTATE_CEPH_FILE=/etc/logrotate.d/ceph LOG_MAX_SIZE=500M ROTATE=7 # edit the logrotate file to only rotate a specific daemon log # otherwise we will logrotate log files without reloading certain daemons # this might happen when multiple daemons run on the same machine sed -i "s|*.log|$CEPH_CLIENT_ID.log|" "$LOG_ROTATE_CEPH_FILE" # replace default daily with given user input sed --in-place "s/daily/$PERIODICITY/g" "$LOG_ROTATE_CEPH_FILE" # replace rotate count, default 7 for all ceph daemons other than rbd-mirror sed --in-place "s/rotate 7/rotate $ROTATE/g" "$LOG_ROTATE_CEPH_FILE" if [ "$LOG_MAX_SIZE" != "0" ]; then # adding maxsize $LOG_MAX_SIZE at the 4th line of the logrotate config file with 4 spaces to maintain indentation sed --in-place "4i \ \ \ \ maxsize $LOG_MAX_SIZE" "$LOG_ROTATE_CEPH_FILE" fi while true; do # we don't force the logrorate but we let the logrotate binary handle the rotation based on user's input for periodicity and size logrotate --verbose "$LOG_ROTATE_CEPH_FILE" sleep 15m done │ │ └─94951 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 15m │ ├─ef83442fe3b0d1bdec69599d94b9c79e8eea48e19513490a1b9d5dd718a18bc6 │ │ ├─47646 haproxy -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ │ └─49247 haproxy -sf 15 -x sockpair@4 -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ ├─abc72f159b50194d0a397b4ad53cfdf1e5a18f7ee68bb30cf98d27af3cc8b4e3 │ │ └─55464 /pause │ ├─404e81771fc3ebabefbc57511eb8557e48ed87e5130a4e5c3866c2bf2ace705c │ │ └─94330 /pause │ ├─b5e4ffd73e76b9f2ebc1dd96d43144f55c5d7f4e212e643893fc16eca0fb6f99 │ │ └─45596 /pause │ ├─69376e2edd2ee7806d1224534c36a7a11cfadd348c48f62405509986dd6d48fd │ │ └─45952 /server │ ├─0ae3b78fa032746179735bf719ba52b9e1b5a1d052ec5ad0b624c7f8ae3b6585 │ │ └─42880 /pause │ ├─23793921405478d70b5351fb55e1a0bf9a016a7c4f3214cadf58228c6b63068e │ │ └─65022 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.162-253-55-12.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml │ ├─cc0e1cabd6ea59b7c920853375ec690d0272be9f1b4b7e01efeeb8a2e665f3a5 │ │ └─90500 /pause │ ├─8354038004fdf2227e15b0ed29d688fa0043361f03dbf2535ecf633c78077914 │ │ └─42927 /coredns -conf /etc/coredns/Corefile │ ├─2aa4034949cc3368ff1a749b839b8883a436f282f8d30e4bc042e00858cb9ac3 │ │ └─55514 nfd-worker -metrics=8081 │ ├─adb5b1659a0fb3a8c6f8620816b2002429d49ee4eb771a32202a1f273840db53 │ │ └─59913 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments │ ├─3416c39524102a176275e3a073c43f63462e112e8360775206d00b16c130a604 │ │ └─40002 etcd --advertise-client-urls=https://162.253.55.12:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://162.253.55.12:2380 --initial-cluster=instance=https://162.253.55.12:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://162.253.55.12:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://162.253.55.12:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─11a7c41eda99d90f292969cfe9b177ec8d5bf4d8e8294f055511e6718214b449 │ │ └─43238 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true │ ├─8d1e5cc52670225e0c682d27365f0a3fefe8851f1f5dcbbe6d428ac207c36977 │ │ └─54551 /pause │ ├─be970f9e80ac07a944f7f7afafb20194aa71b2e52cb0048303eeb8d5e1aa6e7f │ │ └─64195 /usr/bin/loki -config.file=/etc/loki/config/config.yaml -target=all │ ├─ba74543bffd478efa5b24518958a7c92b40ad25efda11e3e17e57c9899a27bfc │ │ └─42914 /pause │ ├─cd42c13d0be44b58d66c65ea94b5fde8c48f16b3c36f36a5d731e2921f3bab70 │ │ └─59468 /pause │ ├─dc9482564c54e29157d98ac178ed2339db1f5651661a17f0533768f409c348ca │ │ └─42828 /pause │ ├─9d81d4ad04462a031afae97fe2f272fb225e518280fad07a4be6516096fdd6bb │ │ └─59584 /pause │ ├─43ec1f98df03547a7d66a28224d2ad60343c1be2ebffb67328826c3e45a21790 │ │ └─47619 /pause │ ├─233217d4b44292104ac90df5c0a5733ef17e71e0b99c3aefe21ad56d5a31f648 │ │ ├─103719 glance-api: uWSGI master │ │ ├─103733 glance-api: uWSGI worker 1 │ │ └─103734 glance-api: uWSGI worker 2 │ ├─00182cedc01af68615bbef6558dec6d692b9d6be1be3fc6e81f100e645016ac4 │ │ └─61778 /pause │ ├─b406c856d743b0fa97ccc18afc074d3994cb19e7db1dfa2b22a3d26fa7ceb6c9 │ │ └─43011 /coredns -conf /etc/coredns/Corefile │ ├─7414397e517e96e53718ec32aab73bd22490407b6ef84c190f92e7d8f8be0c8e │ │ ├─71103 apache2 -DFOREGROUND │ │ ├─71146 (wsgi:k -DFOREGROUND │ │ ├─71147 (wsgi:k -DFOREGROUND │ │ ├─71148 (wsgi:k -DFOREGROUND │ │ ├─71149 (wsgi:k -DFOREGROUND │ │ ├─71150 apache2 -DFOREGROUND │ │ ├─71152 apache2 -DFOREGROUND │ │ ├─71155 apache2 -DFOREGROUND │ │ ├─71157 apache2 -DFOREGROUND │ │ ├─71160 apache2 -DFOREGROUND │ │ ├─71161 apache2 -DFOREGROUND │ │ ├─71162 apache2 -DFOREGROUND │ │ ├─71163 apache2 -DFOREGROUND │ │ ├─71164 apache2 -DFOREGROUND │ │ └─71165 apache2 -DFOREGROUND │ ├─4ed7aaa79b0af8b78d8adc28ddb48ce369d2ac5479b6097461cf62c18013374e │ │ └─65269 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config │ ├─5af8e3dfa6af90fb505b9a63e0dbeb35b0f2b3c54f38b96969c9869e570c6d05 │ │ └─39982 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─4ef0e61c1fd3bd33641c9b6fa74d6f43679fccfd7defe252059b414ec0156627 │ │ └─64994 /pause │ ├─e2e09b5c5b6f865adb4e53b4212a1dadbf86ebbbfb214377c7b476d4d08aeec7 │ │ └─39871 /pause │ ├─b71a5bea7f26e2a0f44e4cfb933e8ca4ce3683f912a281a21861acad19a828a8 │ │ ├─63805 /usr/bin/vector --config-dir /etc/vector/ │ │ └─63838 journalctl --follow --all --show-cursor --output=json --boot --since=2000-01-01 │ ├─77327fe00e765e089ef46d76ffdbe9b9d4baefc07969b83654c8cbf3397da8d4 │ │ └─59514 /pause │ ├─c2deec5032b30d0ad185eb8e85d4663ba802c27ee01d78309b08bc023cbc8a94 │ │ └─62708 /pause │ ├─e0d058f42a6b4ece4854252b8c769f2a054769eb9c3df535e8f85443c7651e14 │ │ └─55658 /secretgen-controller │ ├─e64e25daebf84309e6dde01e906c27a96fe01a89e07fe9ff4fefbcc81ed4a33d │ │ └─59335 /pause │ ├─56cc867d8f073d814a97793d7ed6c2fd2e9ea301737729a29841933df7ebe75a │ │ └─65113 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─fb6d707b2432f255aa178c1884a57ac11506eb3551cdab633d8c41c040b57a80 │ │ └─63120 memcached -m 8192 --extended=modern,track_sizes -I 5m -c 16384 -v -u 11211 │ ├─ffc7821f0268384c644f3d1f37b6c4aac15c18d01a5ceccfbf5ff8cce41508c2 │ │ └─63922 /goldpinger --static-file-path /static │ ├─5df1125c788b6bf861d54605454d2b046a807b35618cd2de91686438dcf175cc │ │ └─39860 /pause │ ├─58782a9bdaa7df348b9441eaedf9b3e445052cf5e43ee31c1d22dd70540b332d │ │ └─55310 /pause │ ├─50690767faefc40746aa217f008128e5546d373ed5a9b2bf6d5ab773d274e01e │ │ └─65165 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.162-253-55-12.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml │ ├─438a6c2e0cd6f4dd7fccc61251eaf771c4a297434dd38d43db5a90034818be23 │ │ └─43143 /pause │ ├─dc05fb6edce2d94d14a082ccdf55e4b8422c9f2790aca6778efa4b0abc7d4674 │ │ └─61704 /pause │ ├─60fdefeb2faa7b5b34a042fc4068ca46410ff39f492923e58b3c4967b745837c │ │ ├─83286 barbiacan-api: uWSGI master │ │ └─83302 barbiacan-api: uWSGI worker 1 │ ├─4debae75448527321a7ecc1171d5af179e217a7f783cf2cc6e0ab5934770be84 │ │ └─43286 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager │ ├─a4a957744e74e09a4b8061882ff98da8dcf512aab38381d0844ed6e8f7299aab │ │ └─39863 /pause │ ├─40540810eb27d736791926aad529eb7c2dcf53cd50e3bff84b25618e19d2805d │ │ └─63586 /pause │ ├─36f107147ea36f9b1b8f14df3411b1ab54632133148cc44605537343369d7379 │ │ └─94377 /pause │ ├─2be49b5ce2cbe75b813442c7d6198b91250cce983e0b4cc2a5fb65f838ddd642 │ │ └─39888 /pause │ ├─58f4c3a9655d68c053210d49e5d993d99540344abde3cac2177a3990757013b8 │ │ └─65604 /bin/oauth2-proxy │ ├─67c44df2b312e89a6c81c67528300c2f7040da840bd7dfc31ad3921da7036d50 │ │ └─63183 memcached -m 1024 --extended=modern,track_sizes -I 5m -c 16384 -v -u 11211 │ ├─8bfd41b676875cee8280d88ca0a9c415a6bfc1f998f067dc7f7ed83682d5f8cf │ │ └─46897 /pause │ ├─caed602b8fea5243ace5b5788ab0ca76f3e352ec0f93416e2bb42e6c1c9fdd25 │ │ └─55230 /pause │ ├─bd4cfcf70e7b0890f23c8504fe2ef1bdf0e019e7f97d191ead0d9741aad59929 │ │ └─62115 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning │ ├─362b0372f03a0824a19b759b9a6a7e2fcf54c4c74aa8e908f32bffa2c2a0e6d1 │ │ └─47727 /opt/percona/peer-list -on-change=/opt/percona/haproxy_add_pxc_nodes.sh -service=percona-xtradb-pxc -protocol=$(PEER_LIST_SRV_PROTOCOL) │ ├─ba0a6fb9008dd9a72a4f42674913e14b3ff4d3bbeb9f71b1cf5a97699f0cb28c │ │ └─50282 /pause │ ├─123145ec86173fa2431f65fd7a4ac09eb05ede7f1fbe45274d7b215ce13769ea │ │ └─85053 /usr/local/bin/rook ceph operator │ ├─7baf6ebd09b70784ea81628b44c5531efb63db6ab75d604af42e75051c0f65c4 │ │ ├─90655 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─90671 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─90677 erl_child_setup 1048576 │ │ ├─90896 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─90897 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─90907 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon │ │ └─91110 /bin/sh -s rabbit_disk_monitor │ ├─2c9f3d2d02d55b93599520983a371cbbc4fb2adcaa64e837ab58ca54b1d916c8 │ │ ├─105072 staffeln-conductor: master process [/var/lib/openstack/bin/staffeln-conductor --config-file /etc/staffeln/staffeln.conf] │ │ ├─105424 staffeln-conductor: Staffeln conductor backup controller worker(0) │ │ └─105426 staffeln-conductor: Staffeln conductor rotation controller worker(0) │ ├─a16d2b08f04520237e212a03e46c906038edb74cfa03281cc80eb4c4cd207110 │ │ ├─76822 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─76839 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─76845 erl_child_setup 1048576 │ │ ├─76903 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─76904 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─76914 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon │ │ └─77096 /bin/sh -s rabbit_disk_monitor │ ├─50ef3f8131e80702e698a5bf9c1428f45fa297ee4c2ffd93afa792044025b34d │ │ └─55553 nfd-gc -gc-interval=1h │ ├─27906fb4c6f75865834122bb334de38aa4b3c5988f8495ed4c10917240e2e958 │ │ └─50835 redis_exporter │ ├─e51588c9962ee40743992d54d3860bd1b29531bf37501a0308b3f6f50592fcb0 │ │ ├─48080 mysqld --wsrep_start_position=bdd74c23-19e7-11f1-b216-1244816258f3:20 │ │ └─48196 /var/lib/mysql/mysql-state-monitor │ ├─4d70e7b903f232dcbe6677f427fd2df3170921760366e846cb60cfe1a11ef0c2 │ │ └─39756 /pause │ ├─9cd70e3b2eec8ae586f5be7b892f5ecbd8b306b4966c323084cab9b7649a75c7 │ │ ├─66099 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─66115 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─66121 erl_child_setup 1048576 │ │ ├─66210 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─66211 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─66221 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon │ │ └─66274 /bin/sh -s rabbit_disk_monitor │ ├─16a01ae464ce67bb94b81e718ad3793b2046a623898a7db6724fc446dc4d263d │ │ └─45685 /pause │ ├─dea27d4517cf05d0c2a313bdf66459912074c3287b85af4aafff52fc0556ec52 │ │ └─47203 /pause │ ├─56a47b88ab33783e58b770a2a716566ad6ef89f8fa811d80f363d43caf068217 │ │ └─39949 kube-apiserver --advertise-address=162.253.55.12 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─3089752e296e08f9634b2c962386faa0f6befa37375055b65b2ad80bab57abc0 │ │ ├─ 45903 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─ 45915 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─ 45973 nginx: master process /usr/bin/nginx -c /etc/nginx/nginx.conf │ │ ├─102687 nginx: worker process │ │ ├─102688 nginx: worker process │ │ └─102689 nginx: cache manager process │ ├─6be833a0cc7db220070fd9826039794d199b67d72710fe2ae923e5606f4a3041 │ │ └─60208 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.35.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 │ ├─0a68526d9abbe7fd40aa3e5abe1bb8cb6da20e72e0289e87186a2096f4c2e386 │ │ ├─62478 nginx: master process nginx -g daemon off; │ │ ├─62506 nginx: worker process │ │ ├─62507 nginx: worker process │ │ ├─62508 nginx: worker process │ │ ├─62509 nginx: worker process │ │ └─62510 nginx: worker process │ ├─caf5b5d5dcb1545c0dd13588f3ddf7b02b25e8be5bd062398c451df437fb02cb │ │ └─103355 /pause │ ├─ab3b868bac3ea3121c9c6ebea55daac6a9db441c92d8c16aee007001334eb1c3 │ │ └─61213 /pause │ ├─f70e0c8e85390fa94c49ed79689d09a46aef47649259ad95bd27ed0ae09968d1 │ │ └─64680 /bin/memcached_exporter --memcached.address=localhost:11211 --web.listen-address=0.0.0.0:9150 │ ├─85a8dad32995c2dd206216098e7d6d9140819feed6bfe4be7b42e877afcf6482 │ │ └─50438 valkey-server *:6379 │ ├─75393f846a92d79b6622292c665e3ad073099023186191116747f7d7e77fc3f6 │ │ └─46510 /pause │ ├─32475549b13f39856b845ff3630d7dabae4bc62d401fe7ae3ee4ba43375a6488 │ │ └─105565 /pause │ ├─1a5f6db11338e8a2cc66b3987e5f69f3422b7cca897dcff80ed64d3ca1c9dda3 │ │ └─50371 /pause │ ├─c5063fa6648d032936419e867d75f20c0701925e7971c18bb51a46bc34e78159 │ │ └─41824 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─58ad8a9c94061201d1dee39e2855982afe3e9e6a916fe0a0ad15674e45855e61 │ │ └─67642 /pause │ ├─82c35fc2d216b9f6ea3d3042f28435c77fa9e86688e0f84b2f17d03a4d91a38e │ │ └─42783 /pause │ ├─ef4cb8defd22ba604398c63b5a5dd4c0d2efcb75e745589ce520f86a5aeaf0ca │ │ └─47025 percona-xtradb-cluster-operator │ ├─c087510da280fa256d31f61d1e3aa51177e9305d3fb06959861a2c201136dfc8 │ │ └─65061 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 │ ├─c139b82abf131382ee77478d2cd539027d5d011c98fc66f9fe0c423f7d2a3d61 │ │ └─64710 /bin/memcached_exporter --memcached.address=localhost:11211 --web.listen-address=0.0.0.0:9150 │ ├─1f5f1b1d9ebc476aaac56fc664513e01765c2d1f4e9103fa5fd62fc5ffc85b53 │ │ ├─60182 python -u /app/sidecar.py │ │ ├─60516 python -u /app/sidecar.py │ │ └─60517 python -u /app/sidecar.py │ ├─72f67ad5653f7c1cd237d2af26a0f78296c0f6c8156f42c4c1661aae907ec4b4 │ │ └─62298 /pause │ ├─23022f7e70eda03b6760d2ba9117b28fb851482f77c204eec8fdca498fae5399 │ │ └─55586 nfd-master -crd-controller=true -metrics=8081 │ ├─93c39438170ede88bf41e1e22c0887a451d777379791172166ab6ff71644ba7f │ │ └─40121 /kube-vip manager │ ├─05eedf22c6a4b995c4918d8577ff1ebbf2fb93f86f5523dddf9da8310db90e00 │ │ └─61154 /pause │ ├─70f4068e268bb883e82d327c7a54952b55d2d77787c2f61ff900b20aa36ff70f │ │ └─60760 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─96f4108bb586a602d01281a628e59534f5f0a8de4b9d8be4b34fd94e65161d2d │ │ └─43372 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.17 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 │ ├─3a8b42abb9f5d60b80981b31f51aa9256be444e4d562493109b3a89c65946ab2 │ │ └─62742 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml │ ├─1f69013a88d5bc7f1bad282c3ab2de80585ac4a40c62c0309a8941bad878ed15 │ │ └─94805 radosgw --fsid=4837cbf8-4f90-4300-b3f6-726c9b9f89b4 --keyring=/etc/ceph/keyring-store/keyring --default-log-to-stderr=true --default-err-to-stderr=true --default-mon-cluster-log-to-stderr=true --default-log-stderr-prefix=debug --default-log-to-file=false --default-mon-cluster-log-to-file=false --mon-host=[v2:10.96.240.200:3300,v1:10.96.240.200:6789] --mon-initial-members=instance --id=rgw.ceph.a --setuser=ceph --setgroup=ceph --foreground --rgw-frontends=beast port=8080 --host=rook-ceph-rgw-ceph-a-699b8bdb59-l74n5 --rgw-mime-types-file=/etc/ceph/rgw/mime.types --rgw-realm=ceph --rgw-zonegroup=ceph --rgw-zone=ceph │ ├─d7efba1bce4fe2f042a79a5e1f33a9a71dcd8487ee5d32da5c056d9f94cff77f │ │ └─65288 /bin/memcached_exporter │ ├─e50c7300abaa2b790b6830397814ed96fbcdfb911c998908f48c9800da16e965 │ │ └─64636 /bin/pushgateway │ ├─c555971bae7cc4059591e559035b9761b60b9b98bcef83576001eb6c97f29cc5 │ │ └─40337 /pause │ ├─883b0cc81217b5411e4ca854e79fdc5b57d987b92de4d7144ec36eb5fc2c46dd │ │ └─46557 /manager --metrics-bind-address=:9782 │ ├─9d05a5963c4a8e224fffb08c1021cd0659e072ff5fb37c17d1ef1fb3c489c376 │ │ └─104918 kubernetes-entrypoint │ ├─7b5b639aa7e4cc14caedb01c9c640c8ab19fd7b4bb878553fa59969cd49d3f30 │ │ └─78248 /pause │ ├─e4f102af7a17cfaa96de525e2d35c7eaa85389105377bc6a6fe959e3166b3a03 │ │ └─43419 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─a71758b95f492e4844c08aed2794b32413d5fd535d953a0aa332a475375e59a7 │ │ └─50530 java -Dkc.config.built=true -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 -XX:FlightRecorderOptions=stackdepth=512 -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:MaxRAMPercentage=70 -XX:MinRAMPercentage=70 -XX:InitialRAMPercentage=50 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dpicocli.disable.closures=true -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true │ ├─a78d9ae619304d20e9600bbb44cd627aa94a14fee3ba3b32a59925b0eff16c4b │ │ └─40001 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─d99b6d86f7a4172fbc714c30b386941e182b78bc36497b8e8f83f40289e594fc │ │ └─61385 /pause │ ├─166d78ffbaace8a516d1c784951530723e8388656a82c8019b6d0c720ba1df94 │ │ └─46696 /manager --metrics-bind-address=:8080 │ ├─3af186198d8a6845539e1676191c8bbe880aa4065492864f9acc8a575fb3b648 │ │ └─84175 /pause │ ├─4015f9b0e0ed67631ef6a7e25d6fc6169db489edbea4eb746116cefec650ad87 │ │ ├─60156 python -u /app/sidecar.py │ │ ├─60471 python -u /app/sidecar.py │ │ └─60475 python -u /app/sidecar.py │ ├─5103fb43e8756ea490965dd582a4b00a86be04b74af202449f4fe45518297c92 │ │ └─41122 /pause │ ├─e10076f59d0718f8321ea1983117423984e1c48e08386e0e0ea8be025c7d2d84 │ │ └─41115 /pause │ ├─c0ca97b810097fa083ad574cf6bef829df41ba7512a4097faf39ebe25fea34b3 │ │ └─76715 /pause │ ├─61836d4f16ada2db7bb2e1678fd23ba306e3ad8c81b77e0c6d303598dbfc06f8 │ │ └─47972 /bin/mysqld_exporter --mysqld.username=monitor --collect.info_schema.processlist │ ├─a48ecb3ed532919cc13bc94b8a981a471e020ebe0019cb894d5feabe06ddbc33 │ │ └─104887 /pause │ ├─cde1a348184b2044f0fd98c2251db859a0989db484df58334f5bc93276e9f692 │ │ └─55256 /pause │ ├─d69d6da050bf27a2215ff6c5f939f065e89fa6ace2c1f1d6e094976d9b227151 │ │ └─104773 /pause │ ├─16920a135e9c98172fc45c3a3900642a4494144cdb053eb16806e8d144c1fd2b │ │ └─43210 /pause │ ├─ff41c9a1da5b70219ca56cf50de9833b45b6b5ebb501f8fee83b48a6f0607d19 │ │ └─40368 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─12067e2090837d9ff4b8983b0b75fc4991d1f2e83ccec88bdb2872f9dbef1bb4 │ │ └─61929 /pause │ └─a91d9b4ae9e1dc6687f50c7588f480180139f86857db2d02ad03342eff993fb3 │ └─95558 /usr/bin/python3 -s /usr/bin/ceph-crash ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount