Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ ├─user-1002.slice │ │ ├─user@1002.service … │ │ │ ├─app.slice │ │ │ └─init.scope │ │ │ ├─21537 /lib/systemd/systemd --user │ │ │ └─21538 (sd-pam) │ │ └─session-22.scope │ │ ├─21528 sshd: cephadm [priv] │ │ └─21554 sshd: cephadm@notty │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─712 /lib/systemd/systemd --user │ │ └─713 (sd-pam) │ ├─session-4.scope │ │ ├─1462 sshd: zuul [priv] │ │ └─1475 sshd: zuul@notty │ └─session-1.scope │ ├─ 709 sshd: zuul [priv] │ ├─ 727 sshd: zuul@notty │ ├─ 872 /usr/bin/python3 │ ├─ 1466 ssh: /home/zuul/.ansible/cp/199.19.213.35-22-zuul [mux] │ ├─108000 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-yyouicjbzkkgwzdefdpivmgijxdmotbg ; /usr/bin/python3'"'"' && sleep 0' │ ├─108001 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-yyouicjbzkkgwzdefdpivmgijxdmotbg ; /usr/bin/python3' && sleep 0 │ ├─108002 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-yyouicjbzkkgwzdefdpivmgijxdmotbg ; /usr/bin/python3 │ ├─108003 /bin/sh -c echo BECOME-SUCCESS-yyouicjbzkkgwzdefdpivmgijxdmotbg ; /usr/bin/python3 │ ├─108004 /usr/bin/python3 │ ├─108005 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─108007 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 42 nofb ├─system.slice │ ├─docker-caf47af979dc330122fa525c249ff71fc1f94abaafa1074ea0217d4fad983d27.scope … │ │ ├─24641 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─24653 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─containerd.service … │ │ ├─ 9802 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─12194 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 1c54a59515357934846c576f78482bcbb96bc7270d20e1ed2031f38991e556b3 -address /run/containerd/containerd.sock │ │ ├─12556 /usr/bin/containerd-shim-runc-v2 -namespace moby -id a027f217b080ad04bd53223f18eed74542873c9eab32a17b7854c6d777ce7542 -address /run/containerd/containerd.sock │ │ ├─19190 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 1160137bd36c1dd8ff0c0fac63f046b2ff31e9d3bfda81155f096c8292a92bb8 -address /run/containerd/containerd.sock │ │ ├─24616 /usr/bin/containerd-shim-runc-v2 -namespace moby -id caf47af979dc330122fa525c249ff71fc1f94abaafa1074ea0217d4fad983d27 -address /run/containerd/containerd.sock │ │ ├─29243 /usr/bin/containerd-shim-runc-v2 -namespace moby -id c84c89350b6b0c5905df3e739704519a5cca9b88deb1598b5b5edbdccd96c537 -address /run/containerd/containerd.sock │ │ ├─34013 /usr/bin/containerd-shim-runc-v2 -namespace moby -id d3081ceadb7c5db3f08e380144a0e86cb64b9dff7e0f74e2e9dd565f76e82f8d -address /run/containerd/containerd.sock │ │ ├─39690 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d24da7cdbdf1e8add6112e8dc81456789cb85504186b718861b97305c34a465f -address /run/containerd/containerd.sock │ │ ├─39699 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ef56a297ebf9e753587282943c00e631c5654ff48fa17c29984434c8dded71aa -address /run/containerd/containerd.sock │ │ ├─39748 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0db8e941f2b59f63cad827f6840c5822c19ab04166e1d6a5e86cf8456a6d7017 -address /run/containerd/containerd.sock │ │ ├─39756 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4302764d830e3be91b701812927e8d2f3bd82b3cd75c8682951b4f929bbc8815 -address /run/containerd/containerd.sock │ │ ├─39783 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 757b72c652b62f81747b84736b85d8860c1009395a566269d59b4bb4cd545935 -address /run/containerd/containerd.sock │ │ ├─40289 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6152edc67337a564511c14ee7bd424f1828212976939d77a5f5db062f5081bc0 -address /run/containerd/containerd.sock │ │ ├─41044 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5b71d75ac276e116bfb55a276cd3e119201ced3f860762bfdd34a6dbe3c00094 -address /run/containerd/containerd.sock │ │ ├─41093 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id fba78db790a1a439a14a72cdb114594860f8f37ec5b1aab4f247b5139b1db87c -address /run/containerd/containerd.sock │ │ ├─42530 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8346950b982ef71762928bc231f607d71e1e292bc39b097f7e1feb5291b77129 -address /run/containerd/containerd.sock │ │ ├─42977 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7dac5759f5e53f1494f6ee5d9e86fcf06cc0a3c720df2488e216008c6f51fddb -address /run/containerd/containerd.sock │ │ ├─43014 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b5cd748ef22749596730f947df6c8756794992b546ae748e7a21d8d8ab635baf -address /run/containerd/containerd.sock │ │ ├─43494 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b2cfc6decd65282dfd1ef9ff73bad742dd2d548ddf6ed3fb8750c06fe7c8090a -address /run/containerd/containerd.sock │ │ ├─44057 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4cbdc12e03668868c42525031343ab401fa9d5c5ac41993c7290af34b281b718 -address /run/containerd/containerd.sock │ │ ├─44107 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a0579b07326e7c4cecbf828c861f1a15078f36b16d62f63b20b49d0d22c9e849 -address /run/containerd/containerd.sock │ │ ├─44151 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c67e31c26439a144411dff784a6201533d34f578e462fb4f54a1c5de92e10248 -address /run/containerd/containerd.sock │ │ ├─46229 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 117337eaf1fa3fb22b2650a620f2cf0b12f7696cd2b0c7a7028274a0ae98e1f5 -address /run/containerd/containerd.sock │ │ ├─46315 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 031c84e4c6cf82e09d07fbbbd5af18e5ce83d1528568ac633b3cc2bc6cc21da1 -address /run/containerd/containerd.sock │ │ ├─47043 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9091c3f5a50646ff914811e177c9d138829e91b82e386e884f67e7abd05f063d -address /run/containerd/containerd.sock │ │ ├─47142 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 21e0bb350e05431a36afc5fb3b60f25a4149a953d3525367e61535f6e6af82df -address /run/containerd/containerd.sock │ │ ├─47534 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 475360426af823252f01937a7e81fa9a1169598cc7988a3f7efa2dc3ca887c01 -address /run/containerd/containerd.sock │ │ ├─47879 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 185362e39d13d053cdfe116eb81519c051296f7d6509f5f4392da7b943278af2 -address /run/containerd/containerd.sock │ │ ├─48385 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 2a60d89d0716f6a6b7d3363fa1b91240490ef551dbd6718def075f89aac7bd12 -address /run/containerd/containerd.sock │ │ ├─50334 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5f7576fe3b7ed810102c4d436584a5e345cd311902dffb4f9c7e2cd5601ba950 -address /run/containerd/containerd.sock │ │ ├─52744 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 40a0c1f7d6f043877e0d4ee66f138b375b9571b1893c4e14c217e3e9c15b8bc2 -address /run/containerd/containerd.sock │ │ ├─53404 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7dbf7a6f02aa1f04084cb3405d9ad8c8c47450184d60f07cf874d2f926c5fd66 -address /run/containerd/containerd.sock │ │ ├─53456 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 30bcf443fb8b7274908b8c8dd211f7f943e1379d4112032fa6bee960adc01066 -address /run/containerd/containerd.sock │ │ ├─53594 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 473c805f61fd4a3d908e1b55abbe5fbb26c42b9ff9619e8b5a6b377ff3c3f8f5 -address /run/containerd/containerd.sock │ │ ├─57182 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 088d51082a5e57da5cf316db45686508823de94b1808d8503abc3604de29612c -address /run/containerd/containerd.sock │ │ ├─57228 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5b0440340e206ac4ba2d05f9adbfa697d06c096370179d92ab0ad878c06bbc34 -address /run/containerd/containerd.sock │ │ ├─57239 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id eeb9d8426366b6f4e5b150e2aa8816b883ec94b012ddb3813fafe1ba74725619 -address /run/containerd/containerd.sock │ │ ├─57332 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 038c7416480706eb5ec0b03a643f204b8b4b4cefd44341f6f158c40cfb3b74ac -address /run/containerd/containerd.sock │ │ ├─58685 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 88724d3759568ed5c5053a7db0139372e8549282d22440cafb122b9d5567d03d -address /run/containerd/containerd.sock │ │ ├─58853 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a6b1bfb69b207a7de8f6bba48f9113cf3aa4522cdd7a14246b769335c71cb2f5 -address /run/containerd/containerd.sock │ │ ├─58996 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 54f6ad4518d30f950e52563bb709a67f89028c9195c57d9d13c735884f585918 -address /run/containerd/containerd.sock │ │ ├─59411 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 81e0114f10721e3c4777b019f6b6b69918d71b85e22322ec0858210df53e9b8e -address /run/containerd/containerd.sock │ │ ├─59549 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ad7997f8fbed705d3af106d07dad572cfeba5b9dbfb2fc4437f37131c153e23d -address /run/containerd/containerd.sock │ │ ├─60170 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 03178f3ee28002b07ded9a976a0e53c4a535488967a97f6f07ccc4f58c471831 -address /run/containerd/containerd.sock │ │ ├─60223 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 27eab878c3ad722c7c8b5a710897fd504a060632335553dfdd616e56570eb313 -address /run/containerd/containerd.sock │ │ ├─61076 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 91b55e102933ae82e002f0314f4b6005d076e66c2ab4041325051cd3e8adbd25 -address /run/containerd/containerd.sock │ │ ├─62123 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c08d70839f096ec00ce5f1be50f1747a0178851027f864f34228f33d334c82b8 -address /run/containerd/containerd.sock │ │ ├─64511 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7b4d71c5db7a1d9dcda90c5b6a3195b5a9fa9356a776dfa2ce63eb216dfccfb3 -address /run/containerd/containerd.sock │ │ ├─71463 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a6b538a03975764c2aec1b94a04059c767811e90a3a0cdac85ff02e34323c8dd -address /run/containerd/containerd.sock │ │ ├─72765 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b9acfb5e91bdbe7ab60d6fbd5811faa48cbf59b6f7b14606aa73aa46a0d1fc7a -address /run/containerd/containerd.sock │ │ ├─77591 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 839160878461f02fc602cb32780afd4c7395b8213280bafee760650baa3ccd9b -address /run/containerd/containerd.sock │ │ ├─82574 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4defbd52081e011be222c4affdc625b6926dff28ef8b20dc37d8be5c3eb7a6e8 -address /run/containerd/containerd.sock │ │ ├─90504 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id bcab94ad0851a7da4be373a6c674ea11d69381137e953ffab0e1e4411770346e -address /run/containerd/containerd.sock │ │ ├─90567 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e89dd020ba90525036feaec53c4e18d135a049f46c1526b528cd55d4dcce4d93 -address /run/containerd/containerd.sock │ │ ├─97903 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8eed06ca44932e8cd81209675bb582ca9f38c06e6f761211fd3e943c6654a506 -address /run/containerd/containerd.sock │ │ ├─99102 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ac8d68bc0e29a4c4e8b472db97083c2adbd2c38a79ccb1f2bdada390f2066ee4 -address /run/containerd/containerd.sock │ │ ├─99224 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b36210e44a8308c06d68e75f64442968f845bd529a79ce2b96a0a2612a7a97c9 -address /run/containerd/containerd.sock │ │ └─99905 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1b813b9a174f09427339c796c4f666335efcd840176f39e61ad4caaa17d8aff6 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1196 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─531 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─398 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─632 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─system-ceph\x2d4837cbf8\x2d4f90\x2d4300\x2db3f6\x2d726c9b9f89b4.slice │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.1.service │ │ │ ├─28951 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/unit.run │ │ │ └─29225 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-1 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1:/var/lib/ceph/osd/ceph-1:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@crash.instance.service │ │ │ ├─19127 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/unit.run │ │ │ └─19148 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-crash --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-crash-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/config:/etc/ceph/ceph.conf:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/keyring:/etc/ceph/ceph.client.crash.instance.keyring quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n client.crash.instance │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.0.service │ │ │ ├─24337 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/unit.run │ │ │ └─24599 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-0 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0:/var/lib/ceph/osd/ceph-0:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@osd.2.service │ │ │ ├─33718 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/unit.run │ │ │ └─33997 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-2 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2:/var/lib/ceph/osd/ceph-2:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ ├─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mon.instance.service │ │ │ ├─12106 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/unit.run │ │ │ └─12138 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mon-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance:/var/lib/ceph/mon/ceph-instance:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev quay.io/ceph/ceph:v18.2.7 -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ │ └─ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4@mgr.instance.dxonjj.service │ │ ├─12471 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.dxonjj/unit.run │ │ └─12499 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mgr --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mgr-instance-dxonjj --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.dxonjj:/var/lib/ceph/mgr/ceph-instance.dxonjj:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.dxonjj/config:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v18.2.7 -n mgr.instance.dxonjj -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker-d3081ceadb7c5db3f08e380144a0e86cb64b9dff7e0f74e2e9dd565f76e82f8d.scope … │ │ ├─34039 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─34051 /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker.service … │ │ └─10170 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock │ ├─polkit.service │ │ └─651 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─626 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─docker-a027f217b080ad04bd53223f18eed74542873c9eab32a17b7854c6d777ce7542.scope … │ │ ├─12579 /sbin/docker-init -- /usr/bin/ceph-mgr -n mgr.instance.dxonjj -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─12599 /usr/bin/ceph-mgr -n mgr.instance.dxonjj -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─kubelet.service │ │ └─40149 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.19.213.35 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─364 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─656 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1397 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─chrony.service │ │ ├─10676 /usr/sbin/chronyd -F 1 │ │ └─10677 /usr/sbin/chronyd -F 1 │ ├─docker-1160137bd36c1dd8ff0c0fac63f046b2ff31e9d3bfda81155f096c8292a92bb8.scope … │ │ ├─19217 /sbin/docker-init -- /usr/bin/ceph-crash -n client.crash.instance │ │ └─19235 /usr/bin/python3 -s /usr/bin/ceph-crash -n client.crash.instance │ ├─docker.socket │ ├─docker-c84c89350b6b0c5905df3e739704519a5cca9b88deb1598b5b5edbdccd96c537.scope … │ │ ├─29266 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ │ └─29279 /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug │ ├─docker-1c54a59515357934846c576f78482bcbb96bc7270d20e1ed2031f38991e556b3.scope … │ │ ├─12218 /sbin/docker-init -- /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ │ └─12238 /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true │ ├─systemd-resolved.service │ │ └─406 /lib/systemd/systemd-resolved │ ├─system-lvm2\x2dpvscan.slice │ ├─dbus.service │ │ └─622 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─631 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─628 /lib/systemd/systemd-logind ├─k8s.io │ ├─8a6d1b0a5cdab25ceded237ad9128a1e402cee6080d9320bcda9a207d44be204 │ │ ├─90889 /bin/bash -x -e -m -c CEPH_CLIENT_ID=ceph-client.rgw.ceph.a PERIODICITY=daily LOG_ROTATE_CEPH_FILE=/etc/logrotate.d/ceph LOG_MAX_SIZE=500M # edit the logrotate file to only rotate a specific daemon log # otherwise we will logrotate log files without reloading certain daemons # this might happen when multiple daemons run on the same machine sed -i "s|*.log|$CEPH_CLIENT_ID.log|" "$LOG_ROTATE_CEPH_FILE" # replace default daily with given user input sed --in-place "s/daily/$PERIODICITY/g" "$LOG_ROTATE_CEPH_FILE" if [ "$LOG_MAX_SIZE" != "0" ]; then # adding maxsize $LOG_MAX_SIZE at the 4th line of the logrotate config file with 4 spaces to maintain indentation sed --in-place "4i \ \ \ \ maxsize $LOG_MAX_SIZE" "$LOG_ROTATE_CEPH_FILE" fi while true; do # we don't force the logrorate but we let the logrotate binary handle the rotation based on user's input for periodicity and size logrotate --verbose "$LOG_ROTATE_CEPH_FILE" sleep 15m done │ │ └─90905 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 15m │ ├─c636547627b28b008c01099863bda59092b383859b10d85120199dad972e6d35 │ │ └─50497 java -Dkc.config.built=true -Xms64m -Xmx512m -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --auto-build --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true │ ├─1d7acf34ee1f945f089698faab4ca3c2c3f82189e43028143bfed336c2bea7c1 │ │ └─39978 kube-apiserver --advertise-address=199.19.213.35 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─b36210e44a8308c06d68e75f64442968f845bd529a79ce2b96a0a2612a7a97c9 │ │ └─99248 /pause │ ├─ad7997f8fbed705d3af106d07dad572cfeba5b9dbfb2fc4437f37131c153e23d │ │ └─59573 /pause │ ├─4cbdc12e03668868c42525031343ab401fa9d5c5ac41993c7290af34b281b718 │ │ └─44084 /pause │ ├─5b0440340e206ac4ba2d05f9adbfa697d06c096370179d92ab0ad878c06bbc34 │ │ └─57300 /pause │ ├─78f373e88a38397e8908d931ddcff7edba8c266f3893c085a402a9ec8b908bf8 │ │ ├─77044 barbiacan-api: uWSGI master │ │ └─77061 barbiacan-api: uWSGI worker 1 │ ├─c4274efe5503e1560f97c4f0b1cc73b4a1c0ec59e3e3254a9b3bf9bb4795819b │ │ └─78375 /usr/local/bin/rook ceph operator │ ├─fdba64896bbf037d2a77be2e12053e228c55010d45fd9ea29acb00251e277f00 │ │ └─40338 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─ef56a297ebf9e753587282943c00e631c5654ff48fa17c29984434c8dded71aa │ │ └─39797 /pause │ ├─694e7635060a993efaaab62a1cff4da99a18221e74ad5b5d527b9f1be22e9d49 │ │ └─53812 nfd-master -featurerules-controller=true │ ├─e89dd020ba90525036feaec53c4e18d135a049f46c1526b528cd55d4dcce4d93 │ │ └─90592 /pause │ ├─6152edc67337a564511c14ee7bd424f1828212976939d77a5f5db062f5081bc0 │ │ └─40312 /pause │ ├─090418a6d0ff064f73be49fd9bf890b58795698540f13427ed617c4fd6b6c37a │ │ └─58585 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments │ ├─5b71d75ac276e116bfb55a276cd3e119201ced3f860762bfdd34a6dbe3c00094 │ │ └─41067 /pause │ ├─03178f3ee28002b07ded9a976a0e53c4a535488967a97f6f07ccc4f58c471831 │ │ └─60194 /pause │ ├─2e181c249e04e68591f84d36246aca82c293fc9ca5b5b7f0bf402228aa212e5b │ │ └─62467 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 │ ├─54f6ad4518d30f950e52563bb709a67f89028c9195c57d9d13c735884f585918 │ │ └─59020 /pause │ ├─8b2964fa8b7c53446cac512c717ced402af9d585ce26a415263f3ced9d2ce27b │ │ └─60312 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─8127451f8c9e6df62f81897dc658602866bfa6092b4d04ef0f07d4cb95796d07 │ │ ├─48277 haproxy -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ │ └─49756 haproxy -sf 11 -x sockpair@3 -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ ├─475360426af823252f01937a7e81fa9a1169598cc7988a3f7efa2dc3ca887c01 │ │ └─47558 /pause │ ├─a73d89ad2d3032eabce52a3951eda6195ed47475a777b48f2b88feace447955e │ │ └─41318 /kube-vip manager │ ├─4302764d830e3be91b701812927e8d2f3bd82b3cd75c8682951b4f929bbc8815 │ │ └─39837 /pause │ ├─bb8f168b3a7ec51c60fb6f337240a40a48d858b5080ffc2096a0746d3bca5dd3 │ │ └─53882 /secretgen-controller │ ├─9c1ffddfe07fcbae0fd1cd9eda9b1d7e0d3abb2216fd8d6c13f466d955a7bda9 │ │ └─63221 /bin/oauth2-proxy │ ├─f5808040f4f7f10721264a6ca589bf147fec5d90642deb4995fe22d87a952216 │ │ └─59820 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning │ ├─fba78db790a1a439a14a72cdb114594860f8f37ec5b1aab4f247b5139b1db87c │ │ └─41115 /pause │ ├─91905cff9096d22ead3427b596834be0806273e56866c3c350e5de827937ce39 │ │ └─48808 /bin/mysqld_exporter --collect.info_schema.processlist │ ├─0db8e941f2b59f63cad827f6840c5822c19ab04166e1d6a5e86cf8456a6d7017 │ │ └─39856 /pause │ ├─7dbf7a6f02aa1f04084cb3405d9ad8c8c47450184d60f07cf874d2f926c5fd66 │ │ └─53431 /pause │ ├─c67e31c26439a144411dff784a6201533d34f578e462fb4f54a1c5de92e10248 │ │ └─44175 /pause │ ├─19285483ce2adb39a48c5fbe21ca10a0de3519634852de44e38045687f09f7c4 │ │ ├─58197 python -u /app/sidecar.py │ │ ├─58538 python -u /app/sidecar.py │ │ └─58539 python -u /app/sidecar.py │ ├─8dc0b0204133b43f49e170dcb83db737250749a4eceb3358a8617beee5232a0c │ │ ├─63096 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─63131 /usr/local/lib/erlang/erts-12.3.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /usr/local/lib/erlang -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false │ │ ├─63157 erl_child_setup 1048576 │ │ ├─63287 /usr/local/lib/erlang/erts-12.3.2/bin/epmd -daemon │ │ ├─63353 inet_gethost 4 │ │ └─63355 inet_gethost 4 │ ├─7a6b944ebff94384934371fa004a6b3470cab7914c0c3e80abbf81fdf6ec1426 │ │ ├─58259 python -u /app/sidecar.py │ │ ├─58574 python -u /app/sidecar.py │ │ └─58575 python -u /app/sidecar.py │ ├─cdf1dbb83135cc00953b45581da664ecfad951486aeadcf0ed2cbfbac63df811 │ │ └─41374 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─0a63b0800c1caa398d77acdfb1b04be7d59cdf6f2924eb6c37a7efb19055fb75 │ │ └─62510 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─27eab878c3ad722c7c8b5a710897fd504a060632335553dfdd616e56570eb313 │ │ └─60247 /pause │ ├─91b55e102933ae82e002f0314f4b6005d076e66c2ab4041325051cd3e8adbd25 │ │ └─61106 /pause │ ├─a6b1bfb69b207a7de8f6bba48f9113cf3aa4522cdd7a14246b769335c71cb2f5 │ │ └─58892 /pause │ ├─e835ad2c9c60d0d766df8c1c2f4e559e5cc48b2a2aea2cc11d04093cf3260269 │ │ └─47348 /manager --metrics-bind-address=:8080 │ ├─e8d9eeb48a1fee7eee906f0e76aabcf7fbe6b743934d012a09ecd45bf4c9bfb5 │ │ └─41533 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─4defbd52081e011be222c4affdc625b6926dff28ef8b20dc37d8be5c3eb7a6e8 │ │ └─82598 /pause │ ├─5ac47eab58b40fcf74341d1dab8eb266263c682c492c905d6672ed6d1dbb40d5 │ │ └─62606 memcached -v -p 11211 -U 0 -c 8192 -m 1024 │ ├─eeb9d8426366b6f4e5b150e2aa8816b883ec94b012ddb3813fafe1ba74725619 │ │ └─57317 /pause │ ├─fe457af81bd1fc982e0648e64d98b200f5c9cf190731422cee4959cce71fbc66 │ │ └─62314 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config │ ├─031c84e4c6cf82e09d07fbbbd5af18e5ce83d1528568ac633b3cc2bc6cc21da1 │ │ └─46337 /pause │ ├─b9acfb5e91bdbe7ab60d6fbd5811faa48cbf59b6f7b14606aa73aa46a0d1fc7a │ │ └─72788 /pause │ ├─b2cfc6decd65282dfd1ef9ff73bad742dd2d548ddf6ed3fb8750c06fe7c8090a │ │ └─43519 /pause │ ├─66391793a5575ca381000c2403d755e31dd9c5e1497658f585c5acad42d949b1 │ │ └─57715 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.34.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 │ ├─b5cd748ef22749596730f947df6c8756794992b546ae748e7a21d8d8ab635baf │ │ └─43039 /pause │ ├─8346950b982ef71762928bc231f607d71e1e292bc39b097f7e1feb5291b77129 │ │ └─42554 /pause │ ├─bc80f3df17bc626fbfb54cd08c29b96e15dc22ce3db97f6b4bb874ce47af2918 │ │ └─63442 /bin/memcached_exporter │ ├─1cf93f4a9e69017ed3123d64ce9b61db486c60e139ceda98b55000e7bc2e418f │ │ └─47672 percona-xtradb-cluster-operator │ ├─1b813b9a174f09427339c796c4f666335efcd840176f39e61ad4caaa17d8aff6 │ │ └─99927 /pause │ ├─40a0c1f7d6f043877e0d4ee66f138b375b9571b1893c4e14c217e3e9c15b8bc2 │ │ └─52768 /pause │ ├─a0579b07326e7c4cecbf828c861f1a15078f36b16d62f63b20b49d0d22c9e849 │ │ └─44137 /pause │ ├─435df800b55fb8939fc882c8fd90e1a54ec327e288c2fdd1f56e47136e3f89c6 │ │ └─53762 nfd-worker --server=node-feature-discovery-master:8080 │ ├─8c1bdf2fbf03766aa0d5e1ddc1815a0cb607562a07633589180f98facf679118 │ │ └─53115 /kubernetes-entrypoint │ ├─5972d59ccbcd15f36fbcbb8381f0a9d38281cfeaf953c75dad951912f1c3277a │ │ ├─98209 glance-api: uWSGI master │ │ ├─98222 glance-api: uWSGI worker 1 │ │ └─98223 glance-api: uWSGI worker 2 │ ├─185362e39d13d053cdfe116eb81519c051296f7d6509f5f4392da7b943278af2 │ │ └─47902 /pause │ ├─a4f53d88f7a94f2b0a26fc40abdf847657d784355a49c052662abf129220cb0a │ │ └─61868 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml │ ├─2a60d89d0716f6a6b7d3363fa1b91240490ef551dbd6718def075f89aac7bd12 │ │ └─48409 /pause │ ├─e7ee864ad9e02a8eeebb91dd9b4e6ccec5246017a2a38e8ffc0d1072cb980382 │ │ ├─46560 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─46572 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─46719 nginx: master process /usr/bin/nginx -c /etc/nginx/nginx.conf │ │ ├─94572 nginx: worker process │ │ ├─94573 nginx: worker process │ │ └─94574 nginx: cache manager process │ ├─ea3c35bce144acb19a3712592036af8f2c40f6943d1a1d29ee594892b9713f5e │ │ └─46655 /server │ ├─53b8c376f08af1c08196f614fb6df1b3a780d39380e72d9d1e774427eb762e9d │ │ └─62272 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.199-19-213-35.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml │ ├─d2b9673cd756971b56f6ac0f1741178697887e216e35abb9dd4817eb2dc68068 │ │ └─90773 radosgw --fsid=4837cbf8-4f90-4300-b3f6-726c9b9f89b4 --keyring=/etc/ceph/keyring-store/keyring --log-to-stderr=true --err-to-stderr=true --mon-cluster-log-to-stderr=true --log-stderr-prefix=debug --default-log-to-file=false --default-mon-cluster-log-to-file=false --mon-host=[v2:10.96.240.200:3300,v1:10.96.240.200:6789] --mon-initial-members=instance --id=rgw.ceph.a --setuser=ceph --setgroup=ceph --foreground --rgw-frontends=beast port=8080 --host=rook-ceph-rgw-ceph-a-67fd8975c6-hft4w --rgw-mime-types-file=/etc/ceph/rgw/mime.types --rgw-realm=ceph --rgw-zonegroup=ceph --rgw-zone=ceph │ ├─c08d70839f096ec00ce5f1be50f1747a0178851027f864f34228f33d334c82b8 │ │ └─62147 /pause │ ├─684e74d9919867ff50adb707c949fa55dc6398186055b868fad27bb6247add27 │ │ └─62425 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.199-19-213-35.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml │ ├─30bcf443fb8b7274908b8c8dd211f7f943e1379d4112032fa6bee960adc01066 │ │ └─53479 /pause │ ├─3ed8d29a4bd7d4270636514325f85dd07baa5119acef43ac0ae80919f59a1386 │ │ └─39964 etcd --advertise-client-urls=https://199.19.213.35:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.19.213.35:2380 --initial-cluster=instance=https://199.19.213.35:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.19.213.35:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.19.213.35:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─0ed5f46f273d267a720da42b2aa7dc75937dfe1c669b3926648878b7bf9d3dac │ │ ├─67279 apache2 -DFOREGROUND │ │ ├─67299 (wsgi:k -DFOREGROUND │ │ ├─67300 (wsgi:k -DFOREGROUND │ │ ├─67301 (wsgi:k -DFOREGROUND │ │ ├─67302 (wsgi:k -DFOREGROUND │ │ ├─67303 apache2 -DFOREGROUND │ │ ├─67307 apache2 -DFOREGROUND │ │ ├─67309 apache2 -DFOREGROUND │ │ ├─67312 apache2 -DFOREGROUND │ │ ├─67313 apache2 -DFOREGROUND │ │ ├─67314 apache2 -DFOREGROUND │ │ ├─67315 apache2 -DFOREGROUND │ │ ├─67316 apache2 -DFOREGROUND │ │ ├─67317 apache2 -DFOREGROUND │ │ └─67318 apache2 -DFOREGROUND │ ├─c693dde15803dec3468d3139e0badc213d41f704611faf3fddbf2d251f670d54 │ │ └─48434 /usr/bin/peer-list -on-change=/usr/bin/add_pxc_nodes.sh -service=percona-xtradb-pxc │ ├─ac8d68bc0e29a4c4e8b472db97083c2adbd2c38a79ccb1f2bdada390f2066ee4 │ │ └─99127 /pause │ ├─81e0114f10721e3c4777b019f6b6b69918d71b85e22322ec0858210df53e9b8e │ │ └─59436 /pause │ ├─bcab94ad0851a7da4be373a6c674ea11d69381137e953ffab0e1e4411770346e │ │ └─90526 /pause │ ├─088d51082a5e57da5cf316db45686508823de94b1808d8503abc3604de29612c │ │ └─57295 /pause │ ├─78dc5db1d8d897a335abbdd5df0b8e29d042f3ed49730ebc5f085411427bf7ed │ │ └─43766 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─7dac5759f5e53f1494f6ee5d9e86fcf06cc0a3c720df2488e216008c6f51fddb │ │ └─43001 /pause │ ├─839160878461f02fc602cb32780afd4c7395b8213280bafee760650baa3ccd9b │ │ └─77614 /pause │ ├─3feec6b89d28da9e3d62baa17fb68ea1b315c7e8cc8004b3409c0fcd2c894ebc │ │ └─61962 /usr/bin/loki -config.file=/etc/loki/config/config.yaml -target=all │ ├─0bda649d4cc4b847e68ddb894ab83ba7c014fc5b0bd00f7a16b2f0008ded6276 │ │ └─99195 kubernetes-entrypoint │ ├─6ee2907989d610c1081b09cbfdf4c5ea1c6945e165b8b1808eed8529eeb529fe │ │ └─61558 /goldpinger --static-file-path /static │ ├─462acaeedb14c3dcb903763f11695b3bd2732e273204bbf80becc769f5d6753d │ │ └─63239 /bin/oauth2-proxy │ ├─fc0f234a185579bf6cf2113ebe477e3c1334a36b67a0dad7af1c0c8b4c684a98 │ │ └─62219 /bin/pushgateway │ ├─0e841056e147ad7ef5df6a8aaca7109f3de0b1a8f39eb29accb29049bf4b6df2 │ │ ├─82784 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─82801 /usr/local/lib/erlang/erts-12.3.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /usr/local/lib/erlang -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false │ │ ├─82808 erl_child_setup 1048576 │ │ ├─82863 /usr/local/lib/erlang/erts-12.3.2/bin/epmd -daemon │ │ ├─82914 inet_gethost 4 │ │ └─82915 inet_gethost 4 │ ├─80a55730439291319d72c67398e7ab2c6cb9ceddca9f31e3c7931e0a5d77433b │ │ ├─99695 staffeln-conductor: master process [/var/lib/openstack/bin/staffeln-conductor --config-file /etc/staffeln/staffeln.conf] │ │ ├─99811 staffeln-conductor: Staffeln conductor backup controller worker(0) │ │ └─99813 staffeln-conductor: Staffeln conductor rotation controller worker(0) │ ├─ad0b2313226c356c5ce1de819ada4c01b797584b6adf2605b4e1ab4ea8321c90 │ │ └─48597 mysqld --wsrep_start_position=71697b7a-3339-11f1-b9e4-16dcf758f51c:22 │ ├─738f08cf35faaec80c78293f3137ea68cca37bdb4cce42b01f1bb3119ab4a665 │ │ └─47300 /manager --metrics-bind-address=:9782 │ ├─473c805f61fd4a3d908e1b55abbe5fbb26c42b9ff9619e8b5a6b377ff3c3f8f5 │ │ └─53618 /pause │ ├─a6b538a03975764c2aec1b94a04059c767811e90a3a0cdac85ff02e34323c8dd │ │ └─71486 /pause │ ├─8eed06ca44932e8cd81209675bb582ca9f38c06e6f761211fd3e943c6654a506 │ │ └─97927 /pause │ ├─c50a535dc726f3705a78bf0c77268022ef701fc963b0302195f743cde941a1c0 │ │ └─39934 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─aa6e2f66cd5448385697bcfbd865927a74621f321932fc5618861ad09020e1bf │ │ ├─71620 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─71637 /usr/local/lib/erlang/erts-12.3.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /usr/local/lib/erlang -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false │ │ ├─71644 erl_child_setup 1048576 │ │ ├─71707 /usr/local/lib/erlang/erts-12.3.2/bin/epmd -daemon │ │ ├─71758 inet_gethost 4 │ │ └─71759 inet_gethost 4 │ ├─117337eaf1fa3fb22b2650a620f2cf0b12f7696cd2b0c7a7028274a0ae98e1f5 │ │ └─46251 /pause │ ├─851ee146600bd93322f38b632ad298fb1fbbd8e222bcc340ed55daa189a71f03 │ │ ├─61347 /usr/bin/vector --config-dir /etc/vector/ │ │ └─61384 journalctl --follow --all --show-cursor --output=json --boot --since=2000-01-01 │ ├─bce7df9da7b18a66ad39e9123132111ca14da790fae6925661ab617ea53226e5 │ │ └─91609 /usr/libexec/platform-python -s /usr/bin/ceph-crash │ ├─757b72c652b62f81747b84736b85d8860c1009395a566269d59b4bb4cd545935 │ │ └─39857 /pause │ ├─8011373e0eebaa0fc50c3a9aee56db0cf36e93d8daf5a29e07547fe72d295d82 │ │ ├─41656 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─42039 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─66ceec4941577867b5fd1ef4a941f10160bdd52bd3d965b3cb4e3fd73b997b46 │ │ ├─60896 nginx: master process nginx -g daemon off; │ │ ├─60923 nginx: worker process │ │ ├─60924 nginx: worker process │ │ ├─60925 nginx: worker process │ │ ├─60926 nginx: worker process │ │ └─60927 nginx: worker process │ ├─88724d3759568ed5c5053a7db0139372e8549282d22440cafb122b9d5567d03d │ │ └─58708 /pause │ ├─038c7416480706eb5ec0b03a643f204b8b4b4cefd44341f6f158c40cfb3b74ac │ │ └─57356 /pause │ ├─ce7b14c804158745bedfc31dfe42513d289d4d791f93fad846ee732d79d7f062 │ │ └─44288 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true │ ├─9c5fa70c3c2794e0937c785428ff3548556062efcc90bcabded88dce4ed25ed7 │ │ └─42599 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─e1ead3df21ffaf0ca1d1ccc5a0ebda5467939520aa909ca32108fc976b75b39b │ │ └─43100 /coredns -conf /etc/coredns/Corefile │ ├─5f7576fe3b7ed810102c4d436584a5e345cd311902dffb4f9c7e2cd5601ba950 │ │ └─50357 /pause │ ├─9091c3f5a50646ff914811e177c9d138829e91b82e386e884f67e7abd05f063d │ │ └─47077 /pause │ ├─7b4d71c5db7a1d9dcda90c5b6a3195b5a9fa9356a776dfa2ce63eb216dfccfb3 │ │ └─64544 /pause │ ├─d24da7cdbdf1e8add6112e8dc81456789cb85504186b718861b97305c34a465f │ │ └─39771 /pause │ ├─21e0bb350e05431a36afc5fb3b60f25a4149a953d3525367e61535f6e6af82df │ │ └─47166 /pause │ ├─24b5cd711b3b48949e5e296e8cd004308747d09d94d1cd0531e02f1c41982ff2 │ │ └─44379 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager │ ├─6991b91b7d99038afee040718cd784b7e1d3d219c8e3f7cb4b5fd557746c3b14 │ │ └─43070 /coredns -conf /etc/coredns/Corefile │ └─8313256af339bb37342e3ca3751e08cc24c48d7911a16acfe24ec8f65cbfac67 │ └─44332 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.10 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount