Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─616 /lib/systemd/systemd --user │ │ └─617 (sd-pam) │ ├─session-4.scope │ │ ├─1348 sshd: zuul [priv] │ │ └─1361 sshd: zuul@notty │ └─session-1.scope │ ├─ 613 sshd: zuul [priv] │ ├─ 630 sshd: zuul@notty │ ├─ 767 /usr/bin/python3 │ ├─ 1352 ssh: /home/zuul/.ansible/cp/199.204.45.250-22-zuul [mux] │ ├─35497 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-srmobwodqqlivxqdazjivibifoetbknx ; /usr/bin/python3'"'"' && sleep 0' │ ├─35498 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-srmobwodqqlivxqdazjivibifoetbknx ; /usr/bin/python3' && sleep 0 │ ├─35499 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-srmobwodqqlivxqdazjivibifoetbknx ; /usr/bin/python3 │ ├─35500 /bin/sh -c echo BECOME-SUCCESS-srmobwodqqlivxqdazjivibifoetbknx ; /usr/bin/python3 │ ├─35501 /usr/bin/python3 │ ├─35502 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─35504 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 2869 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─ 4422 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f92d9c21160541da2c8e86a6fde9d97ba434a51c9ad8e20316197b7cbe1638b2 -address /run/containerd/containerd.sock │ │ ├─ 4449 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7b3c809872f5d0e7bea5c9d9e335142523631a58c68b81a2d9b696920f106324 -address /run/containerd/containerd.sock │ │ ├─ 4462 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3a2c4918de0c8e5d77dc270943f2e6010ff114e928aceffaceb17f939136c39f -address /run/containerd/containerd.sock │ │ ├─ 4472 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c0e4e187c82f179361dbde8ea6dbc11bdec0d7f96727b1ab12bd7f1e02abb0c1 -address /run/containerd/containerd.sock │ │ ├─ 4490 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d87a29659dccb6d0fc71b7209ebdbc7aa91d16c39824ec1dba52dee68df79b9e -address /run/containerd/containerd.sock │ │ ├─ 4987 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a70c8c3d0e73baca977831a39d8e3144c18ab76ea790cf1d54402f72f5b949a7 -address /run/containerd/containerd.sock │ │ ├─ 5639 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 65502bcf068c11851b542f75503e59de87ba6bc5fe50c3dee2722129f0616e19 -address /run/containerd/containerd.sock │ │ ├─ 5660 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4ca912f8213441f60d2885cfdf53123986e743f3f66f236fd02cbe9056da9e56 -address /run/containerd/containerd.sock │ │ ├─ 6846 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 54c4b1c8fd7c93382448c3d0d46fff7c2ddf4facce974497dd699e63c17c135a -address /run/containerd/containerd.sock │ │ ├─ 7194 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 361efd0a3ce8e074262ac340a5f7a70b0e23bd5557f835a0982e3ebaf6fedc2f -address /run/containerd/containerd.sock │ │ ├─ 7249 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 127cfb6a4ed78294b2a76818001a10a140a0ce2575a5e83b4be0ccc318cba1a2 -address /run/containerd/containerd.sock │ │ ├─ 7591 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3cffff5e6f269060e4120e7b11dcca6efd986fdd68599e765825fb02789c46be -address /run/containerd/containerd.sock │ │ ├─ 8100 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 41033eeee467829df1ce2884fc5d236872381062269377efbe1725fd2df29bee -address /run/containerd/containerd.sock │ │ ├─ 8146 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c7c14e37d03a3245eda9f1cd8a01adb00fd2cd61a9318c0a63cb60ddeca42124 -address /run/containerd/containerd.sock │ │ ├─ 8184 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 4b34884b7e4c6de261b0d683f23303d0ae4d994a566d0e0d7c256f8fb496eaea -address /run/containerd/containerd.sock │ │ ├─10072 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e914e16ad83fc30478f89911e0cb2d238c51b455847dceb9cd91df9342dbe7de -address /run/containerd/containerd.sock │ │ ├─10130 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 71d13d78e87806e530c97fba032f6a72a192f3d0ad0ec1b6097bd6d5f8ccab21 -address /run/containerd/containerd.sock │ │ ├─10765 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1a50ac3b361d515ac2fc322d45ec9c4dcc395f5bacd62867f652d9ee4781042c -address /run/containerd/containerd.sock │ │ ├─10872 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 696e06ecc52f07d8539e170fa81c8b035f346ab92f4bc143ff99d39d51d37ee9 -address /run/containerd/containerd.sock │ │ ├─11345 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ecf0a13f2d5e033369eb07482a65ed629214b7ea0f52bc37ea09a08c02a4460e -address /run/containerd/containerd.sock │ │ ├─11651 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9054f0f3721fc98044af65137f51479cd39155e8961d29188195aacb13c56992 -address /run/containerd/containerd.sock │ │ ├─11971 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 61e66db6568e2f39527fa98e22c038511823efd151282626a5cbf269405e92b2 -address /run/containerd/containerd.sock │ │ ├─13455 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 37857acbe5728bab81fe0d018efa12b6fa9e18241955e767a058a754e3bf7f7f -address /run/containerd/containerd.sock │ │ ├─14053 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 891ca2bc3e2bb85c9ab19a8f3610c78742a6300ce3787d24e521d84d9b7e8781 -address /run/containerd/containerd.sock │ │ ├─16022 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 29c9c85b118d1caeeda94e648f40a231e231f63b0cb543fc311d524cabf3e718 -address /run/containerd/containerd.sock │ │ ├─18322 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d9baf1368bb809c366fae9dc0c3c7167b67ac9fe09c882a2242c096d0aac73ac -address /run/containerd/containerd.sock │ │ ├─18412 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 17fe67d62b698fa8097d4218fc7ea879e7ccba0d932aa9f2e6618ec86495ee65 -address /run/containerd/containerd.sock │ │ ├─18450 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f08311367a83ae9170fb25054125ec7f7eeb9d66ac50bdfa91d0e8279afada5c -address /run/containerd/containerd.sock │ │ ├─18496 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 35e0444a4b7045925a7a49b5c2717873cc37e7639b62b557c00a6758327cb543 -address /run/containerd/containerd.sock │ │ ├─19522 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e1ca18ac546e59cc85b033f30e21d016c484138440044caf0926c8821e5ec259 -address /run/containerd/containerd.sock │ │ ├─19644 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 58f4939ae34a5bea35eca1456f4386492aafe996f9fdd69c1dd8f4e02d1820e9 -address /run/containerd/containerd.sock │ │ ├─20412 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1fa59d7a9d35cdca105c9c93b54fdf27ebc7bb276a1ec6d80d63fd3b525f20c9 -address /run/containerd/containerd.sock │ │ ├─21685 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id afc0f8bdab1f5d4d90217a9284f2776d05913718fd37d3826b689c184a4a2707 -address /run/containerd/containerd.sock │ │ ├─26960 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id fa892dc637e3e861e6fe493e39aebd97fb85be7a5555cff79e6769aff6e36e41 -address /run/containerd/containerd.sock │ │ ├─27055 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 62a02c8e6cea1f0093052e468f4a0f7e5ef72350c43ef9fb91850ccb384fe8a0 -address /run/containerd/containerd.sock │ │ └─27093 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6ee2db379cad4e8c5d8f3d95f0ecfe44c94d3015e3e80e9c08a139bd4ddf8cf9 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1094 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─437 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─328 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─544 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─557 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─531 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─4856 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.250 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─cloud-init-local.service │ ├─systemd-journald.service │ │ └─295 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─559 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1283 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─332 /lib/systemd/systemd-resolved │ ├─dbus.service │ │ └─528 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─333 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─543 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─534 /lib/systemd/systemd-logind ├─k8s.io │ ├─41033eeee467829df1ce2884fc5d236872381062269377efbe1725fd2df29bee │ │ └─8123 /pause │ ├─35e0444a4b7045925a7a49b5c2717873cc37e7639b62b557c00a6758327cb543 │ │ └─18521 /pause │ ├─f92d9c21160541da2c8e86a6fde9d97ba434a51c9ad8e20316197b7cbe1638b2 │ │ └─4535 /pause │ ├─6242d8fa8589abbfde304cfd71f6d2f1143677ac7217907632a799c2be6ecf13 │ │ └─19850 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning │ ├─6ee2db379cad4e8c5d8f3d95f0ecfe44c94d3015e3e80e9c08a139bd4ddf8cf9 │ │ └─27120 /pause │ ├─73b72b56c63e5f8bdaa1411a570743fa9d437899f69efed274a5590ead56fff5 │ │ └─7253 /coredns -conf /etc/coredns/Corefile │ ├─c72d6729e1122b170692c9d2d67efa1fc313b8fe07f8d780b4a8914a2f463d15 │ │ └─8388 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager │ ├─8d3fc2b89b0fdddca2cdcd1c5c152ecd6b7b81ebe9903ab0c4b4981aac10a691 │ │ └─8313 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.10 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 │ ├─a435213e852fc8d3334fd3cc92197348d3d2a7ad8a283c55eb323105b483d663 │ │ └─4740 etcd --advertise-client-urls=https://199.204.45.250:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.250:2380 --initial-cluster=instance=https://199.204.45.250:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.250:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.250:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─3d071dba0a5802d40039eace212e62bab396769a4e126037cf49b56c4c0fa059 │ │ └─20282 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 │ ├─4c90fc508bc0ca02b214bae20e06da51567f7d1cdd8a8d26351a147d9cb93b4e │ │ └─10403 /server │ ├─18c29bd25870a04307f044ff2bb1b2cc03937bb73aadf701e36b5f250e0adedd │ │ └─12084 mysqld --wsrep_start_position=2d051675-3463-11f1-b8a6-a7a36830bf92:22 │ ├─106bb5144d29c102eb9d4b481db65f91494d6a455f219cc0a0b3375b04630462 │ │ └─13819 memcached -v -p 11211 -U 0 -c 8192 -m 1024 │ ├─29c9c85b118d1caeeda94e648f40a231e231f63b0cb543fc311d524cabf3e718 │ │ └─16047 /pause │ ├─60bbb77ac201e72de2f9a7f364d9940a521926e92d90deab08f5c38d9e256c1c │ │ └─6900 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─3d9538b72adcae544757cd70c2a08a82f3b84fb607a055015cc803ad231cb3d4 │ │ ├─20801 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─20819 /usr/local/lib/erlang/erts-12.3.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /usr/local/lib/erlang -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false │ │ ├─20826 erl_child_setup 1048576 │ │ ├─20869 /usr/local/lib/erlang/erts-12.3.2/bin/epmd -daemon │ │ ├─20912 inet_gethost 4 │ │ └─20913 inet_gethost 4 │ ├─f971b565881dc64ab704068e6839574d5e94f241e798f5056288624eed6afca1 │ │ └─13866 /bin/memcached_exporter │ ├─c0e4e187c82f179361dbde8ea6dbc11bdec0d7f96727b1ab12bd7f1e02abb0c1 │ │ └─4595 /pause │ ├─bceab781efa5dfe0726a6426157de72e1384a7cc0e4d2f5c8f7c3eff8dd05c3a │ │ └─7331 /coredns -conf /etc/coredns/Corefile │ ├─4ca912f8213441f60d2885cfdf53123986e743f3f66f236fd02cbe9056da9e56 │ │ └─5702 /pause │ ├─3cffff5e6f269060e4120e7b11dcca6efd986fdd68599e765825fb02789c46be │ │ └─7617 /pause │ ├─65502bcf068c11851b542f75503e59de87ba6bc5fe50c3dee2722129f0616e19 │ │ └─5684 /pause │ ├─8840fea6758145fcce1ffc74845b0bebb48f51d946dc1adb11e65868bd63f6b2 │ │ └─20506 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config │ ├─d87a29659dccb6d0fc71b7209ebdbc7aa91d16c39824ec1dba52dee68df79b9e │ │ └─4567 /pause │ ├─ef141f94f1eed9d317ad6d442880a179599d4200f45f23851619b956ef4cc575 │ │ ├─23783 apache2 -DFOREGROUND │ │ ├─23819 (wsgi:k -DFOREGROUND │ │ ├─23820 (wsgi:k -DFOREGROUND │ │ ├─23821 (wsgi:k -DFOREGROUND │ │ ├─23822 (wsgi:k -DFOREGROUND │ │ ├─23826 apache2 -DFOREGROUND │ │ ├─23827 apache2 -DFOREGROUND │ │ ├─23829 apache2 -DFOREGROUND │ │ ├─23830 apache2 -DFOREGROUND │ │ ├─23832 apache2 -DFOREGROUND │ │ ├─23833 apache2 -DFOREGROUND │ │ ├─23834 apache2 -DFOREGROUND │ │ ├─23836 apache2 -DFOREGROUND │ │ ├─23837 apache2 -DFOREGROUND │ │ └─23838 apache2 -DFOREGROUND │ ├─d9baf1368bb809c366fae9dc0c3c7167b67ac9fe09c882a2242c096d0aac73ac │ │ └─18363 /pause │ ├─58f4939ae34a5bea35eca1456f4386492aafe996f9fdd69c1dd8f4e02d1820e9 │ │ └─19667 /pause │ ├─696e06ecc52f07d8539e170fa81c8b035f346ab92f4bc143ff99d39d51d37ee9 │ │ └─10896 /pause │ ├─a5a4ad3e7fbb41eaa5f9a7786989aaa395857b15a161a7c7e35db474734e8830 │ │ └─7699 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─b615b6841bf680a65b19688fa42684ad308be90226ca845c54d289d416fa0a9d │ │ └─4664 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─62a02c8e6cea1f0093052e468f4a0f7e5ef72350c43ef9fb91850ccb384fe8a0 │ │ └─27078 /pause │ ├─453e6fd24c47fe94489584405bea9cc78d301da4b78e6393ac6facfa449682d0 │ │ └─4735 kube-apiserver --advertise-address=199.204.45.250 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─e914e16ad83fc30478f89911e0cb2d238c51b455847dceb9cd91df9342dbe7de │ │ └─10099 /pause │ ├─1fa59d7a9d35cdca105c9c93b54fdf27ebc7bb276a1ec6d80d63fd3b525f20c9 │ │ └─20433 /pause │ ├─afc0f8bdab1f5d4d90217a9284f2776d05913718fd37d3826b689c184a4a2707 │ │ └─21711 /pause │ ├─d1ceae771d788750f0933312ae84fe66b8f4e699812cc1bbacf84894d0f125a4 │ │ ├─10359 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─10372 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─10498 nginx: master process /usr/bin/nginx -c /etc/nginx/nginx.conf │ │ ├─28098 nginx: worker process │ │ ├─28099 nginx: worker process │ │ └─28100 nginx: cache manager process │ ├─c3396f336d58f11d39169326dda64cc00ad37f24860815ff2812c0fa57e788b1 │ │ └─20325 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─a2b47b52107a7bca1dbec9db36c79a2e88fc13ec7a5e225560a02946fe8b7018 │ │ ├─28463 apache2 -DFOREGROUND │ │ ├─31413 (wsgi:h -DFOREGROUND │ │ ├─31414 (wsgi:h -DFOREGROUND │ │ ├─31415 (wsgi:h -DFOREGROUND │ │ ├─31416 (wsgi:h -DFOREGROUND │ │ ├─31417 (wsgi:h -DFOREGROUND │ │ ├─31418 apache2 -DFOREGROUND │ │ └─31419 apache2 -DFOREGROUND │ ├─4b34884b7e4c6de261b0d683f23303d0ae4d994a566d0e0d7c256f8fb496eaea │ │ └─8207 /pause │ ├─e5e21cc9f96df37627b4ef3be870115b8983ca8335cfebdc24d2f815e6728564 │ │ ├─19179 python -u /app/sidecar.py │ │ ├─19386 python -u /app/sidecar.py │ │ └─19387 python -u /app/sidecar.py │ ├─891ca2bc3e2bb85c9ab19a8f3610c78742a6300ce3787d24e521d84d9b7e8781 │ │ └─14077 /pause │ ├─4b49cbc7b91e552c0714f9c9e48a31bec7ac2376f5707af16de1a9bc666fd82e │ │ └─19467 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─f08311367a83ae9170fb25054125ec7f7eeb9d66ac50bdfa91d0e8279afada5c │ │ └─18476 /pause │ ├─c988a31e283382f4e88ea364ec87e62ae7e976b68e76024cc53b62679d2c2b4b │ │ └─5038 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─d919b1519fb825be591171c11a3b276df497ab68c3d459715dea24961c6c103d │ │ └─16159 /secretgen-controller │ ├─12f6ada71abd097ee586a7fefe5a10336e93555d9990b11ce453f3d2d6616558 │ │ └─20759 /bin/oauth2-proxy │ ├─17fe67d62b698fa8097d4218fc7ea879e7ccba0d932aa9f2e6618ec86495ee65 │ │ └─18436 /pause │ ├─ecf0a13f2d5e033369eb07482a65ed629214b7ea0f52bc37ea09a08c02a4460e │ │ └─11369 /pause │ ├─d56ba3334b41c7f6c376db29fba71b6b8f19e05b35df5db9a256d64c3e916fdf │ │ ├─19203 python -u /app/sidecar.py │ │ ├─19388 python -u /app/sidecar.py │ │ └─19389 python -u /app/sidecar.py │ ├─127cfb6a4ed78294b2a76818001a10a140a0ce2575a5e83b4be0ccc318cba1a2 │ │ └─7282 /pause │ ├─e104eaf2c490a792a690b3538a3cd4b4603ed6d8864e40a9509135c68aba32c6 │ │ └─12181 /bin/mysqld_exporter --collect.info_schema.processlist │ ├─f58a77428ef097d5d122bb9f88f3077dbd3971fd06f39230b61f1f22e530d473 │ │ ├─6107 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─6445 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─54c4b1c8fd7c93382448c3d0d46fff7c2ddf4facce974497dd699e63c17c135a │ │ └─6869 /pause │ ├─7b3c809872f5d0e7bea5c9d9e335142523631a58c68b81a2d9b696920f106324 │ │ └─4592 /pause │ ├─5ecfd44d6e7e9132c81ff31e946c7f95f9a73a3595d42c6f5fc65224b14c0ead │ │ └─4808 /kube-vip manager │ ├─fa892dc637e3e861e6fe493e39aebd97fb85be7a5555cff79e6769aff6e36e41 │ │ └─26988 /pause │ ├─37857acbe5728bab81fe0d018efa12b6fa9e18241955e767a058a754e3bf7f7f │ │ └─13480 /pause │ ├─c343af74a9d4ea7eab963852eb08b35ebd41554d3c001366ae117df18533a1b4 │ │ └─18730 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.34.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 │ ├─9054f0f3721fc98044af65137f51479cd39155e8961d29188195aacb13c56992 │ │ └─11675 /pause │ ├─a0405dc03f904bcb52f17632d9100902fdac11558b7d3e72b8c3e9628ed226cf │ │ └─14146 java -Dkc.config.built=true -Xms64m -Xmx512m -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --auto-build --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true │ ├─6ad63f7865bf875c0c32fcf73bc7aa6689c5b4e2cefc68e64aaba427b733fb79 │ │ └─11898 /usr/bin/peer-list -on-change=/usr/bin/add_pxc_nodes.sh -service=percona-xtradb-pxc │ ├─5470eaa21feda60e47b061d5e2f4652e6540b24324441b1f5af91337dc5cd1bd │ │ └─19357 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments │ ├─bcf1e8d7b2e9394740c150030f1fd985a94f3ebc7d4448f51809d8d093834cfe │ │ └─20543 /bin/oauth2-proxy │ ├─d9548c0d85eb6f61f21300f434556a77ca834ae34c3c3cffda6c86f6fdadae90 │ │ └─11110 /manager --metrics-bind-address=:8080 │ ├─29e8379341dec99d2c419da7e59d74a046ef359b65975494d1c9775a395ef7b7 │ │ └─20473 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.199-204-45-250.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml │ ├─ccdcdbfca2893b03af4a48a6d903b988f086041f851006cdcc4e72dc4d4e7547 │ │ ├─11858 haproxy -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ │ └─13033 haproxy -sf 11 -x sockpair@3 -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ ├─b50783d832d963bc39c052c7c5c285e128e5f85b16a2d6928c5fcf0e373defa7 │ │ └─10972 /manager --metrics-bind-address=:9782 │ ├─8a7ef27b5b6b1685d50b16c688fdc67e1f842345a0c8b103623108c6e444f90a │ │ ├─28431 apache2 -DFOREGROUND │ │ ├─31222 (wsgi:h -DFOREGROUND │ │ ├─31223 (wsgi:h -DFOREGROUND │ │ ├─31224 (wsgi:h -DFOREGROUND │ │ ├─31225 (wsgi:h -DFOREGROUND │ │ ├─31226 (wsgi:h -DFOREGROUND │ │ ├─31227 apache2 -DFOREGROUND │ │ └─31228 apache2 -DFOREGROUND │ ├─7b5fe420e55e1d43d3298a57eb424416067d9f286ea58722b425fda17457b0df │ │ └─20248 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.199-204-45-250.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml │ ├─1a50ac3b361d515ac2fc322d45ec9c4dcc395f5bacd62867f652d9ee4781042c │ │ └─10789 /pause │ ├─cb500ae47278e8e91e4f7905db7d94d3194010708b988267bbeafd869087cebe │ │ └─11491 percona-xtradb-cluster-operator │ ├─df239e2d0d239e7550158467057796e46d8547524d7c13e6d38d3e29fb29eb78 │ │ └─20616 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml │ ├─a921e20693d6b4db262240276351343176f83795a8c5870109e89ce0447fdbe3 │ │ ├─28451 apache2 -DFOREGROUND │ │ ├─31233 (wsgi:h -DFOREGROUND │ │ ├─31234 (wsgi:h -DFOREGROUND │ │ ├─31235 (wsgi:h -DFOREGROUND │ │ ├─31236 (wsgi:h -DFOREGROUND │ │ ├─31237 (wsgi:h -DFOREGROUND │ │ ├─31238 apache2 -DFOREGROUND │ │ └─31239 apache2 -DFOREGROUND │ ├─e1ca18ac546e59cc85b033f30e21d016c484138440044caf0926c8821e5ec259 │ │ └─19547 /pause │ ├─361efd0a3ce8e074262ac340a5f7a70b0e23bd5557f835a0982e3ebaf6fedc2f │ │ └─7217 /pause │ ├─3a2c4918de0c8e5d77dc270943f2e6010ff114e928aceffaceb17f939136c39f │ │ └─4575 /pause │ ├─5343dac73a7d00a00afc41a604b9bd1720e62249ded0449e58787eab31df7513 │ │ └─8352 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true │ ├─c7c14e37d03a3245eda9f1cd8a01adb00fd2cd61a9318c0a63cb60ddeca42124 │ │ └─8170 /pause │ ├─a70c8c3d0e73baca977831a39d8e3144c18ab76ea790cf1d54402f72f5b949a7 │ │ └─5011 /pause │ ├─71d13d78e87806e530c97fba032f6a72a192f3d0ad0ec1b6097bd6d5f8ccab21 │ │ └─10154 /pause │ ├─61e66db6568e2f39527fa98e22c038511823efd151282626a5cbf269405e92b2 │ │ └─11995 /pause │ ├─c368b8a6f140c7aa252a9b13d2712640f223a645c30beb4b1adace02db4bb81d │ │ └─4684 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ └─851b7a30e6ae12fe123efb8efac6058d9273e5a3f810728774d5b33ac8b7e620 │ └─5905 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount