Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─616 /lib/systemd/systemd --user │ │ └─617 (sd-pam) │ ├─session-4.scope │ │ ├─1356 sshd: zuul [priv] │ │ └─1369 sshd: zuul@notty │ └─session-1.scope │ ├─ 613 sshd: zuul [priv] │ ├─ 630 sshd: zuul@notty │ ├─ 767 /usr/bin/python3 │ ├─ 1360 ssh: /home/zuul/.ansible/cp/162.253.55.204-22-zuul [mux] │ ├─41818 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-kzjlbdjlgcawzblfsxovywwfequzxsig ; /usr/bin/python3'"'"' && sleep 0' │ ├─41819 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-kzjlbdjlgcawzblfsxovywwfequzxsig ; /usr/bin/python3' && sleep 0 │ ├─41820 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-kzjlbdjlgcawzblfsxovywwfequzxsig ; /usr/bin/python3 │ ├─41821 /bin/sh -c echo BECOME-SUCCESS-kzjlbdjlgcawzblfsxovywwfequzxsig ; /usr/bin/python3 │ ├─41822 /usr/bin/python3 │ ├─41823 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─41825 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 2879 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─ 4439 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 55b8f2b60eb42afd2ade1c3a00d0b9268ce1c0851d27f9002657908635729205 -address /run/containerd/containerd.sock │ │ ├─ 4481 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 707dc700defa6e942c1d5d207260b0d413ac5f4e4254f209cf48cbf92ae7f6e5 -address /run/containerd/containerd.sock │ │ ├─ 4490 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b9c2df726cb59823c03a22f1e9b058d6c26451d7908ac89805d38fe6e0bcf342 -address /run/containerd/containerd.sock │ │ ├─ 4497 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id de148e354d306db6de9e55a51c174566c8e3b9751b48a2ef9d512c0acda31fb9 -address /run/containerd/containerd.sock │ │ ├─ 4498 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 76b997980e185b9e0119e00f76df2cbf6e9b3228eda1cae34d2263820d23f28b -address /run/containerd/containerd.sock │ │ ├─ 5026 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e8b8eee185cea3bc6c85f9c1dd2adfcf8ba7eaa750b7f20754ee7d536c30b996 -address /run/containerd/containerd.sock │ │ ├─ 5673 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7b8753ea48e8300516f934e46fa7630431161e4bd154d9f9e128c6c7f09d4a4a -address /run/containerd/containerd.sock │ │ ├─ 5706 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 35d174648d629d227fd0796560273df9bc0e182bdf0955aeccac6323d2bd59d9 -address /run/containerd/containerd.sock │ │ ├─ 6955 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a8f5facd77dcc217be23b0c7c1c201b6d4c37c8134f89b8aa78dc0d1a047d16a -address /run/containerd/containerd.sock │ │ ├─ 7303 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0eb4eba9f7ea47f74af9cd2501a428743682e59cc4fc9fc722e0d06e6c0e8c1b -address /run/containerd/containerd.sock │ │ ├─ 7341 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5b8045b88846415b40ff6512cf8b20b62bf005ba61372622183e45417134f916 -address /run/containerd/containerd.sock │ │ ├─ 7691 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5fc02800060274b4fe13fa57b7abf57f0458c8e806cfd395e1e4eb84e5aa8a90 -address /run/containerd/containerd.sock │ │ ├─ 8160 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3e90766aa314315d22c1f5abadb5520e91b21bab45c5cc06275d3950fd039388 -address /run/containerd/containerd.sock │ │ ├─ 8201 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 28650ea3fa81ea122e165277ce0aa1dd53d869927969616589f770dfbcb0cece -address /run/containerd/containerd.sock │ │ ├─ 8222 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e1896c26bf82cfcf381da05897be3eb3c941be3214ecb56dca4822ba3381b51f -address /run/containerd/containerd.sock │ │ ├─10416 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d51edf5b6e565972c7b23c03358567389f19cc1ab53225196db76e78a5ed7d00 -address /run/containerd/containerd.sock │ │ ├─10505 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a9a8ad8ae026f8a29f2db31f11073d9f9d1170cd00b8fb5af7c4b1f8d4d6d27c -address /run/containerd/containerd.sock │ │ ├─11116 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ea1082c43220fc3074423b620aafe696999ca2c67b431d92bc1b8d317d578288 -address /run/containerd/containerd.sock │ │ ├─11209 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 24dd8f95c2280d79be65975441e299045fc156c66a88860d03f8d4b3a3ee1de6 -address /run/containerd/containerd.sock │ │ ├─11478 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 09e1b1b8b16907af2e67559becd6e5e32957f091b5ff770f3805435287899c82 -address /run/containerd/containerd.sock │ │ ├─11949 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7af3f82ba6e6963f8ae8c00ec65f66585d34f0c2e8421329a465ea76320dec13 -address /run/containerd/containerd.sock │ │ ├─12323 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 386efef1e08aaa3a7268eb7831b3ba7a4b3a3c2f377533596edae79e43f13856 -address /run/containerd/containerd.sock │ │ ├─13990 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7f42e2691e436319611af6081df84a282f2cc9b5ff0021b081eaeedbde8592e4 -address /run/containerd/containerd.sock │ │ ├─14656 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 635438801fb6a6a24162ff05b0332fc32c0447ebb75c7f095491143d6ed9579d -address /run/containerd/containerd.sock │ │ ├─17959 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1dd8e8b09acdbf42d26aeefe5e7e01aa02a7f3adbe5a5989a39dd048fda0f9b7 -address /run/containerd/containerd.sock │ │ ├─20594 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id dddd195166465be44f764d00e4aeb3c90d138640bfd2de71ddcc41d69d4a0190 -address /run/containerd/containerd.sock │ │ ├─20651 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 098b7e9366765708a09e85685c43ae6313ef60231bf945342af1d3088e937a51 -address /run/containerd/containerd.sock │ │ ├─20681 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 429d9ad12c54a018649325cf1c26d3fca0278251da214e904d5f5c796da3f180 -address /run/containerd/containerd.sock │ │ ├─20697 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 00c51079d491cb585144eb3e0cd90f97a6cfc6f30a234310b8777a08e6fab76a -address /run/containerd/containerd.sock │ │ ├─21773 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 463de3d6a2f71e15aaade2e49413e230868c3998096a4eba42817b72046a1832 -address /run/containerd/containerd.sock │ │ ├─21858 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ff6c6e335b5f134a8fbfd91731cd38479a3caccd5a060708f94c52ef7aab1a6e -address /run/containerd/containerd.sock │ │ ├─22737 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 99ad99b8b6e5c9ba7544f5ff24ae8f161673d8f4b0ba9c0e86ee53da773529d4 -address /run/containerd/containerd.sock │ │ ├─24403 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0781853ac2b9dc49218b83fa7e611e84c8e814d693dcabda307f570ffa631046 -address /run/containerd/containerd.sock │ │ ├─29909 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b5446b133443f0b3e4c18a8773c6dd52894c0f87a2c18eec975d213795225869 -address /run/containerd/containerd.sock │ │ ├─29946 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 065119532a7b0036ae145736209a9309ad0c26debbce01d76ac3162cd8075e33 -address /run/containerd/containerd.sock │ │ └─30058 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 299b41a8b9ba001ef8beffe4d65fd62964888c42449c5a97d00d16304bbc7d71 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1091 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─437 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─326 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─542 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─558 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─532 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─4875 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=162.253.55.204 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─cloud-init-local.service │ ├─systemd-journald.service │ │ └─294 /lib/systemd/systemd-journald │ ├─ssh.service │ │ └─563 sshd: /usr/sbin/sshd -D [listener] 0 of 10-100 startups │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1283 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─332 /lib/systemd/systemd-resolved │ ├─dbus.service │ │ └─529 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─333 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─538 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─535 /lib/systemd/systemd-logind ├─k8s.io │ ├─99ad99b8b6e5c9ba7544f5ff24ae8f161673d8f4b0ba9c0e86ee53da773529d4 │ │ └─22760 /pause │ ├─b9c2df726cb59823c03a22f1e9b058d6c26451d7908ac89805d38fe6e0bcf342 │ │ └─4579 /pause │ ├─6c7ab8cb1ae7acf90ac550adf8ee7f4c4140d47dd045b79d5d0a7b1163bdd033 │ │ └─4741 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─76b997980e185b9e0119e00f76df2cbf6e9b3228eda1cae34d2263820d23f28b │ │ └─4608 /pause │ ├─0af8948fc681e1375876bfab0b81f03ecb05a3e661413cde5702b6f2881b04ab │ │ └─10842 /server │ ├─e1ba945b407815355997103f5b5c59a21b731c1a61f21bb946a9387b1e6a429b │ │ └─12447 /opt/percona/peer-list -on-change=/opt/percona/haproxy_add_pxc_nodes.sh -service=percona-xtradb-pxc -protocol=$(PEER_LIST_SRV_PROTOCOL) │ ├─1f0efd8b95805e77666d5326779b4d6919267f25a8efb7422d420462480b15e7 │ │ └─11824 percona-xtradb-cluster-operator │ ├─a8f5facd77dcc217be23b0c7c1c201b6d4c37c8134f89b8aa78dc0d1a047d16a │ │ └─6978 /pause │ ├─77811c7fdb28898ddf1c98c75901003f51153546c63448a99da776e2ef6e97cc │ │ ├─10761 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─10773 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─10872 nginx: master process /usr/bin/nginx -c /etc/nginx/nginx.conf │ │ ├─31081 nginx: worker process │ │ ├─31082 nginx: worker process │ │ └─31083 nginx: cache manager process │ ├─6d87c80ceae425d26c6658ce215ebd19b247c491a10bec18abde8974863342ad │ │ └─7427 /coredns -conf /etc/coredns/Corefile │ ├─463de3d6a2f71e15aaade2e49413e230868c3998096a4eba42817b72046a1832 │ │ └─21797 /pause │ ├─72a82a97b9a482225c77a3c9ce068f4869bf287107ce83e57fc1f9e65c0bf5ae │ │ └─5077 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─c33f63eda54a4fd29e33c6476bfdd02d452bc50cbbf7acbd8b8e6820d008550a │ │ ├─6232 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─6601 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─d98bf8ed5f0794ded1ba80030a6f3562fc822977323605400a8b01e07c4739ef │ │ └─7395 /coredns -conf /etc/coredns/Corefile │ ├─de148e354d306db6de9e55a51c174566c8e3b9751b48a2ef9d512c0acda31fb9 │ │ └─4593 /pause │ ├─fe22718055d798e88d6fb933c77702cfbc22b776df5357f3ee2047789d1d302c │ │ └─12641 /bin/mysqld_exporter --mysqld.username=monitor --collect.info_schema.processlist │ ├─d51edf5b6e565972c7b23c03358567389f19cc1ab53225196db76e78a5ed7d00 │ │ └─10444 /pause │ ├─941adfe6d0e5eeac1ee13b969171a2a07ee075e6317a8c93877c8b6e71ed161b │ │ └─11341 /manager --metrics-bind-address=:9782 │ ├─098b7e9366765708a09e85685c43ae6313ef60231bf945342af1d3088e937a51 │ │ └─20707 /pause │ ├─ea1082c43220fc3074423b620aafe696999ca2c67b431d92bc1b8d317d578288 │ │ └─11138 /pause │ ├─ea5bb0d49f98a6540fa047d2257dedea96af3d7cea9030a12c6a20bc934137d6 │ │ └─14452 /bin/memcached_exporter │ ├─28650ea3fa81ea122e165277ce0aa1dd53d869927969616589f770dfbcb0cece │ │ └─8246 /pause │ ├─887ddbb1e2144b74c76ea258f7bcad4b0872fc1a3bbe2556b6bcd7ad9544fa51 │ │ └─18100 /secretgen-controller │ ├─7e1412d61826b4fb5feb3604dab2dee458c5516ef0a7069abb3291da5bfe58ec │ │ └─8377 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true │ ├─1dd8e8b09acdbf42d26aeefe5e7e01aa02a7f3adbe5a5989a39dd048fda0f9b7 │ │ └─17983 /pause │ ├─3e90766aa314315d22c1f5abadb5520e91b21bab45c5cc06275d3950fd039388 │ │ └─8184 /pause │ ├─c90d91d5c59027d81445b20b3608c8104d7283c614643a21b4e3001ef62010d6 │ │ └─21912 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─2af69ab14c3c50bc8b38f7fcd8aef6813432948a73392fb35c54925d839b2815 │ │ └─4719 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─55b8f2b60eb42afd2ade1c3a00d0b9268ce1c0851d27f9002657908635729205 │ │ └─4525 /pause │ ├─065119532a7b0036ae145736209a9309ad0c26debbce01d76ac3162cd8075e33 │ │ └─29971 /pause │ ├─dcee822b2721e48a6e5ea3e4c296e5720bc15722675bf35c530327dd87a058fa │ │ └─22840 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 │ ├─e1896c26bf82cfcf381da05897be3eb3c941be3214ecb56dca4822ba3381b51f │ │ └─8263 /pause │ ├─bae02e9fc208cc50779f1fca2a97107fd8a58de15fb2039ff04df68b01f44aea │ │ └─4683 kube-apiserver --advertise-address=162.253.55.204 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─5b8045b88846415b40ff6512cf8b20b62bf005ba61372622183e45417134f916 │ │ └─7366 /pause │ ├─b5446b133443f0b3e4c18a8773c6dd52894c0f87a2c18eec975d213795225869 │ │ └─29933 /pause │ ├─39099284a4eaac9baaaa1fc82316766dc5103db968961d071713efce490179d3 │ │ └─11588 /manager --metrics-bind-address=:8080 │ ├─b32534b6235b30d4b0777398cd58d2274760cc542f8cbb6101fe51874004f613 │ │ ├─21485 python -u /app/sidecar.py │ │ ├─21624 python -u /app/sidecar.py │ │ └─21625 python -u /app/sidecar.py │ ├─ff6c6e335b5f134a8fbfd91731cd38479a3caccd5a060708f94c52ef7aab1a6e │ │ └─21882 /pause │ ├─0781853ac2b9dc49218b83fa7e611e84c8e814d693dcabda307f570ffa631046 │ │ └─24427 /pause │ ├─340805b4655073551240123d7750b55d38e6082dadd81a0787e728279a2fa64b │ │ └─14760 java -Dkc.config.built=true -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 -XX:FlightRecorderOptions=stackdepth=512 -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:MaxRAMPercentage=70 -XX:MinRAMPercentage=70 -XX:InitialRAMPercentage=50 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dpicocli.disable.closures=true -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true │ ├─a9a8ad8ae026f8a29f2db31f11073d9f9d1170cd00b8fb5af7c4b1f8d4d6d27c │ │ └─10528 /pause │ ├─e8b8eee185cea3bc6c85f9c1dd2adfcf8ba7eaa750b7f20754ee7d536c30b996 │ │ └─5049 /pause │ ├─5bec3b04ae1bdbd83b174f69f87f62119fec1a57b300b7102e5020ea257270d2 │ │ └─23250 /bin/oauth2-proxy │ ├─a1a7439296804904cba7fb9b1cdc2088802ac9278f3fec755ac91f26314c12c5 │ │ └─22237 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning │ ├─707dc700defa6e942c1d5d207260b0d413ac5f4e4254f209cf48cbf92ae7f6e5 │ │ └─4607 /pause │ ├─09e1b1b8b16907af2e67559becd6e5e32957f091b5ff770f3805435287899c82 │ │ └─11501 /pause │ ├─429d9ad12c54a018649325cf1c26d3fca0278251da214e904d5f5c796da3f180 │ │ └─20747 /pause │ ├─b1a41f26cf8a62b9852c5e679091e0a8a70d7e1cf2ab4af1878f00fda8509966 │ │ ├─21408 python -u /app/sidecar.py │ │ ├─21622 python -u /app/sidecar.py │ │ └─21623 python -u /app/sidecar.py │ ├─3e3a48e01c1f7ec390208835d78d65c56f2d8992ea3500288df52142c66df89f │ │ ├─12371 haproxy -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ │ └─13571 haproxy -sf 15 -x sockpair@4 -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ ├─acc810a1b1c69431f1e3f1d172fe05da56db396b05bf672e195a3dee35ad64a4 │ │ ├─12542 mysqld --wsrep_start_position=acc97275-33b7-11f1-ac67-c2ece19bf7b1:20 │ │ └─12666 /var/lib/mysql/mysql-state-monitor │ ├─7af3f82ba6e6963f8ae8c00ec65f66585d34f0c2e8421329a465ea76320dec13 │ │ └─11973 /pause │ ├─635438801fb6a6a24162ff05b0332fc32c0447ebb75c7f095491143d6ed9579d │ │ └─14679 /pause │ ├─23c54982a8b59644fb1a666642db00bda395ef1a0ea93f9623ece98a8e736e70 │ │ └─7012 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─9d82d707ec038e2970f629803e9363033f76a9ae839335dce5f2716753c91fc9 │ │ └─22803 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.162-253-55-204.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml │ ├─00c51079d491cb585144eb3e0cd90f97a6cfc6f30a234310b8777a08e6fab76a │ │ └─20746 /pause │ ├─ab063d1b46c1bc6b87a408a8f00c7dbbd71427c659f378c51d6192f60d2b1188 │ │ └─21659 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments │ ├─7b7dd6a887d98e9513963f709d225c129840c609ea3378a976dd89b738c4721d │ │ ├─23298 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─23315 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─23322 erl_child_setup 1048576 │ │ ├─23397 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─23398 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─23408 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon │ │ └─23450 /bin/sh -s rabbit_disk_monitor │ ├─7f42e2691e436319611af6081df84a282f2cc9b5ff0021b081eaeedbde8592e4 │ │ └─14015 /pause │ ├─06bfe7c38ea621c868bfed02ce6ba6d82c0c318a0a5ace99484adac9f22dbc55 │ │ └─8454 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager │ ├─22d2da04ab6d4407178fe0be6e632d11121466a5c45fdc8a78ccea1708af8611 │ │ └─7857 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─5fc02800060274b4fe13fa57b7abf57f0458c8e806cfd395e1e4eb84e5aa8a90 │ │ └─7718 /pause │ ├─24dd8f95c2280d79be65975441e299045fc156c66a88860d03f8d4b3a3ee1de6 │ │ └─11232 /pause │ ├─37ad7dd1875a0814cb9f27b8e96d4db34b67ef929422825702fcd1d43b21fc1c │ │ └─22972 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config │ ├─0eb4eba9f7ea47f74af9cd2501a428743682e59cc4fc9fc722e0d06e6c0e8c1b │ │ └─7326 /pause │ ├─ff7b4d9224d38b15443a1519a2222930d6dedc9e10437230228e6123fb2c6476 │ │ └─4827 /kube-vip manager │ ├─76a4f48bdcc338331dc075ab6b359affd74c4e484b099bec35949f0b5e8da86b │ │ └─23232 /bin/oauth2-proxy │ ├─bb8f5906a7c67c881da07a6c4118818952e419f7f831a50b35a6605206657242 │ │ └─20991 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.35.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 │ ├─f493ee5736e1f3054c67f9a0e7123373623c03a3aafc51faa6a23680b0d3820a │ │ └─22931 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.162-253-55-204.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml │ ├─386efef1e08aaa3a7268eb7831b3ba7a4b3a3c2f377533596edae79e43f13856 │ │ └─12346 /pause │ ├─cd8a68e1c69cb319e5cb6995fbf6207862f785ea1dc70af8433b14eae547f8bc │ │ └─6033 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─d651ba979ad1875f6983198ec64f83c67e87d32af874c81e7842cb7157e5765f │ │ └─22875 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─299b41a8b9ba001ef8beffe4d65fd62964888c42449c5a97d00d16304bbc7d71 │ │ └─30104 /pause │ ├─7b8753ea48e8300516f934e46fa7630431161e4bd154d9f9e128c6c7f09d4a4a │ │ └─5711 /pause │ ├─f67a7f891081912fead550f5e897eb2473fad7b0d9926ae99713418120fbf2eb │ │ └─4714 etcd --advertise-client-urls=https://162.253.55.204:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://162.253.55.204:2380 --initial-cluster=instance=https://162.253.55.204:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://162.253.55.204:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://162.253.55.204:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─dddd195166465be44f764d00e4aeb3c90d138640bfd2de71ddcc41d69d4a0190 │ │ └─20621 /pause │ ├─5a87a8eb799813e5efe29130e58af0e60b675d7c3210e7857b7167997ce7f8af │ │ └─14363 memcached -v -p 11211 -U 0 -c 8192 -m 1024 │ ├─9e1d78da7ae46d169efc2fb0e186de8188a82a93219012d437bfe4f8e20a85e8 │ │ ├─26670 apache2 -DFOREGROUND │ │ ├─26685 (wsgi:k -DFOREGROUND │ │ ├─26686 (wsgi:k -DFOREGROUND │ │ ├─26687 (wsgi:k -DFOREGROUND │ │ ├─26688 (wsgi:k -DFOREGROUND │ │ ├─26689 apache2 -DFOREGROUND │ │ ├─26692 apache2 -DFOREGROUND │ │ ├─26693 apache2 -DFOREGROUND │ │ ├─26695 apache2 -DFOREGROUND │ │ ├─26698 apache2 -DFOREGROUND │ │ ├─26699 apache2 -DFOREGROUND │ │ ├─26700 apache2 -DFOREGROUND │ │ ├─26702 apache2 -DFOREGROUND │ │ ├─26703 apache2 -DFOREGROUND │ │ └─26704 apache2 -DFOREGROUND │ ├─15e278b59dc154f1d0fdb535e930dbc5b777c242a23151f63e7f5c498646c72f │ │ └─23079 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml │ ├─dae06fc7bed59164da7d5a5af22ca344506fd0d534a8db23cb482a79dd6fb321 │ │ └─8413 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.17 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 │ └─35d174648d629d227fd0796560273df9bc0e182bdf0955aeccac6323d2bd59d9 │ └─5738 /pause ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount