Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─614 /lib/systemd/systemd --user │ │ └─615 (sd-pam) │ ├─session-4.scope │ │ ├─1360 sshd: zuul [priv] │ │ └─1373 sshd: zuul@notty │ └─session-1.scope │ ├─ 611 sshd: zuul [priv] │ ├─ 628 sshd: zuul@notty │ ├─ 765 /usr/bin/python3 │ ├─ 1364 ssh: /home/zuul/.ansible/cp/162.253.55.207-22-zuul [mux] │ ├─41765 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-pfircdyfbjbjlbsrmokupqlmwmzhvnuo ; /usr/bin/python3'"'"' && sleep 0' │ ├─41766 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-pfircdyfbjbjlbsrmokupqlmwmzhvnuo ; /usr/bin/python3' && sleep 0 │ ├─41767 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-pfircdyfbjbjlbsrmokupqlmwmzhvnuo ; /usr/bin/python3 │ ├─41768 /bin/sh -c echo BECOME-SUCCESS-pfircdyfbjbjlbsrmokupqlmwmzhvnuo ; /usr/bin/python3 │ ├─41769 /usr/bin/python3 │ ├─41781 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─41783 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 2882 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─ 4439 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d7573b92fef6e52bce3f16bf83b2f898122064b1455207d28f4175d2484b5579 -address /run/containerd/containerd.sock │ │ ├─ 4470 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b9ed1472f5df4f92562fe3f808323f9723e4ae7f460a19bf828a873e7a2900dd -address /run/containerd/containerd.sock │ │ ├─ 4478 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 548ad9f773119da9db1c967ccb468f2ea41e8e3142de89bf1db5237bfd6eb213 -address /run/containerd/containerd.sock │ │ ├─ 4497 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1d5b79d49a78d36fd5536cedb2ca5988c34067ebecf13360b164a705543a991a -address /run/containerd/containerd.sock │ │ ├─ 4501 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9540a6048c99c48958a9c00f27456d698d5b3c51ef104738b67771d3123b83f3 -address /run/containerd/containerd.sock │ │ ├─ 4969 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d90a00543abdb5e552642946c93f400c6f2d3e277c0dd6876a5d7892eba811ff -address /run/containerd/containerd.sock │ │ ├─ 5675 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 91175e28ed0ce4e4dda516fda1c4ed75a4ff40bf7d40ab0655b17784aa4e5986 -address /run/containerd/containerd.sock │ │ ├─ 5708 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f5abf543cc32dff76a737092278323060b1d86f8557960523da74ffacb67d0c8 -address /run/containerd/containerd.sock │ │ ├─ 6889 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id fe28c78447c969c02bfe1e2f18ee0dd1edc5f107e47a8e8016fe6f93692a7b9e -address /run/containerd/containerd.sock │ │ ├─ 7226 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id fa1135839433766b525972d82fa2bac4bbaa60689763bcaff27c30e91f9b5b46 -address /run/containerd/containerd.sock │ │ ├─ 7264 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8f005d8b60e8e717b528273e3a307394805a6705d1281749297e93bd87b478ba -address /run/containerd/containerd.sock │ │ ├─ 7615 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 55d236db81ab24d5ea34745ae7b878b07e9fcadd6771a9afa181e91e3784e951 -address /run/containerd/containerd.sock │ │ ├─ 8079 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id dfdc2106d5f958fcccb5ba627cd1220c4c1eb7edeb972035831b2d056708cb2e -address /run/containerd/containerd.sock │ │ ├─ 8105 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f6819270c3efd98d8c68783667702c9a61cec83925be9885a8756a0cb372603a -address /run/containerd/containerd.sock │ │ ├─ 8154 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1a0ba61dcfd29f6d09dda7ad351c094cfe6c9559ffd941d0fcfd3de9ee01fba1 -address /run/containerd/containerd.sock │ │ ├─10427 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 526ad8b1450f8d8a466fc62add4a1fe8dd8f6ec6cfcd7006bac35e31b5de3792 -address /run/containerd/containerd.sock │ │ ├─10484 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id cf534781f356b30b0d49f898a6b74581172e828e699db2c07486ee4e3ae0eeac -address /run/containerd/containerd.sock │ │ ├─11162 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6c07769854163510c633f7f3c4863165d577c86f62b8f0db16ff1d7a1c18a211 -address /run/containerd/containerd.sock │ │ ├─11237 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 618a12e5ff6be2911c1fbcfaba51b74fff8f9ebe8b377051780ec1e9369b96df -address /run/containerd/containerd.sock │ │ ├─11509 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6c3750cb6effc684d6e71b3de3e2a657b3ff7564a99e24bca30ec3dcaf3d7311 -address /run/containerd/containerd.sock │ │ ├─11798 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5bdb9eb6fa4f4be9cf5cb10d4b29dc43f9a036ecceb341899f49a894a769a8f7 -address /run/containerd/containerd.sock │ │ ├─12089 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0292aca31882ab742194ab993fe6a69131a62a651bdeac57a324b17880e65504 -address /run/containerd/containerd.sock │ │ ├─13883 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8bf055086dab3180a74324dbd391a1ff7ec28fe498de2da0b5e0cb2e706158a4 -address /run/containerd/containerd.sock │ │ ├─14464 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 33bf1564cd842484a2675e57d37332f53914d23856df564adc0ced59ff439e8b -address /run/containerd/containerd.sock │ │ ├─17932 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 301c5f5ca362bc757e86d2e86742f1ad1d655670f2d958ab683cd56a33833a14 -address /run/containerd/containerd.sock │ │ ├─20628 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ee0c25565467f938d5c9a1769362d2d8920cc90a34dca9321c877eedb74d5054 -address /run/containerd/containerd.sock │ │ ├─20703 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 806bb7addbb2af5fa6295d90366afa8ddc57a328400a4e6cf8193acce108ec50 -address /run/containerd/containerd.sock │ │ ├─20742 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 181e741125f4d6c20d49ea1bfc14b52c863ec97ad99d24c547c4dd0899e155ce -address /run/containerd/containerd.sock │ │ ├─20784 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9936250f794d2c0490eb49d29df61b811cd9d39f4c4d9b59ba21233f73033e9d -address /run/containerd/containerd.sock │ │ ├─21820 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id bb404f65814e191864d888627be995e091929635eeaccb913214ed7d118dc884 -address /run/containerd/containerd.sock │ │ ├─21939 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6a45e62e1f963c337ae88be3ff7b0e8e0a1155d1a5d58d16db879d8f2da2f1d6 -address /run/containerd/containerd.sock │ │ ├─23093 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0430c5932bca81ce0782f827b0d1d99883b2cc83977752feea3730b1d37fb000 -address /run/containerd/containerd.sock │ │ ├─24271 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a7e20f8d92c01ffebe2420cc4783f69d53e4bdd6ffe8b4e1cb8e9d2bc07b6e1f -address /run/containerd/containerd.sock │ │ ├─29811 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 83a005ddb4a3a16e77250f56e6ff7df12023cd556f6c1d6fbbf0a8ff7e786c66 -address /run/containerd/containerd.sock │ │ ├─29855 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c019184e875880dea235747cce0490a9c1a1a782bc2ebcf55e51fb0ea244974b -address /run/containerd/containerd.sock │ │ └─29971 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9fe4064f83c73bcc1d729321351a55ce4cff70223c5de8bc72dae6d025124e60 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1088 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─438 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─328 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─539 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─557 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─532 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─4878 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=162.253.55.207 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─297 /lib/systemd/systemd-journald │ ├─ssh.service │ │ ├─ 558 sshd: /usr/sbin/sshd -D [listener] 1 of 10-100 startups │ │ ├─41114 sshd: root [priv] │ │ └─41115 sshd: root [net] │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1285 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─331 /lib/systemd/systemd-resolved │ ├─cloud-init.service │ ├─dbus.service │ │ └─529 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─332 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─538 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─535 /lib/systemd/systemd-logind ├─k8s.io │ ├─d262e1aeb2a4bd00126919e5c0e5f14f03b85914adc77dabd189a337b54ff548 │ │ └─22817 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 │ ├─c019184e875880dea235747cce0490a9c1a1a782bc2ebcf55e51fb0ea244974b │ │ └─29878 /pause │ ├─91175e28ed0ce4e4dda516fda1c4ed75a4ff40bf7d40ab0655b17784aa4e5986 │ │ └─5710 /pause │ ├─c259bf05d0b99913209649d5c411d6313334e72f5c9e84cf4cadda55092b2ff5 │ │ └─23187 /bin/oauth2-proxy │ ├─548ad9f773119da9db1c967ccb468f2ea41e8e3142de89bf1db5237bfd6eb213 │ │ └─4598 /pause │ ├─b1fd435d79dbbc7c875d03327b4df6e671a52f3d626a0c781dbad25ee9d8e417 │ │ ├─6158 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─6479 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─715ed0ec8c96ff01e632a2ab67c88423a04e1b134f2df02d8836163e0af7e761 │ │ └─8373 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true │ ├─ee0c25565467f938d5c9a1769362d2d8920cc90a34dca9321c877eedb74d5054 │ │ └─20662 /pause │ ├─dfdc2106d5f958fcccb5ba627cd1220c4c1eb7edeb972035831b2d056708cb2e │ │ └─8117 /pause │ ├─9540a6048c99c48958a9c00f27456d698d5b3c51ef104738b67771d3123b83f3 │ │ └─4587 /pause │ ├─8bf055086dab3180a74324dbd391a1ff7ec28fe498de2da0b5e0cb2e706158a4 │ │ └─13906 /pause │ ├─e3744fca1dd7b2705a8f8ebdc1b5c859352d824c209e54fd5fdd593ce12e4801 │ │ └─14593 java -Dkc.config.built=true -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 -XX:FlightRecorderOptions=stackdepth=512 -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:MaxRAMPercentage=70 -XX:MinRAMPercentage=70 -XX:InitialRAMPercentage=50 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dpicocli.disable.closures=true -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true │ ├─c6ae451f8a06479b722e1d36248fa429a09d9bb4bb2077534072abb094e5dc24 │ │ └─22752 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.162-253-55-207.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml │ ├─ef435b546fdd9b46be4985387fa30f2cc644d49612e7edc7908e22ce75d16e35 │ │ └─4722 kube-apiserver --advertise-address=162.253.55.207 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─6a45e62e1f963c337ae88be3ff7b0e8e0a1155d1a5d58d16db879d8f2da2f1d6 │ │ └─21977 /pause │ ├─d168ed87c4fe4079edee527ff8baccf62303853dbb93ac4f515abef79aa61826 │ │ └─5959 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─d96b1ad5e25dc6fa437032799e3d49dc71940494c0ec3a86849d710222966a39 │ │ ├─12275 mysqld --wsrep_start_position=1cda1655-32e2-11f1-8b54-97463680368b:20 │ │ └─12367 /var/lib/mysql/mysql-state-monitor │ ├─cf534781f356b30b0d49f898a6b74581172e828e699db2c07486ee4e3ae0eeac │ │ └─10507 /pause │ ├─72bf50d7ce412a87e5475442280d3572a504c47bbb35eb9d87da535b3544baaa │ │ └─7319 /coredns -conf /etc/coredns/Corefile │ ├─bb404f65814e191864d888627be995e091929635eeaccb913214ed7d118dc884 │ │ └─21844 /pause │ ├─eb067556daabee7344a7d92a4745200a78937b6e4e74523383d9cca23823235f │ │ └─23169 /bin/oauth2-proxy │ ├─6c3750cb6effc684d6e71b3de3e2a657b3ff7564a99e24bca30ec3dcaf3d7311 │ │ └─11532 /pause │ ├─db96ff4afcff7bbc63fd48cc7067c0774d34d30eea8e9239fe92bd41bb9d7f4c │ │ └─4687 etcd --advertise-client-urls=https://162.253.55.207:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://162.253.55.207:2380 --initial-cluster=instance=https://162.253.55.207:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://162.253.55.207:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://162.253.55.207:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─9d20cac086abce9d52d4a8df83f5cd5da1be80d866401a92ba7caf6c53cdbd5b │ │ └─4825 /kube-vip manager │ ├─97e28788fe705442cac1580c2539813cca464367d5cef5c2ee4bf8741cd6db5c │ │ └─22858 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─2f30add17d128fbd10c0eeec7d73200cb3db3ee81b836f1fb7ef289f333a4105 │ │ └─21076 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.35.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 │ ├─9ea3b368cd67c6d32f8cad62f34cb212533824a11efe6bb94141cb3846160ccd │ │ └─6938 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─12d9973c27802f69a5300774ea8bbcd7848ec66d60784735c553bbe9ecd71d9d │ │ ├─41604 /bin/bash /tmp/horizon.sh start │ │ └─41732 python /tmp/manage.py compilemessages │ ├─0292aca31882ab742194ab993fe6a69131a62a651bdeac57a324b17880e65504 │ │ └─12114 /pause │ ├─4a55ac19e00141681551e67333ea2a97c76e23dab2fec99c8558ae796ac3804a │ │ └─8333 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.17 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 │ ├─f5abf543cc32dff76a737092278323060b1d86f8557960523da74ffacb67d0c8 │ │ └─5750 /pause │ ├─ca4f0c4fbc31d78cb3c9471e8915282d176248f2ceeb4b70579021b2703d002a │ │ └─22305 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning │ ├─0430c5932bca81ce0782f827b0d1d99883b2cc83977752feea3730b1d37fb000 │ │ └─23126 /pause │ ├─618a12e5ff6be2911c1fbcfaba51b74fff8f9ebe8b377051780ec1e9369b96df │ │ └─11260 /pause │ ├─fb927e8da11b34034e8772001fc6b1ce4d981b1f011bd3928a48fd11713b1bed │ │ ├─26546 apache2 -DFOREGROUND │ │ ├─26580 (wsgi:k -DFOREGROUND │ │ ├─26581 (wsgi:k -DFOREGROUND │ │ ├─26582 (wsgi:k -DFOREGROUND │ │ ├─26583 (wsgi:k -DFOREGROUND │ │ ├─26586 apache2 -DFOREGROUND │ │ ├─26587 apache2 -DFOREGROUND │ │ ├─26589 apache2 -DFOREGROUND │ │ ├─26590 apache2 -DFOREGROUND │ │ ├─26591 apache2 -DFOREGROUND │ │ ├─26592 apache2 -DFOREGROUND │ │ ├─26593 apache2 -DFOREGROUND │ │ ├─26594 apache2 -DFOREGROUND │ │ ├─26596 apache2 -DFOREGROUND │ │ └─26599 apache2 -DFOREGROUND │ ├─6c07769854163510c633f7f3c4863165d577c86f62b8f0db16ff1d7a1c18a211 │ │ └─11187 /pause │ ├─1d7f6c8ab1f135570c732d033407d65054c0fea871da14798516754cfd1b1049 │ │ └─23003 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.162-253-55-207.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml │ ├─55d236db81ab24d5ea34745ae7b878b07e9fcadd6771a9afa181e91e3784e951 │ │ └─7639 /pause │ ├─0f6cb48b7f8ec3700cbd349b4978617a99979363700c96d86f02ad818bd8d772 │ │ └─12629 /bin/mysqld_exporter --mysqld.username=monitor --collect.info_schema.processlist │ ├─301c5f5ca362bc757e86d2e86742f1ad1d655670f2d958ab683cd56a33833a14 │ │ └─17955 /pause │ ├─bff2206382eef6bbcf162d6dabe41eaa676722d2e3df66c74f965af5e5b03feb │ │ ├─21516 python -u /app/sidecar.py │ │ ├─21714 python -u /app/sidecar.py │ │ └─21715 python -u /app/sidecar.py │ ├─d7573b92fef6e52bce3f16bf83b2f898122064b1455207d28f4175d2484b5579 │ │ └─4584 /pause │ ├─84a6c771d62228708239151b579d727aa12206b2e216fb282de23644e6b6630a │ │ └─14298 /bin/memcached_exporter │ ├─9fe4064f83c73bcc1d729321351a55ce4cff70223c5de8bc72dae6d025124e60 │ │ └─29995 /pause │ ├─d9b7d7f66417e8e3791881cfd0d686ae74bb7475755b2f860c6364df005ab45a │ │ └─12660 /opt/percona/peer-list -on-change=/opt/percona/haproxy_add_pxc_nodes.sh -service=percona-xtradb-pxc -protocol=$(PEER_LIST_SRV_PROTOCOL) │ ├─33bf1564cd842484a2675e57d37332f53914d23856df564adc0ced59ff439e8b │ │ └─14488 /pause │ ├─1a0ba61dcfd29f6d09dda7ad351c094cfe6c9559ffd941d0fcfd3de9ee01fba1 │ │ └─8180 /pause │ ├─b7d6652427104b9f6e4529e48f2301ed317796375f102b51869dc778d9080224 │ │ └─11670 percona-xtradb-cluster-operator │ ├─4d7b4e833b2d1ea03f73e3a632fe5beb174dad4379fbd0ce2419ff5521413154 │ │ ├─21478 python -u /app/sidecar.py │ │ ├─21712 python -u /app/sidecar.py │ │ └─21713 python -u /app/sidecar.py │ ├─1d5b79d49a78d36fd5536cedb2ca5988c34067ebecf13360b164a705543a991a │ │ └─4612 /pause │ ├─b9ed1472f5df4f92562fe3f808323f9723e4ae7f460a19bf828a873e7a2900dd │ │ └─4585 /pause │ ├─d90a00543abdb5e552642946c93f400c6f2d3e277c0dd6876a5d7892eba811ff │ │ └─4994 /pause │ ├─118d2f76842ae8b811eb175e808af4b20fc5bb526c3bdc8a2df70876894e424c │ │ └─23077 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config │ ├─3ae7f8d60dae1c15b7536c8024edf129346cf279ad5993e50c5538d520972fda │ │ └─5018 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─5bdb9eb6fa4f4be9cf5cb10d4b29dc43f9a036ecceb341899f49a894a769a8f7 │ │ └─11835 /pause │ ├─ebf131258c4ee6b57b5572f2b85e1c3af771e63bb25c326a960fa8f49ecd9d8a │ │ └─11495 /manager --metrics-bind-address=:9782 │ ├─806bb7addbb2af5fa6295d90366afa8ddc57a328400a4e6cf8193acce108ec50 │ │ └─20727 /pause │ ├─fa1135839433766b525972d82fa2bac4bbaa60689763bcaff27c30e91f9b5b46 │ │ └─7249 /pause │ ├─9936250f794d2c0490eb49d29df61b811cd9d39f4c4d9b59ba21233f73033e9d │ │ └─20806 /pause │ ├─25962c2bfd342cabc3f19fc8c76b9c2a228b229e3f7417943dd0db1c9359f733 │ │ ├─23358 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─23374 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─23380 erl_child_setup 1048576 │ │ ├─23479 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─23480 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─23490 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon │ │ └─23527 /bin/sh -s rabbit_disk_monitor │ ├─b7daaec7760a0680ea13b6f66f5053660e7481b959e088989ca4a872dd9aaedd │ │ └─11339 /manager --metrics-bind-address=:8080 │ ├─ce4596c473967084fcd0d60b75858bab9e577996dcfeeb82a292d56b4d46b753 │ │ └─10803 /server │ ├─43fc1b2419351d8d8a93b1bfff2da17a9c2390201321a0db1ef7c522b1a3e9a3 │ │ └─18074 /secretgen-controller │ ├─7e280e800a939b67338f5fbf4156dab1d727d10b1d908898c5494862f992746f │ │ └─21754 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─c5cb29b7f65d8048f3f16f5985a5c08e23b554ee52d49096f5b01c377b2a89fa │ │ └─4700 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─456f8669dd8cb6512716bbcf5c154570a3dcf73526eb723d83b982da69721197 │ │ └─14241 memcached -v -p 11211 -U 0 -c 8192 -m 1024 │ ├─fd6961d80d7b7176d176d0a850dfef64e8e8e2348b64b4763dcc52cec67e4737 │ │ └─7761 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─83a005ddb4a3a16e77250f56e6ff7df12023cd556f6c1d6fbbf0a8ff7e786c66 │ │ └─29836 /pause │ ├─4e5126e533fb81621edb8391ca071091890090146abc191e09ee9a4c625a65ba │ │ └─7353 /coredns -conf /etc/coredns/Corefile │ ├─bdc7a373ff3a97b2f89644e3de48fb5bd1bd1025ecc3002d8a13bc8e441ebae4 │ │ ├─10723 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─10747 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─10854 nginx: master process /usr/bin/nginx -c /etc/nginx/nginx.conf │ │ ├─30988 nginx: worker process │ │ ├─30989 nginx: worker process │ │ └─30990 nginx: cache manager process │ ├─8f005d8b60e8e717b528273e3a307394805a6705d1281749297e93bd87b478ba │ │ └─7287 /pause │ ├─a7e20f8d92c01ffebe2420cc4783f69d53e4bdd6ffe8b4e1cb8e9d2bc07b6e1f │ │ └─24294 /pause │ ├─181e741125f4d6c20d49ea1bfc14b52c863ec97ad99d24c547c4dd0899e155ce │ │ └─20768 /pause │ ├─b5a14cd1b5a87174a7e6feb1ce4a20e0289e6309500fd0342eedd75a13467a45 │ │ └─21542 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments │ ├─fc4ccac2b89ac6b93f29064a18d942ed2ffd7f0a17a4ed63522c53d614c06f95 │ │ └─22771 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml │ ├─6668d282f746f31466a9d0dad78c22da4255be831583cb9e93a74b89b371a8fe │ │ └─8296 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager │ ├─739cb78be940f26b3f66efdb8e5e5a805c44efebdb9aaa595845eba97177be69 │ │ ├─12588 haproxy -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ │ └─13444 haproxy -sf 15 -x sockpair@4 -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ ├─f6819270c3efd98d8c68783667702c9a61cec83925be9885a8756a0cb372603a │ │ └─8139 /pause │ ├─fe28c78447c969c02bfe1e2f18ee0dd1edc5f107e47a8e8016fe6f93692a7b9e │ │ └─6912 /pause │ ├─526ad8b1450f8d8a466fc62add4a1fe8dd8f6ec6cfcd7006bac35e31b5de3792 │ │ └─10449 /pause │ └─2b511219273182552a4e4d9db46e78e789cc4892e91ac61d1ec4bbef9f8a10a1 │ └─4740 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount