Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─616 /lib/systemd/systemd --user │ │ └─617 (sd-pam) │ ├─session-4.scope │ │ ├─1364 sshd: zuul [priv] │ │ └─1377 sshd: zuul@notty │ └─session-1.scope │ ├─ 613 sshd: zuul [priv] │ ├─ 631 sshd: zuul@notty │ ├─ 770 /usr/bin/python3 │ ├─ 1368 ssh: /home/zuul/.ansible/cp/199.19.213.9-22-zuul [mux] │ ├─39119 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-popynvxblowjrgxnuxxeluyqmicsjtgi ; /usr/bin/python3'"'"' && sleep 0' │ ├─39120 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-popynvxblowjrgxnuxxeluyqmicsjtgi ; /usr/bin/python3' && sleep 0 │ ├─39121 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-popynvxblowjrgxnuxxeluyqmicsjtgi ; /usr/bin/python3 │ ├─39122 /bin/sh -c echo BECOME-SUCCESS-popynvxblowjrgxnuxxeluyqmicsjtgi ; /usr/bin/python3 │ ├─39123 /usr/bin/python3 │ ├─39124 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─39126 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 2891 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─ 4462 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e1358170e968c7d6199d8edd7bde8de6f0d1c7fcf4658ff4a533b48d347dc47f -address /run/containerd/containerd.sock │ │ ├─ 4472 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 8b90da47eb22914c82bd3064df8b309843388bc9bc928b3868ed9a9c263bb6b2 -address /run/containerd/containerd.sock │ │ ├─ 4539 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c55816254022f455489a0e7cf2f0519e05cc8f8ffc3a2903b7137256491c48cc -address /run/containerd/containerd.sock │ │ ├─ 4546 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9215f299615c4993a6124aff424b69be580ba35a3eb98b9dafd4d952e2b28d00 -address /run/containerd/containerd.sock │ │ ├─ 4558 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 121f8c85d39ec51492d198a136463286a42140771c0bacc6b3e72f347a5db6fd -address /run/containerd/containerd.sock │ │ ├─ 4986 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9bd82114e9fe33c926b49928cd1df75242649ce9c96feba0133b4325bc4f4891 -address /run/containerd/containerd.sock │ │ ├─ 5684 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 72bc8a1f000a7978d303d73802d8bcd227c6f39fe22331cd37db141b7a403132 -address /run/containerd/containerd.sock │ │ ├─ 5737 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7d11194326be97cd2bdf30fae89f9992e2a65368fca78cc43073a3482fd97ad5 -address /run/containerd/containerd.sock │ │ ├─ 6890 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 465ae0b2161bacecef1371b5f50e99676d9bfc8bf195dea146f53d3a99118dd1 -address /run/containerd/containerd.sock │ │ ├─ 7238 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 06964da8ab9995ea9cfbe5faebafeefbb907b500774fdd01991999b9ac947d09 -address /run/containerd/containerd.sock │ │ ├─ 7278 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 2c1297d4b87860b17df1353b520936ac711f05c8c0c853d10ea3a8a073299f27 -address /run/containerd/containerd.sock │ │ ├─ 7636 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id df275e8a61319e006da1a1f87df968f44c9d29b5718a988531132f7a5941871c -address /run/containerd/containerd.sock │ │ ├─ 8103 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e6d56139c14f3d044cc99bde64808bf160e7613b3c55e13802ee21d219a0d4c3 -address /run/containerd/containerd.sock │ │ ├─ 8105 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9067cfa66773128b3ff8e6653a1a52e1062f1b84d5e60a28e0e0e40c2076efcf -address /run/containerd/containerd.sock │ │ ├─ 8186 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 783470187875f4319554df18254a926787d116ec474c8d2b9a92f863cf56ba97 -address /run/containerd/containerd.sock │ │ ├─10333 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id db49fddf6adb2dea4fbfe15b9e1beb8d15cd0fa9a088bac226629ccd08f577f2 -address /run/containerd/containerd.sock │ │ ├─10418 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id afb2e48d73b3d002838ab836368d11c159cb6d400966f95c72e3ecf9350d4311 -address /run/containerd/containerd.sock │ │ ├─11109 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 49487b47f189d89fa118a5b879078ef2a0ce6546d203991ec8faa38f7d14832d -address /run/containerd/containerd.sock │ │ ├─11140 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6c4379e6130ab688c22791d8de27adb333e8f5f8b6ee6b2e8cca4dbae1d995f1 -address /run/containerd/containerd.sock │ │ ├─11593 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id dc73570cedb521518bc0a956e6d49c2c332626d56f9c12e479b2084743368e1f -address /run/containerd/containerd.sock │ │ ├─11863 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 529c79aa7dbd91f9162433738aa246768d3f0072374c6b8de7410463cc818eb8 -address /run/containerd/containerd.sock │ │ ├─12236 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 2c234389a108d9e1d15e6d4f629e0df51fe0210db03a01be500e9cd4604c9cdf -address /run/containerd/containerd.sock │ │ ├─14261 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7c600afcae6aadd8d61a26ca2dfffe859390bb2cf4ed16540d455cc9eeb85461 -address /run/containerd/containerd.sock │ │ ├─14864 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 04e1bf5c3a8c40a256e119c6ea8bf6103cda2e950dcb64fe4d5b22d8c206b29d -address /run/containerd/containerd.sock │ │ ├─18460 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 56b03628a68c9caabf20966b00c62532779ce09aecb5eb6ffeda784b306551ea -address /run/containerd/containerd.sock │ │ ├─21118 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id cf789bec3c32f3019f9595a309510949960894d7f58343c2cd4f31a9afe3270c -address /run/containerd/containerd.sock │ │ ├─21259 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 72453abac6b9da7043c626844ff76457d93638f9baaba47ce20f32da709288bc -address /run/containerd/containerd.sock │ │ ├─21312 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id f23ddb231597389c775191fbfd5326d6ba84780715e29389dd02f63217a3f267 -address /run/containerd/containerd.sock │ │ ├─21358 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id caf550d85cbfcaba6918b9954fc44de981bd4f3ad8c345797686953c5f0eb1e5 -address /run/containerd/containerd.sock │ │ ├─22504 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e195745f6425332023e771e225d8e62e2fb18a0ec1bf1f23f8cef6e8fb9acc89 -address /run/containerd/containerd.sock │ │ ├─22649 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b7bda0ea83fec73fb86a4be2242cd54d99bab450e57d6f91220e81f6c15cc8a9 -address /run/containerd/containerd.sock │ │ ├─23368 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 16060513c4d91eb8b0d6d566725c0d4f45d1c713182cdb48001e006014c53576 -address /run/containerd/containerd.sock │ │ ├─24906 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id fd8b209d3511daa1aef641f83becab7371408bc1473f71fe0f45f4ea0ba56d0b -address /run/containerd/containerd.sock │ │ ├─30433 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id ef5ca3c1901c949f7e12fd1459e28fcc5df4dc9c518de47595e057aa0cf267d5 -address /run/containerd/containerd.sock │ │ ├─30475 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5de07eb4b6d21ac595bf89031330b5e60b3f37f2a53dec96b8dc8934f69fe74e -address /run/containerd/containerd.sock │ │ └─30512 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a406e856297601ea1efa22dc9e5ae2ce61f3c1ac0970d1a1142fb4fea16e209e -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1096 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─435 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─328 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─538 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─557 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─529 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─4895 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.19.213.9 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─cloud-init-local.service │ ├─systemd-journald.service │ │ └─296 /lib/systemd/systemd-journald │ ├─ssh.service │ │ ├─ 562 sshd: /usr/sbin/sshd -D [listener] 1 of 10-100 startups │ │ ├─38473 sshd: root [priv] │ │ └─38502 sshd: root [net] │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1293 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─331 /lib/systemd/systemd-resolved │ ├─dbus.service │ │ └─526 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─332 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─535 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─532 /lib/systemd/systemd-logind ├─k8s.io │ ├─ae0e9d5565e4b66768e94ae1af704ab0b5881e582232303fd772ecdf823fb56b │ │ ├─31860 apache2 -DFOREGROUND │ │ ├─33987 (wsgi:h -DFOREGROUND │ │ ├─33988 (wsgi:h -DFOREGROUND │ │ ├─33989 (wsgi:h -DFOREGROUND │ │ ├─33993 (wsgi:h -DFOREGROUND │ │ ├─33994 (wsgi:h -DFOREGROUND │ │ ├─33996 apache2 -DFOREGROUND │ │ └─33997 apache2 -DFOREGROUND │ ├─fed80449760ee9406a1057f1bad74e399dee1b08f7e457abc6d7de1d4438c7cb │ │ ├─31834 apache2 -DFOREGROUND │ │ ├─34056 (wsgi:h -DFOREGROUND │ │ ├─34057 (wsgi:h -DFOREGROUND │ │ ├─34058 (wsgi:h -DFOREGROUND │ │ ├─34059 (wsgi:h -DFOREGROUND │ │ ├─34060 (wsgi:h -DFOREGROUND │ │ ├─34061 apache2 -DFOREGROUND │ │ └─34062 apache2 -DFOREGROUND │ ├─bc374dd02f465a00360788450ee980e65caafa607c5b63b637f6d12987e2dc37 │ │ └─21850 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.35.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 │ ├─72453abac6b9da7043c626844ff76457d93638f9baaba47ce20f32da709288bc │ │ └─21292 /pause │ ├─afb2e48d73b3d002838ab836368d11c159cb6d400966f95c72e3ecf9350d4311 │ │ └─10444 /pause │ ├─9bd82114e9fe33c926b49928cd1df75242649ce9c96feba0133b4325bc4f4891 │ │ └─5010 /pause │ ├─c55816254022f455489a0e7cf2f0519e05cc8f8ffc3a2903b7137256491c48cc │ │ └─4622 /pause │ ├─56b03628a68c9caabf20966b00c62532779ce09aecb5eb6ffeda784b306551ea │ │ └─18486 /pause │ ├─48d0b64ea1844e1927efb19a98fb8926ccc9f5d7c064032067aeea9d6f4335e1 │ │ └─5035 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─d407812295ed2fb776af70309e7b85bc9a066f3bf054d0b0929c0a5e424f1d7c │ │ └─23504 /bin/oauth2-proxy │ ├─f23ddb231597389c775191fbfd5326d6ba84780715e29389dd02f63217a3f267 │ │ └─21336 /pause │ ├─08451a0028853818bee2efb16cf8b34618546bb8dbd11f270fecd5cbbd7134e7 │ │ └─23142 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.199-19-213-9.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml │ ├─3d6ef1be4c4dd5d1de8200a2f0c36d9d35601f503b8aac444aebe328f1651f6d │ │ └─14672 /bin/memcached_exporter │ ├─db49fddf6adb2dea4fbfe15b9e1beb8d15cd0fa9a088bac226629ccd08f577f2 │ │ └─10369 /pause │ ├─033f579697e9893a9b8d92a0316e736f00e52cbb7c0736ec6ad1f002948f5bf8 │ │ └─5956 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─df275e8a61319e006da1a1f87df968f44c9d29b5718a988531132f7a5941871c │ │ └─7661 /pause │ ├─3b6e39b6ed944abc9c77d0d99dc12ef316cb1abb19562faab4c982537ed4e5b9 │ │ └─7334 /coredns -conf /etc/coredns/Corefile │ ├─465ae0b2161bacecef1371b5f50e99676d9bfc8bf195dea146f53d3a99118dd1 │ │ └─6913 /pause │ ├─e6d56139c14f3d044cc99bde64808bf160e7613b3c55e13802ee21d219a0d4c3 │ │ └─8148 /pause │ ├─f4569053fc285c73699826868b75f7770fb162a560ae84267dbd9b5f3cebfd90 │ │ └─14996 java -Dkc.config.built=true -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 -XX:FlightRecorderOptions=stackdepth=512 -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:MaxRAMPercentage=70 -XX:MinRAMPercentage=70 -XX:InitialRAMPercentage=50 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dpicocli.disable.closures=true -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true │ ├─9c7a728326c6d7c9fda079d4b34e0af5cf5da7fac8b5d88a731acdaaac70b239 │ │ └─23646 /bin/oauth2-proxy │ ├─568ddfd74618976e2f9dbd15615f3da9bb8a6ee42e7e5c24eaef2098a9b99509 │ │ └─8306 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true │ ├─caf550d85cbfcaba6918b9954fc44de981bd4f3ad8c345797686953c5f0eb1e5 │ │ └─21382 /pause │ ├─81a6f7c79fc0df9b317567fcce7767dd0186ddf019545139248fa71d05dd01d2 │ │ ├─12421 mysqld --wsrep_start_position=8d249b10-3e52-11f1-adfe-ae1bab363544:20 │ │ └─12547 /var/lib/mysql/mysql-state-monitor │ ├─121f8c85d39ec51492d198a136463286a42140771c0bacc6b3e72f347a5db6fd │ │ └─4620 /pause │ ├─6c4379e6130ab688c22791d8de27adb333e8f5f8b6ee6b2e8cca4dbae1d995f1 │ │ └─11172 /pause │ ├─e1358170e968c7d6199d8edd7bde8de6f0d1c7fcf4658ff4a533b48d347dc47f │ │ └─4508 /pause │ ├─31afc06ce2b9f2b50c59a4e892c19add8c710bccd1d4932ebf64199bb8e522ff │ │ └─14606 memcached -v -p 11211 -U 0 -c 8192 -m 1024 │ ├─38caad834f311df27d469e908ee1fc55502ae4ce3779bc846310a2c5e114347d │ │ └─22731 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning │ ├─04e1bf5c3a8c40a256e119c6ea8bf6103cda2e950dcb64fe4d5b22d8c206b29d │ │ └─14889 /pause │ ├─9215f299615c4993a6124aff424b69be580ba35a3eb98b9dafd4d952e2b28d00 │ │ └─4624 /pause │ ├─b7bda0ea83fec73fb86a4be2242cd54d99bab450e57d6f91220e81f6c15cc8a9 │ │ └─22674 /pause │ ├─e4ba9394ca17baf8814e101e66e80fc5c5a127a7e01ef5af0c1f84a2e0eac794 │ │ └─23217 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─53057a65fc0800b73122459b670472a252e8a28dec1b148d31d7bcf55eda54c4 │ │ └─11753 percona-xtradb-cluster-operator │ ├─76f4bc8ef59d51e5f84fd0482ed8445244b07080abcdd24164126b999b39b1f2 │ │ ├─12376 haproxy -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ │ └─13894 haproxy -sf 15 -x sockpair@4 -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ ├─5ceb94557ea4bded4617b9db54391c6b83a02cf11fbd47213888eb71057d3044 │ │ └─7793 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─f2c06e76844bc033bddf5bbc09a2a2ae226484746374b13fce3fed08e29e9629 │ │ └─23181 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 │ ├─5fba2a820ce40e915a54c30adee6b87a7009eebfb1ca7d30d76399398bcfe6c9 │ │ └─4760 kube-apiserver --advertise-address=199.19.213.9 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─dc73570cedb521518bc0a956e6d49c2c332626d56f9c12e479b2084743368e1f │ │ └─11618 /pause │ ├─7d11194326be97cd2bdf30fae89f9992e2a65368fca78cc43073a3482fd97ad5 │ │ └─5761 /pause │ ├─c34b63881981803ba52dbd14985e72e0f82d89bce5d06e7fe787c32eb2d01bb3 │ │ ├─21815 python -u /app/sidecar.py │ │ ├─22062 python -u /app/sidecar.py │ │ └─22063 python -u /app/sidecar.py │ ├─cf789bec3c32f3019f9595a309510949960894d7f58343c2cd4f31a9afe3270c │ │ └─21155 /pause │ ├─9067cfa66773128b3ff8e6653a1a52e1062f1b84d5e60a28e0e0e40c2076efcf │ │ └─8151 /pause │ ├─e5106cba62bcb1fd0a20f76259791964002e7665268daa07d1b9ac3c418d4c5c │ │ ├─21791 python -u /app/sidecar.py │ │ ├─22036 python -u /app/sidecar.py │ │ └─22037 python -u /app/sidecar.py │ ├─d988ac23f774da9b618aa40d873026bd160a13fa12b953b4ad5b33d1634a6f3c │ │ └─8345 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager │ ├─7158ee5192a9daf5a2ac3f067d674421c5453e16ede418d2cebc9f9ec1ba2df7 │ │ └─12639 /bin/mysqld_exporter --mysqld.username=monitor --collect.info_schema.processlist │ ├─798c7367dc2cc89a119c0b3378b3ff6e2c0cc583a357f21fcf7835e8fc65b7a9 │ │ └─18525 /secretgen-controller │ ├─fd8b209d3511daa1aef641f83becab7371408bc1473f71fe0f45f4ea0ba56d0b │ │ └─24929 /pause │ ├─45d7083d4c50d288f032ac8c36c094e5cc1266a465ad97d1fe9e1e5edf3d0783 │ │ └─22181 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─16b1f4775d376169c5fb56e85bc9a9291d610915151a5225c69d5164aee3adda │ │ └─4735 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─a406e856297601ea1efa22dc9e5ae2ce61f3c1ac0970d1a1142fb4fea16e209e │ │ └─30541 /pause │ ├─9b6969e371f0be6726fcfdd46a2a625bd7916fa1c8f98679feb23401b5e89621 │ │ ├─31841 apache2 -DFOREGROUND │ │ ├─33969 (wsgi:h -DFOREGROUND │ │ ├─33970 (wsgi:h -DFOREGROUND │ │ ├─33971 (wsgi:h -DFOREGROUND │ │ ├─33972 (wsgi:h -DFOREGROUND │ │ ├─33973 (wsgi:h -DFOREGROUND │ │ ├─33974 apache2 -DFOREGROUND │ │ └─33975 apache2 -DFOREGROUND │ ├─40f9cdd1e3240a1c2b71cfc3a5138beca74a30637d476b76bd9d696a999cb67d │ │ └─23456 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config │ ├─72bc8a1f000a7978d303d73802d8bcd227c6f39fe22331cd37db141b7a403132 │ │ └─5708 /pause │ ├─2c234389a108d9e1d15e6d4f629e0df51fe0210db03a01be500e9cd4604c9cdf │ │ └─12260 /pause │ ├─5de07eb4b6d21ac595bf89031330b5e60b3f37f2a53dec96b8dc8934f69fe74e │ │ └─30510 /pause │ ├─2f3cda7f91de0bf205b6c107f73ffe898410188d634b60ea72fce021344b6785 │ │ ├─27335 apache2 -DFOREGROUND │ │ ├─27354 (wsgi:k -DFOREGROUND │ │ ├─27355 (wsgi:k -DFOREGROUND │ │ ├─27356 (wsgi:k -DFOREGROUND │ │ ├─27357 (wsgi:k -DFOREGROUND │ │ ├─27363 apache2 -DFOREGROUND │ │ ├─27365 apache2 -DFOREGROUND │ │ ├─27366 apache2 -DFOREGROUND │ │ ├─27367 apache2 -DFOREGROUND │ │ ├─27368 apache2 -DFOREGROUND │ │ ├─27369 apache2 -DFOREGROUND │ │ ├─27370 apache2 -DFOREGROUND │ │ ├─27371 apache2 -DFOREGROUND │ │ ├─27372 apache2 -DFOREGROUND │ │ └─27373 apache2 -DFOREGROUND │ ├─f631f89682eaaf420fb8a33f3fb234f9baedab1a6c37960246b074cded66f117 │ │ └─23575 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml │ ├─e195745f6425332023e771e225d8e62e2fb18a0ec1bf1f23f8cef6e8fb9acc89 │ │ └─22529 /pause │ ├─ef5ca3c1901c949f7e12fd1459e28fcc5df4dc9c518de47595e057aa0cf267d5 │ │ └─30457 /pause │ ├─2c1297d4b87860b17df1353b520936ac711f05c8c0c853d10ea3a8a073299f27 │ │ └─7301 /pause │ ├─06964da8ab9995ea9cfbe5faebafeefbb907b500774fdd01991999b9ac947d09 │ │ └─7263 /pause │ ├─1024aea4aedf94351d416c4367d9c4fa1e2b76bebdf2ec1659a43269dec0f034 │ │ └─10721 /server │ ├─068763db672ef469d3e78c0ea6cc3111e2483eb70eae1ffa4cdf387101622e0c │ │ └─21623 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments │ ├─529c79aa7dbd91f9162433738aa246768d3f0072374c6b8de7410463cc818eb8 │ │ └─11888 /pause │ ├─cc1a4632dfb34b8f0593934723ee69667a3e54f8a4d37cda6a5a393cdf47529c │ │ └─6948 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─babc04d44bcb5a6b5f70b8fb5113ccc3177d50cd7a130bcdc4cf37cf27c25262 │ │ └─4734 etcd --advertise-client-urls=https://199.19.213.9:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.19.213.9:2380 --initial-cluster=instance=https://199.19.213.9:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.19.213.9:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.19.213.9:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─8b90da47eb22914c82bd3064df8b309843388bc9bc928b3868ed9a9c263bb6b2 │ │ └─4543 /pause │ ├─92e1957758856de5fbfb3aa612b3e10b5b9e6be57afa3e97d31692c7291363a6 │ │ └─11401 /manager --metrics-bind-address=:9782 │ ├─b3bc3379a490948c6a8997525fe06b34bd2a0705afab9a702f95efd3c86df86d │ │ └─4847 /kube-vip manager │ ├─5ab2176745d6a022d3ca987a62b86851bb4eaa4ffdd03739ee53a92e1f65d214 │ │ ├─10641 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─10655 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─10786 nginx: master process /usr/bin/nginx -c /etc/nginx/nginx.conf │ │ ├─31614 nginx: worker process │ │ ├─31616 nginx: worker process │ │ └─31617 nginx: cache manager process │ ├─145e5241db812dad6487175ba19c539be7d390968faa255e3223c491520c6689 │ │ └─11482 /manager --metrics-bind-address=:8080 │ ├─037d5f02ce40a88358d2d1cf37b8010c218f7b3ac99df56345e4b1152cfe9363 │ │ ├─6160 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─6526 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─16060513c4d91eb8b0d6d566725c0d4f45d1c713182cdb48001e006014c53576 │ │ └─23394 /pause │ ├─f4ae60b861335476449a587c599b9031b1052db0bc0f633931606947cf468a9e │ │ ├─23780 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─23797 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─23803 erl_child_setup 1048576 │ │ ├─23847 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─23848 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─23858 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon │ │ └─23906 /bin/sh -s rabbit_disk_monitor │ ├─336ccb3070a6b06fe6cc87d9e6f7258c7cadbac9dd9a08b10a48b86408a15a94 │ │ └─7362 /coredns -conf /etc/coredns/Corefile │ ├─e9ba74b70bcf16fc49c1f05fae6bd3193bdda362425dd14fbb95b4796790f8c4 │ │ └─8384 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.17 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 │ ├─abb11647cfa24d21c4b30fd292684c7d4557857dce022b4bba5e0a26e2f24911 │ │ └─12519 /opt/percona/peer-list -on-change=/opt/percona/haproxy_add_pxc_nodes.sh -service=percona-xtradb-pxc -protocol=$(PEER_LIST_SRV_PROTOCOL) │ ├─fdff55935b93f7ed49308d5e47a92bafbc0d5cc0f522df467d5f1e25d93517ea │ │ └─4698 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─49487b47f189d89fa118a5b879078ef2a0ce6546d203991ec8faa38f7d14832d │ │ └─11144 /pause │ ├─783470187875f4319554df18254a926787d116ec474c8d2b9a92f863cf56ba97 │ │ └─8211 /pause │ ├─7c600afcae6aadd8d61a26ca2dfffe859390bb2cf4ed16540d455cc9eeb85461 │ │ └─14283 /pause │ └─fd8b89b3aa1b454c949c44e6b0a27021b81acdc50b722dc5076b07d81b21c394 │ └─23421 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.199-19-213-9.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount