Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─615 /lib/systemd/systemd --user │ │ └─616 (sd-pam) │ ├─session-4.scope │ │ ├─1352 sshd: zuul [priv] │ │ └─1365 sshd: zuul@notty │ └─session-1.scope │ ├─ 612 sshd: zuul [priv] │ ├─ 629 sshd: zuul@notty │ ├─ 766 /usr/bin/python3 │ ├─ 1356 ssh: /home/zuul/.ansible/cp/199.204.45.233-22-zuul [mux] │ ├─38434 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-gifftshxhwawatruzcnpzavefpskchfk ; /usr/bin/python3'"'"' && sleep 0' │ ├─38435 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-gifftshxhwawatruzcnpzavefpskchfk ; /usr/bin/python3' && sleep 0 │ ├─38436 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-gifftshxhwawatruzcnpzavefpskchfk ; /usr/bin/python3 │ ├─38437 /bin/sh -c echo BECOME-SUCCESS-gifftshxhwawatruzcnpzavefpskchfk ; /usr/bin/python3 │ ├─38438 /usr/bin/python3 │ ├─38439 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─38441 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 2876 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─ 4438 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id dc234c8e854d38868dd0d7da8303890a6c2a718d03c1042e0d96ab85bfbd4ab4 -address /run/containerd/containerd.sock │ │ ├─ 4446 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5444dda947da23a71651ef3ed5c1b94ea6d6face16558638a8b3da83839d127d -address /run/containerd/containerd.sock │ │ ├─ 4467 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e9c535926b33d8859c8a624ae5bf3a17ddaea992a94b5ea5af5e93038a8e2ea9 -address /run/containerd/containerd.sock │ │ ├─ 4473 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a54aa110cda875f718831e3dd6c33986b10ce4994aeb97b26756f617d2f23784 -address /run/containerd/containerd.sock │ │ ├─ 4488 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5d58bff8d6a2e772fcc9b68c2f8446f14e1d0cc875da8e9fb68b537ba525d225 -address /run/containerd/containerd.sock │ │ ├─ 4998 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3e26acef51efcabc75ab2f19d511575852e085fa9accdaa2dd7a1f4691b84662 -address /run/containerd/containerd.sock │ │ ├─ 5652 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7396fbb34989975e707cf182f1e8604b79751c742f043fec95b169bd82694cfb -address /run/containerd/containerd.sock │ │ ├─ 5669 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0cca1ae40782c7758aa46db1a83f16662d3e5bbe4e1356d0afc945767d638723 -address /run/containerd/containerd.sock │ │ ├─ 7246 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 159ce2e5a41bda7bced5138a7bdb06f3e57e0232d93cf20417a0d61c4c46297f -address /run/containerd/containerd.sock │ │ ├─ 7594 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a36844e6645e1b5d2286ec56a7bbf89e21976f92edcbab3f4df1eb1d0ce00340 -address /run/containerd/containerd.sock │ │ ├─ 7628 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 26082e18cba93b967c568c085ae310383744ba8d1957e03f8b523b692de78dc1 -address /run/containerd/containerd.sock │ │ ├─ 7985 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e79841bc13a5f529825e27a8e9b28431928f8ab13dc196e1f181070a8afc7c36 -address /run/containerd/containerd.sock │ │ ├─ 8477 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 104751202346d8c091f0c789e80804d827972cfc1e8cb5b42ba9502903cc76d8 -address /run/containerd/containerd.sock │ │ ├─ 8516 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id bb7e0b8cc6620e793c7a0333b0b5a4ca706644bc20ddd6ceb82a4b83cdbf349e -address /run/containerd/containerd.sock │ │ ├─ 8566 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 36353d40a65ba91047d9db19bf6c7ccf082ec12e7df7ba981a44667d0788c98b -address /run/containerd/containerd.sock │ │ ├─10758 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0f6389e54fea3c9b19f13c97152e3e8a1175aab47944e1c65fda26464a2db070 -address /run/containerd/containerd.sock │ │ ├─10803 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b162e555f51361252b89eb744fc11a20926a13e541cb568a67c646a291247277 -address /run/containerd/containerd.sock │ │ ├─11418 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e0d5f09e8d35b8ab6f3f19a8fcde214140e32afd66cb2d5d7c3e04bb04bb0597 -address /run/containerd/containerd.sock │ │ ├─11526 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 44d835cb02104b6414c66496c3121a745e9313cfa7c791b936735c3a989d4211 -address /run/containerd/containerd.sock │ │ ├─11822 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7b49a2642d8b4b3fffd3d2bf7f29f1c2931af909a6975e8ec419a24e7e652a8f -address /run/containerd/containerd.sock │ │ ├─12221 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 06e0e5327a50e729ae3750ed9169cf075bd97bdc66959d44d3c97ad1b7264e20 -address /run/containerd/containerd.sock │ │ ├─12559 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5324d47b830b4b609de3d9aa48dc052b424f77d4cc21d8d20defc9422e0583cd -address /run/containerd/containerd.sock │ │ ├─14239 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 028f31cd9a68fd110babce0119101c3a4bff3b32033ba93a150a271df3798605 -address /run/containerd/containerd.sock │ │ ├─14889 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e14631171e7874f60f4b46a566916c6e91bcbcf5abc19c56c2a11e1b2f09f48a -address /run/containerd/containerd.sock │ │ ├─18200 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9a1ac0ab25ab67c4a0d7b960b04981dede15a281903ba7b0323a186969c5725c -address /run/containerd/containerd.sock │ │ ├─20751 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c6f444bbe9cc9c31e0c00917fc5ae21609a33ad68d5db6141e3d96ba096ee19e -address /run/containerd/containerd.sock │ │ ├─20886 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 95369ad47a36e911195285061d509b312f88fa1357e00b82634442a42a2828c2 -address /run/containerd/containerd.sock │ │ ├─20931 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6e191cd8bda38c5a45519b2f4bdca05c6633a1521cae613b4eed60a718ae1ac8 -address /run/containerd/containerd.sock │ │ ├─20958 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 68ff97d1477021afd9688da60bd5943b64db0a180f7f8080e4b91acd2f54a520 -address /run/containerd/containerd.sock │ │ ├─22115 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id b0c5ed6e7227f289f06affa658bcbea4db8525b8416a3deefa9a0216fe3f8ca9 -address /run/containerd/containerd.sock │ │ ├─22195 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d423602897954a5b6abfabddecbaef9795441e49c4a1fc6e1efbd58c28038e46 -address /run/containerd/containerd.sock │ │ ├─23074 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6249799958434239dd105b8a2c09bd92863aeb9452aee3e717a10e0ac6a2910a -address /run/containerd/containerd.sock │ │ ├─24376 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 40c85262375f6c5a21ad5ea97662c1b3e7853f12b4d7c492aea5fbb78a824b19 -address /run/containerd/containerd.sock │ │ ├─29859 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a2b04d086f0ee283a6998eb1f14013d93382c1de29dddbfa5779f4604c132264 -address /run/containerd/containerd.sock │ │ ├─29899 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 95c68d794e8c51a88a68ff6c78a4aa81fe812fb73cf072810d23a4bff1c682ed -address /run/containerd/containerd.sock │ │ └─29994 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 24195e363cf2878bd1233126845e2b505e2415c65dd3096e33f79a8165799e78 -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1089 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─437 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─327 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─543 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─557 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─533 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─4864 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.233 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─cloud-init-local.service │ ├─systemd-journald.service │ │ └─294 /lib/systemd/systemd-journald │ ├─ssh.service │ │ ├─ 561 sshd: /usr/sbin/sshd -D [listener] 1 of 10-100 startups │ │ ├─38378 sshd: root [priv] │ │ └─38379 sshd: root [net] │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1285 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─331 /lib/systemd/systemd-resolved │ ├─cloud-init.service │ ├─dbus.service │ │ └─530 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─333 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─539 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─536 /lib/systemd/systemd-logind ├─k8s.io │ ├─726d969e1e4301e4fa624e4fe7e7b00a9daa51117cfc38475bbdc7fcd693693a │ │ └─10939 /server │ ├─5444dda947da23a71651ef3ed5c1b94ea6d6face16558638a8b3da83839d127d │ │ └─4603 /pause │ ├─1e0bd170a7a019dfa2e16e3189ee9357e1486e62ba623112c4195f2b8cb7468d │ │ └─12970 /bin/mysqld_exporter --mysqld.username=monitor --collect.info_schema.processlist │ ├─06e0e5327a50e729ae3750ed9169cf075bd97bdc66959d44d3c97ad1b7264e20 │ │ └─12247 /pause │ ├─954e47b82b6375d7329e405a0789906169804fe78cb2a932f8300f8817669ab7 │ │ └─22967 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 │ ├─8e267907d08ff949ada157674559d149880c1e9ff138887bf46e348a8ddf2b8b │ │ └─11600 /manager --metrics-bind-address=:9782 │ ├─9ce89220de16250546e94a5bfec5473d400db9c1c4cc97d9f7950c36513a27ff │ │ └─14801 /bin/memcached_exporter │ ├─3e26acef51efcabc75ab2f19d511575852e085fa9accdaa2dd7a1f4691b84662 │ │ └─5023 /pause │ ├─159ce2e5a41bda7bced5138a7bdb06f3e57e0232d93cf20417a0d61c4c46297f │ │ └─7270 /pause │ ├─69a3f502186cafb99f7dd688911d9f9bb97389202f04df2c0117b25eaf485773 │ │ └─21542 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments │ ├─8ba78912689470a94933136842107afe31e3b47a1cacc612b526f2b9a3a739a8 │ │ ├─23406 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─23422 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─23428 erl_child_setup 1048576 │ │ ├─23459 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─23460 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─23470 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon │ │ └─23518 /bin/sh -s rabbit_disk_monitor │ ├─0f6389e54fea3c9b19f13c97152e3e8a1175aab47944e1c65fda26464a2db070 │ │ └─10783 /pause │ ├─b00fa04d46f182fd01f2f9fe7b5543ba1232cebf244db1e10eadb79fdd43d1d2 │ │ └─4714 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ ├─4741550f80226870e91738163cecd46b18854b42c297ec8e90c8e3e1ffe1c3a6 │ │ └─21917 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─7396fbb34989975e707cf182f1e8604b79751c742f043fec95b169bd82694cfb │ │ └─5704 /pause │ ├─6249799958434239dd105b8a2c09bd92863aeb9452aee3e717a10e0ac6a2910a │ │ └─23096 /pause │ ├─ffc305353843ffb1149a79550622c37b0635270099f4201a1e770e7fc24d3a36 │ │ └─11692 /manager --metrics-bind-address=:8080 │ ├─e0b1ca81195f0c5e66f8eaa461d8b2711158b150a527beec67186da24bd6b2b6 │ │ └─22786 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config │ ├─2c95e6209560b2cbb77ad73493d020f95cd1bb2cbeb273cee5d0ac6f7c58f930 │ │ └─4819 /kube-vip manager │ ├─1ea04b6bab340fa564269f44889241d9417a7da5f947aa3280e571ced10783e3 │ │ ├─31261 apache2 -DFOREGROUND │ │ ├─33416 (wsgi:h -DFOREGROUND │ │ ├─33417 (wsgi:h -DFOREGROUND │ │ ├─33418 (wsgi:h -DFOREGROUND │ │ ├─33419 (wsgi:h -DFOREGROUND │ │ ├─33420 (wsgi:h -DFOREGROUND │ │ ├─33421 apache2 -DFOREGROUND │ │ └─33422 apache2 -DFOREGROUND │ ├─85ccbe643d70fa018a3af029e8bde1f40b82767d2caa7233eaab15adfaa74e38 │ │ └─18366 /secretgen-controller │ ├─dd4cb49ef443adb5bf5f0abe4d889f7f7ec769639d8a849aca2927590d927732 │ │ └─22752 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.199-204-45-233.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml │ ├─a36844e6645e1b5d2286ec56a7bbf89e21976f92edcbab3f4df1eb1d0ce00340 │ │ └─7625 /pause │ ├─575cda3ed80180edcbc9a6d9ec79d7581336affa49dda92a3518090c7da8cde9 │ │ └─7707 /coredns -conf /etc/coredns/Corefile │ ├─c6f444bbe9cc9c31e0c00917fc5ae21609a33ad68d5db6141e3d96ba096ee19e │ │ └─20812 /pause │ ├─7313f67a949d7c30a6cdbbba78743412a2217919cbea06be21028f01775abab3 │ │ ├─21288 python -u /app/sidecar.py │ │ ├─21558 python -u /app/sidecar.py │ │ └─21559 python -u /app/sidecar.py │ ├─5324d47b830b4b609de3d9aa48dc052b424f77d4cc21d8d20defc9422e0583cd │ │ └─12582 /pause │ ├─1aef6f8b61e1c034bd724861b3cd63064f6aa2a4d37dcd2c33e13208f18812b9 │ │ └─23240 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml │ ├─b0c5ed6e7227f289f06affa658bcbea4db8525b8416a3deefa9a0216fe3f8ca9 │ │ └─22139 /pause │ ├─58e7941927b5adc7de97819f17e3b8d33b776c55fa43b522ce37bcb27f037cc8 │ │ └─7687 /coredns -conf /etc/coredns/Corefile │ ├─c0b3ecd5b28426a3e8eeef85ebee64674e31cd95b0662088cdb03af03aa78764 │ │ └─6327 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─ac3b25d19577ead51de8ddb701b47f97bebccb24bb8e2acb64e6b4c4d7bf8790 │ │ └─22933 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.199-204-45-233.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml │ ├─5ca1791517346892cea9d64f8099416b614450e8d8b5457d3f22243cfd0ceb7e │ │ └─8131 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─c9da96e9833fa0cfccc8d1222f932376197d451ef887c6d4a019e60a53d282c9 │ │ ├─6527 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─6864 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─dc234c8e854d38868dd0d7da8303890a6c2a718d03c1042e0d96ab85bfbd4ab4 │ │ └─4598 /pause │ ├─104751202346d8c091f0c789e80804d827972cfc1e8cb5b42ba9502903cc76d8 │ │ └─8502 /pause │ ├─68ff97d1477021afd9688da60bd5943b64db0a180f7f8080e4b91acd2f54a520 │ │ └─20995 /pause │ ├─5d58bff8d6a2e772fcc9b68c2f8446f14e1d0cc875da8e9fb68b537ba525d225 │ │ └─4577 /pause │ ├─26082e18cba93b967c568c085ae310383744ba8d1957e03f8b523b692de78dc1 │ │ └─7657 /pause │ ├─f8df0144bf3b02a5cfa436d29552add5fe7e60a8ba46496fff9c71bbdfc76a5d │ │ └─7297 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─95c68d794e8c51a88a68ff6c78a4aa81fe812fb73cf072810d23a4bff1c682ed │ │ └─29922 /pause │ ├─e14631171e7874f60f4b46a566916c6e91bcbcf5abc19c56c2a11e1b2f09f48a │ │ └─14913 /pause │ ├─40c85262375f6c5a21ad5ea97662c1b3e7853f12b4d7c492aea5fbb78a824b19 │ │ └─24410 /pause │ ├─95369ad47a36e911195285061d509b312f88fa1357e00b82634442a42a2828c2 │ │ └─20910 /pause │ ├─e79841bc13a5f529825e27a8e9b28431928f8ab13dc196e1f181070a8afc7c36 │ │ └─8010 /pause │ ├─028f31cd9a68fd110babce0119101c3a4bff3b32033ba93a150a271df3798605 │ │ └─14263 /pause │ ├─3437175fe24f9aeb95577e15b19b2f20c907831d6e7a4ba313801ccdb3c52883 │ │ ├─21168 python -u /app/sidecar.py │ │ ├─21432 python -u /app/sidecar.py │ │ └─21433 python -u /app/sidecar.py │ ├─a2b04d086f0ee283a6998eb1f14013d93382c1de29dddbfa5779f4604c132264 │ │ └─29884 /pause │ ├─6a42efcad8caab4cedce2c71c7e35ff882aa180b1fbf4612a1e4de699866cc78 │ │ └─21351 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.35.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 │ ├─e145ce0e9f489872cc287b7ea4c4228db815421887ff472addd8ed1963086f12 │ │ ├─31252 apache2 -DFOREGROUND │ │ ├─33293 (wsgi:h -DFOREGROUND │ │ ├─33294 (wsgi:h -DFOREGROUND │ │ ├─33298 (wsgi:h -DFOREGROUND │ │ ├─33299 (wsgi:h -DFOREGROUND │ │ ├─33300 (wsgi:h -DFOREGROUND │ │ ├─33301 apache2 -DFOREGROUND │ │ └─33302 apache2 -DFOREGROUND │ ├─24195e363cf2878bd1233126845e2b505e2415c65dd3096e33f79a8165799e78 │ │ └─30043 /pause │ ├─a17dbbeac52da80d5c2416dcabc0b7efcf0b868a57dda15a71843df0e1b5071a │ │ └─23115 /bin/oauth2-proxy │ ├─b162e555f51361252b89eb744fc11a20926a13e541cb568a67c646a291247277 │ │ └─10828 /pause │ ├─e9c535926b33d8859c8a624ae5bf3a17ddaea992a94b5ea5af5e93038a8e2ea9 │ │ └─4569 /pause │ ├─a91671d06e280dfd3f59ab8312a33f5251f71008af021a8d8d1b4abe17f39e04 │ │ └─4754 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─36353d40a65ba91047d9db19bf6c7ccf082ec12e7df7ba981a44667d0788c98b │ │ └─8591 /pause │ ├─1d12f3c30a193b7cc1265900121f0ffcbc18732e347420285e5dc52b5431900d │ │ └─5052 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─44d835cb02104b6414c66496c3121a745e9313cfa7c791b936735c3a989d4211 │ │ └─11549 /pause │ ├─e526d42df8dfcb9c2cc0966b9e87168595c1d1081e0cb830bfe0c2119bf18717 │ │ └─23147 /bin/oauth2-proxy │ ├─0da826d961b6e63356259da4d0c48cca20b0eacfdf337892d3d173cb5edf9291 │ │ ├─11064 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─11078 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─11170 nginx: master process /usr/bin/nginx -c /etc/nginx/nginx.conf │ │ ├─31012 nginx: worker process │ │ ├─31013 nginx: worker process │ │ └─31014 nginx: cache manager process │ ├─a54aa110cda875f718831e3dd6c33986b10ce4994aeb97b26756f617d2f23784 │ │ └─4570 /pause │ ├─5ba5e06a16d5b34c86e82e1a6f66a2200d2fb417f66362add8b17862a464d6bd │ │ ├─26528 apache2 -DFOREGROUND │ │ ├─26542 (wsgi:k -DFOREGROUND │ │ ├─26543 (wsgi:k -DFOREGROUND │ │ ├─26544 (wsgi:k -DFOREGROUND │ │ ├─26545 (wsgi:k -DFOREGROUND │ │ ├─26547 apache2 -DFOREGROUND │ │ ├─26548 apache2 -DFOREGROUND │ │ ├─26550 apache2 -DFOREGROUND │ │ ├─26551 apache2 -DFOREGROUND │ │ ├─26552 apache2 -DFOREGROUND │ │ ├─26554 apache2 -DFOREGROUND │ │ ├─26556 apache2 -DFOREGROUND │ │ ├─26559 apache2 -DFOREGROUND │ │ ├─26560 apache2 -DFOREGROUND │ │ └─26561 apache2 -DFOREGROUND │ ├─bddaa40a268f00eddaed8d9b80ec06833388a3e69603b9bb47b04f6d90599978 │ │ └─8765 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.17 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 │ ├─bb7e0b8cc6620e793c7a0333b0b5a4ca706644bc20ddd6ceb82a4b83cdbf349e │ │ └─8539 /pause │ ├─4ca60ad5b00000742dc9e7973222201ff47d01d95d065a18e076eebaf33d94dd │ │ ├─12709 haproxy -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ │ └─13898 haproxy -sf 15 -x sockpair@4 -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ ├─9f8aa16ee99246fd0c9dcb3f2b7e4540d25380338cb792aa15a5ed735c279e99 │ │ └─4676 kube-apiserver --advertise-address=199.204.45.233 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─ba2e44073ca74878f6edd845689f3c8609579a27249fb7eedf64cf606e420f3d │ │ └─15012 java -Dkc.config.built=true -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 -XX:FlightRecorderOptions=stackdepth=512 -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:MaxRAMPercentage=70 -XX:MinRAMPercentage=70 -XX:InitialRAMPercentage=50 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dpicocli.disable.closures=true -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true │ ├─031d82626a9ee0bf046f085c893e3fa3a642767531c53f0547ab1ed881e1fb9e │ │ ├─31237 apache2 -DFOREGROUND │ │ ├─33253 (wsgi:h -DFOREGROUND │ │ ├─33254 (wsgi:h -DFOREGROUND │ │ ├─33255 (wsgi:h -DFOREGROUND │ │ ├─33256 (wsgi:h -DFOREGROUND │ │ ├─33257 (wsgi:h -DFOREGROUND │ │ ├─33258 apache2 -DFOREGROUND │ │ └─33259 apache2 -DFOREGROUND │ ├─44755f87cdbcbd57651f90c014a3dd863aa49dd3f06cf0a66cb3372061bdec07 │ │ └─4695 etcd --advertise-client-urls=https://199.204.45.233:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.233:2380 --initial-cluster=instance=https://199.204.45.233:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.233:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.233:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─e0d5f09e8d35b8ab6f3f19a8fcde214140e32afd66cb2d5d7c3e04bb04bb0597 │ │ └─11443 /pause │ ├─d423602897954a5b6abfabddecbaef9795441e49c4a1fc6e1efbd58c28038e46 │ │ └─22219 /pause │ ├─0cca1ae40782c7758aa46db1a83f16662d3e5bbe4e1356d0afc945767d638723 │ │ └─5718 /pause │ ├─6e191cd8bda38c5a45519b2f4bdca05c6633a1521cae613b4eed60a718ae1ac8 │ │ └─20965 /pause │ ├─ecfb4c86045ded6eeea6fcd79773241b6faf88aab0048e1d99f694a69eb515b6 │ │ └─14703 memcached -v -p 11211 -U 0 -c 8192 -m 1024 │ ├─7b49a2642d8b4b3fffd3d2bf7f29f1c2931af909a6975e8ec419a24e7e652a8f │ │ └─11845 /pause │ ├─934fed137a70be0838ab760eea713d313e4ac9a6bcec7826eb21288224c912aa │ │ └─8686 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager │ ├─e51a2998805926263931c099b4ba2a56b4459f9d475318084206734d08dfdaec │ │ └─11967 percona-xtradb-cluster-operator │ ├─e8a9dfd07465461fc59f42c608e1e95ae850c8aaae544d2a0bff5e9486705a2b │ │ ├─12756 mysqld --wsrep_start_position=2d529dd3-3bf3-11f1-9921-4b74ddcdd606:20 │ │ └─12882 /var/lib/mysql/mysql-state-monitor │ ├─de73bcdf2d3a1d893d9f90b6b168058053983af140b9a4ee9ce3c32345fe518f │ │ └─23011 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─9a1ac0ab25ab67c4a0d7b960b04981dede15a281903ba7b0323a186969c5725c │ │ └─18225 /pause │ ├─8af2baa7d74b4bdf8024686efeafc21d62c86b86966d51713670ceae8ddc8dbb │ │ └─8725 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true │ ├─02912234d7e6e3847e70d2ddf7a140513c8dc249d245c1672eb3586321913e09 │ │ └─12852 /opt/percona/peer-list -on-change=/opt/percona/haproxy_add_pxc_nodes.sh -service=percona-xtradb-pxc -protocol=$(PEER_LIST_SRV_PROTOCOL) │ └─136276294359d45052ecf16d6b5bcd018577be9adb19a9e47fc05ee166299a1a │ └─22308 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount