Control group /: -.slice ├─sys-fs-fuse-connections.mount ├─sys-kernel-config.mount ├─sys-kernel-debug.mount ├─dev-mqueue.mount ├─user.slice │ └─user-1000.slice │ ├─user@1000.service … │ │ ├─app.slice │ │ └─init.scope │ │ ├─620 /lib/systemd/systemd --user │ │ └─621 (sd-pam) │ ├─session-4.scope │ │ ├─1366 sshd: zuul [priv] │ │ └─1379 sshd: zuul@notty │ └─session-1.scope │ ├─ 617 sshd: zuul [priv] │ ├─ 634 sshd: zuul@notty │ ├─ 771 /usr/bin/python3 │ ├─ 1370 ssh: /home/zuul/.ansible/cp/162.253.55.207-22-zuul [mux] │ ├─42438 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-bufftiaoqpxqwlbmvwjdskokgprcpbmw ; /usr/bin/python3'"'"' && sleep 0' │ ├─42439 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-bufftiaoqpxqwlbmvwjdskokgprcpbmw ; /usr/bin/python3' && sleep 0 │ ├─42440 sudo -H -S -n -u root /bin/sh -c echo BECOME-SUCCESS-bufftiaoqpxqwlbmvwjdskokgprcpbmw ; /usr/bin/python3 │ ├─42441 /bin/sh -c echo BECOME-SUCCESS-bufftiaoqpxqwlbmvwjdskokgprcpbmw ; /usr/bin/python3 │ ├─42442 /usr/bin/python3 │ ├─42443 /bin/bash -c set -x systemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi │ └─42445 systemd-cgls --full --all --no-pager ├─sys-kernel-tracing.mount ├─init.scope │ └─1 /lib/systemd/systemd --system --deserialize 39 nofb ├─system.slice │ ├─containerd.service … │ │ ├─ 2892 /usr/bin/containerd --config /etc/containerd/config.toml │ │ ├─ 4464 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e57cc1762bf432d71242161f2ae4ad75e8078388716a3a5c9c6df2bd614608b1 -address /run/containerd/containerd.sock │ │ ├─ 4502 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 79ea4d3732df22a638f9707642fc59dc2ffff095cd2b01008e17e59057c01998 -address /run/containerd/containerd.sock │ │ ├─ 4511 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id bd77b368f383d51c42a24fcbe9d1b40dbaeca1d2865019dd5955b4028d1dc724 -address /run/containerd/containerd.sock │ │ ├─ 4513 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 558fce71a534ddb6fba05dd320feb548a18a2eb156a8f813a653ee102b44b9f5 -address /run/containerd/containerd.sock │ │ ├─ 4541 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 67253f8c964bfbfb74c3906bf98beac07176cf89f38f4db0578bf4e0969d7cab -address /run/containerd/containerd.sock │ │ ├─ 5059 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 50040042906ac64405c8d71b4aab3a7d0ec16ed1182a23925f6e3db6351796eb -address /run/containerd/containerd.sock │ │ ├─ 5683 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e3d9d10ad84c75f24572c894af9cf8618894613bbc74ba063c4923caa6d90773 -address /run/containerd/containerd.sock │ │ ├─ 5724 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1b7b9b0beac57b30061778fd7dc1762ac13e5edaea823f338882b0b274af4bff -address /run/containerd/containerd.sock │ │ ├─ 6940 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 955482f8efc3c9bd49360051bc83f8f4319972c5a2b5e646328360f108c15189 -address /run/containerd/containerd.sock │ │ ├─ 7297 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9518fb093184988be7ca89ba40709ef028a501f0be2cf970250840147daa92fe -address /run/containerd/containerd.sock │ │ ├─ 7328 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 148c2ef0f83dac5acd4204d3af79290cec92b6e05b3d54876edd0a900fd2f68a -address /run/containerd/containerd.sock │ │ ├─ 7691 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6be30197890350fca21d6155a7224b2f160f7d87cd195e161c4d3955f18b25cf -address /run/containerd/containerd.sock │ │ ├─ 8155 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 47eaf34225cfc381b6878cb03be51dfb0acbf7ac978ea10c8e7c64f1b22b8995 -address /run/containerd/containerd.sock │ │ ├─ 8195 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3228665e075791b8c716e585bcdd41d54a58ba4ca98b383c4befe61ef368c068 -address /run/containerd/containerd.sock │ │ ├─ 8234 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id a6a75415da665682250df9ac3ac3d2dadca9d099232285f6abd26f91b9271430 -address /run/containerd/containerd.sock │ │ ├─10409 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 148b589678e6e2837a70e5d9f1d66ce48d730fec10ca8f8b70efd1fe1b6e4800 -address /run/containerd/containerd.sock │ │ ├─10472 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 16b07971332db164dca0d62ab060650e2750a7f3d9e8a523dca7b9212a2d5910 -address /run/containerd/containerd.sock │ │ ├─11261 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 548379e434adb2a2565fadf82c8fd128d3c61c69f58cf2e855d7705e3b1c6e2a -address /run/containerd/containerd.sock │ │ ├─11370 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3f6978fda0876020edcc1e31a8f86838267bfd7c15e5878be7d53b87272920be -address /run/containerd/containerd.sock │ │ ├─11681 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 98fcbcf8ffb8ac8b6583cd7b1610397a123ff5479c191f11db3d4e616b199c9d -address /run/containerd/containerd.sock │ │ ├─11947 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0658c995251e3ed5fa62609f69352b39418658875ad80b69c0aef403bea424dc -address /run/containerd/containerd.sock │ │ ├─12326 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0eed1c937a9e6c205277d009af1381167eaebb35efb33920f93af2820f71cd3c -address /run/containerd/containerd.sock │ │ ├─14216 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 984cb1be6f06757154ba06098af59e484def76c3e9e1789a9a3e69393c2ad6a5 -address /run/containerd/containerd.sock │ │ ├─14704 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id e30573187bf229811fddd6ec3ef11f22d7bc1c67d9f752294c18781cff38a5c9 -address /run/containerd/containerd.sock │ │ ├─18537 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id c4ce808b00392fc8a551e4ff773fa1f1906308a70424d38f66b341850d345ec7 -address /run/containerd/containerd.sock │ │ ├─21273 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0c56d133009e13db1e15982e08c885b17e8d3da67a1cc28d8cb0d09bda238121 -address /run/containerd/containerd.sock │ │ ├─21342 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9b94ff320e21362210c67631f9e13c7c9f428fe01364ffc277e586593eac14c1 -address /run/containerd/containerd.sock │ │ ├─21385 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 19ecb91af8aee3e8397d3627516e9a840f778d681f50ab23c1a1df910da53402 -address /run/containerd/containerd.sock │ │ ├─21423 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 783d36ecbb754c703cc8918026716a1f5cafb58348614438b4bea30e02cc805c -address /run/containerd/containerd.sock │ │ ├─22496 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6abbaf6ea9fcc754ed44f129dee8aa45cd2202b9e64db033ab616a8f8af4998b -address /run/containerd/containerd.sock │ │ ├─22553 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id d77fd0d4fbdeee0fbdf899c02edf93d4f9eb80045b43a5b8888b3a9cbe4a7fca -address /run/containerd/containerd.sock │ │ ├─23558 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 3b67819f658b6fdaf8cbd00b22142cc16a4e17d8918261306ee849c1aab88ee2 -address /run/containerd/containerd.sock │ │ ├─25088 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 9303babe47f96e11f3bee09bc0906ca0b1d84f6ad8c20c63eca1c407ce8b5cd5 -address /run/containerd/containerd.sock │ │ ├─30525 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id fea9efaeaa9d9814db56c7dbe5c343d575c13d6b403f1e189f3166bc7fc0d4dd -address /run/containerd/containerd.sock │ │ ├─30568 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 337dfefbdac7cfe07581e16144d0942e72155c318449a8b86c420fe3e8539c96 -address /run/containerd/containerd.sock │ │ └─30639 /usr/bin/containerd-shim-runc-v2 -namespace k8s.io -id 5b31f9bcb437f9dc714e0d1088174feefb746432aea2090803bd135fcda77ffa -address /run/containerd/containerd.sock │ ├─packagekit.service │ │ └─1094 /usr/libexec/packagekitd │ ├─systemd-networkd.service │ │ └─438 /lib/systemd/systemd-networkd │ ├─systemd-udevd.service │ │ └─328 /lib/systemd/systemd-udevd │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─542 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 │ ├─polkit.service │ │ └─563 /usr/libexec/polkitd --no-debug │ ├─networkd-dispatcher.service │ │ └─533 /usr/bin/python3 /usr/bin/networkd-dispatcher --run-startup-triggers │ ├─kubelet.service │ │ └─4902 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=162.253.55.207 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 │ ├─system-modprobe.slice │ ├─systemd-journald.service │ │ └─295 /lib/systemd/systemd-journald │ ├─ssh.service │ │ ├─ 562 sshd: /usr/sbin/sshd -D [listener] 1 of 10-100 startups │ │ ├─41684 sshd: root [priv] │ │ └─41685 sshd: root [net] │ ├─cloud-final.service │ ├─uuidd.service │ │ └─1297 /usr/sbin/uuidd --socket-activation │ ├─cloud-config.service │ ├─systemd-resolved.service │ │ └─331 /lib/systemd/systemd-resolved │ ├─cloud-init.service │ ├─dbus.service │ │ └─530 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─systemd-timesyncd.service │ │ └─332 /lib/systemd/systemd-timesyncd │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─539 /sbin/agetty -o -p -- \u --noclear tty1 linux │ └─systemd-logind.service │ └─536 /lib/systemd/systemd-logind ├─k8s.io │ ├─e57cc1762bf432d71242161f2ae4ad75e8078388716a3a5c9c6df2bd614608b1 │ │ └─4556 /pause │ ├─cd894a09048dd2b2b753e0968f1c165a803cf88589457b4fa1973abfcff6cb4d │ │ └─5108 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance │ ├─a6a75415da665682250df9ac3ac3d2dadca9d099232285f6abd26f91b9271430 │ │ └─8264 /pause │ ├─3b67819f658b6fdaf8cbd00b22142cc16a4e17d8918261306ee849c1aab88ee2 │ │ └─23583 /pause │ ├─148b589678e6e2837a70e5d9f1d66ce48d730fec10ca8f8b70efd1fe1b6e4800 │ │ └─10436 /pause │ ├─955482f8efc3c9bd49360051bc83f8f4319972c5a2b5e646328360f108c15189 │ │ └─6964 /pause │ ├─8eb6045c721659a9d8366f696936c6bfeaf1337177e81f513283b261b6b5d666 │ │ └─11538 /manager --metrics-bind-address=:8080 │ ├─9cba56a29bcc706089d8f0b682bbfb60d3dbd669ca2e12550ca56b9d4bc98c99 │ │ ├─12555 mysqld --wsrep_start_position=885b5082-31fb-11f1-8181-d6e3be228c62:20 │ │ └─12677 /var/lib/mysql/mysql-state-monitor │ ├─c5f536357b4697ab04239ce6aa9a66f14657104246ee8575edc7fd02f073fcd1 │ │ └─12457 /opt/percona/peer-list -on-change=/opt/percona/haproxy_add_pxc_nodes.sh -service=percona-xtradb-pxc -protocol=$(PEER_LIST_SRV_PROTOCOL) │ ├─fb68f3f574e39cc74ece35fc2d960498b717e56fad33f946df50e118c2e7e521 │ │ └─6091 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false │ ├─e3d9d10ad84c75f24572c894af9cf8618894613bbc74ba063c4923caa6d90773 │ │ └─5713 /pause │ ├─df02c6435c9ac8b33012339ba2ad92cd435f25af6b06f3c74ae54034b37d7492 │ │ └─7412 /coredns -conf /etc/coredns/Corefile │ ├─0825092dd32270481f100a4847c98feaa76c24b5f679bdfcdb8a4474a737f3bc │ │ └─22602 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─fc6797077d24448fd2410cc51ef987eed3c11055311cad2d7e5c74817199194d │ │ └─23731 /bin/oauth2-proxy │ ├─2d3973f4376d0b528d37325184c9e785d714de288a2e020b6b24a8cc9e2e7f8e │ │ └─21978 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments │ ├─79ea4d3732df22a638f9707642fc59dc2ffff095cd2b01008e17e59057c01998 │ │ └─4616 /pause │ ├─5137ccad6da4c9fdee90acf2e6efae79b1a546a80ff1a5d372d4942874f31d79 │ │ └─7392 /coredns -conf /etc/coredns/Corefile │ ├─0eed1c937a9e6c205277d009af1381167eaebb35efb33920f93af2820f71cd3c │ │ └─12350 /pause │ ├─61a7bb4184b0bc6aec758618f09945bb3ca3df72c9579641586e996323218f2d │ │ ├─10703 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─10715 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true │ │ ├─10843 nginx: master process /usr/bin/nginx -c /etc/nginx/nginx.conf │ │ ├─31671 nginx: worker process │ │ ├─31672 nginx: worker process │ │ └─31673 nginx: cache manager process │ ├─0eb20a44ad869bb8847b632f6c487152fe0164838aa3b8354c4f9140a33aa0a2 │ │ └─10765 /server │ ├─783d36ecbb754c703cc8918026716a1f5cafb58348614438b4bea30e02cc805c │ │ └─21454 /pause │ ├─be2ee6bc33fa1bbf3694ea77471eb71807c7d37031516af3b9ae9f5c131e6174 │ │ └─8372 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.17 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 │ ├─0c56d133009e13db1e15982e08c885b17e8d3da67a1cc28d8cb0d09bda238121 │ │ └─21309 /pause │ ├─9303babe47f96e11f3bee09bc0906ca0b1d84f6ad8c20c63eca1c407ce8b5cd5 │ │ └─25120 /pause │ ├─1571e5d16ee9380f65e42c020cf5078400508106d8d29fa2ae797bf842f73349 │ │ └─8411 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true │ ├─266626f9a83472c5ab703595bf9b6ea7aa2b18b489f787503568f81ae5351c5f │ │ └─23352 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.162-253-55-207.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml │ ├─9b94ff320e21362210c67631f9e13c7c9f428fe01364ffc277e586593eac14c1 │ │ └─21366 /pause │ ├─c4ce808b00392fc8a551e4ff773fa1f1906308a70424d38f66b341850d345ec7 │ │ └─18561 /pause │ ├─485561e2b160b66e1858abe7c1d80b2b4653645f4365c000f8357cd496cea767 │ │ └─6999 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml │ ├─f8eca47b37e6f37efd8e2426f6adb534e16649dbbca26240b16aa6155a9e57f4 │ │ ├─24030 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server │ │ ├─24047 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- │ │ ├─24053 erl_child_setup 1048576 │ │ ├─24084 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─24085 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 │ │ ├─24095 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon │ │ └─24167 /bin/sh -s rabbit_disk_monitor │ ├─38edf572f1719d4861c6ff251f3c84b7eae4ca04b8284fce55510a715d8de1df │ │ └─12653 /bin/mysqld_exporter --mysqld.username=monitor --collect.info_schema.processlist │ ├─48b0343a21824b90035cac117f63815f095ca4d362c141ca7f727ad6b597dbea │ │ └─4855 /kube-vip manager │ ├─a23676ced07bf12d9050dfa75c6897698b154ae5d73b2ea2890ce69df6d6cb84 │ │ └─11824 percona-xtradb-cluster-operator │ ├─d14eab3a65d798d79d5b16350eac66801ebadbcb3ffbe1416c190f9dbfccb925 │ │ └─23812 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml │ ├─fea9efaeaa9d9814db56c7dbe5c343d575c13d6b403f1e189f3166bc7fc0d4dd │ │ └─30549 /pause │ ├─da0a31af9c2e6729b40c256527518915bea03f7c48901ffcb1c7100d5e839c2a │ │ └─4763 kube-apiserver --advertise-address=162.253.55.207 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ ├─98fcbcf8ffb8ac8b6583cd7b1610397a123ff5479c191f11db3d4e616b199c9d │ │ └─11706 /pause │ ├─cf148535b031a10f384453b88654968eee1f2c2f0613b35d7f3b079b6c0ab405 │ │ ├─27347 apache2 -DFOREGROUND │ │ ├─27361 (wsgi:k -DFOREGROUND │ │ ├─27362 (wsgi:k -DFOREGROUND │ │ ├─27363 (wsgi:k -DFOREGROUND │ │ ├─27364 (wsgi:k -DFOREGROUND │ │ ├─27368 apache2 -DFOREGROUND │ │ ├─27369 apache2 -DFOREGROUND │ │ ├─27370 apache2 -DFOREGROUND │ │ ├─27373 apache2 -DFOREGROUND │ │ ├─27374 apache2 -DFOREGROUND │ │ ├─27375 apache2 -DFOREGROUND │ │ ├─27377 apache2 -DFOREGROUND │ │ ├─27378 apache2 -DFOREGROUND │ │ ├─27379 apache2 -DFOREGROUND │ │ └─27380 apache2 -DFOREGROUND │ ├─9c4d66cc776b2600960333ef2f4b9bfd90e60c125552a4b08884e1f2ed4de2d4 │ │ ├─22378 python -u /app/sidecar.py │ │ ├─22429 python -u /app/sidecar.py │ │ └─22430 python -u /app/sidecar.py │ ├─16b07971332db164dca0d62ab060650e2750a7f3d9e8a523dca7b9212a2d5910 │ │ └─10496 /pause │ ├─67d40bcd1affdf8d7a3c02497f9dbd87f5af618809bec19dfbccb54abc8114ef │ │ └─14929 java -Dkc.config.built=true -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 -XX:FlightRecorderOptions=stackdepth=512 -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:MaxRAMPercentage=70 -XX:MinRAMPercentage=70 -XX:InitialRAMPercentage=50 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dpicocli.disable.closures=true -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true │ ├─7b836f3264c8d88ea295c7e39268f193eca6775706de56d467fe43c9f6a932d6 │ │ ├─22323 python -u /app/sidecar.py │ │ ├─22424 python -u /app/sidecar.py │ │ └─22426 python -u /app/sidecar.py │ ├─6be30197890350fca21d6155a7224b2f160f7d87cd195e161c4d3955f18b25cf │ │ └─7714 /pause │ ├─ba4542681124c56d0438d2a3bc8e9fbe4c6d65f1c28e4935b30922124d511c55 │ │ ├─6205 cilium-agent --config-dir=/tmp/cilium/config-map │ │ └─6587 cilium-health-responder --listen 4240 --pidfile /var/run/cilium/state/health-endpoint.pid │ ├─47eaf34225cfc381b6878cb03be51dfb0acbf7ac978ea10c8e7c64f1b22b8995 │ │ └─8181 /pause │ ├─337dfefbdac7cfe07581e16144d0942e72155c318449a8b86c420fe3e8539c96 │ │ └─30594 /pause │ ├─148c2ef0f83dac5acd4204d3af79290cec92b6e05b3d54876edd0a900fd2f68a │ │ └─7359 /pause │ ├─3228665e075791b8c716e585bcdd41d54a58ba4ca98b383c4befe61ef368c068 │ │ └─8220 /pause │ ├─bd77b368f383d51c42a24fcbe9d1b40dbaeca1d2865019dd5955b4028d1dc724 │ │ └─4639 /pause │ ├─19ecb91af8aee3e8397d3627516e9a840f778d681f50ab23c1a1df910da53402 │ │ └─21409 /pause │ ├─9518fb093184988be7ca89ba40709ef028a501f0be2cf970250840147daa92fe │ │ └─7335 /pause │ ├─558fce71a534ddb6fba05dd320feb548a18a2eb156a8f813a653ee102b44b9f5 │ │ └─4608 /pause │ ├─91173a0d0c50ab0ee8763d7c394195e00b76fffaa1c628812312b1d0b0c82038 │ │ └─22939 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning │ ├─6177575d7f11fcabaf23d9d5b81e602305d03bc8453eecc5a5470e9f26b89852 │ │ └─23875 /bin/oauth2-proxy │ ├─c465d8785b8e3591fa54a5753d22e546132d96f8718bebaf9b5f58a18d701ef4 │ │ └─11435 /manager --metrics-bind-address=:9782 │ ├─832263aa2021a7166eef33e5dce5f494bb24400c3ac3f2fc6aaf3418e120712b │ │ └─8450 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager │ ├─af58b08d328958aefed2395e5d5a042b20f521f692a98b0d96aa689bdb4c7be2 │ │ └─18611 /secretgen-controller │ ├─be258460d7528dc487ad0abeffb12827dec31d3cbf2b04186cd84c6ee1d27e80 │ │ └─23681 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key │ ├─67253f8c964bfbfb74c3906bf98beac07176cf89f38f4db0578bf4e0969d7cab │ │ └─4633 /pause │ ├─ea1682bdc1490d43714e72e031dd592972adae1afc0062c7bf8627f20e767813 │ │ └─23385 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config │ ├─3f6978fda0876020edcc1e31a8f86838267bfd7c15e5878be7d53b87272920be │ │ └─11393 /pause │ ├─fdb6de14348ae210ea79e3cc320f2b52f3230fcd45eae5231bd235f5d8df6f8d │ │ └─7836 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config │ ├─e30573187bf229811fddd6ec3ef11f22d7bc1c67d9f752294c18781cff38a5c9 │ │ └─14727 /pause │ ├─f16a170560ad3e5ccea174a586c4a555ef63c39a1793cdcf439049af0b1a8eed │ │ └─15148 memcached -v -p 11211 -U 0 -c 8192 -m 1024 │ ├─984cb1be6f06757154ba06098af59e484def76c3e9e1789a9a3e69393c2ad6a5 │ │ └─14240 /pause │ ├─65a4a4bf7f0d9cee37bda68561a1ec253e1fa114b35b6845c3c090c5d78fb1dc │ │ ├─12377 haproxy -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ │ └─15102 haproxy -sf 240 -x sockpair@5 -W -db -f /etc/haproxy-custom/haproxy-global.cfg -f /etc/haproxy/pxc/haproxy.cfg -p /etc/haproxy/pxc/haproxy.pid -S /etc/haproxy/pxc/haproxy-main.sock │ ├─04ba7e2e68f562f93891686891b399d2071645467895574f440fa5ad88c18680 │ │ └─21561 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.35.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 │ ├─0658c995251e3ed5fa62609f69352b39418658875ad80b69c0aef403bea424dc │ │ └─11971 /pause │ ├─6abbaf6ea9fcc754ed44f129dee8aa45cd2202b9e64db033ab616a8f8af4998b │ │ └─22521 /pause │ ├─1b7b9b0beac57b30061778fd7dc1762ac13e5edaea823f338882b0b274af4bff │ │ └─5757 /pause │ ├─05a9074aa5a02a4aa0fa9a9d6c2605ca3dc82600eea23e6bd11c1cacd1238d9d │ │ └─23645 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 │ ├─5b31f9bcb437f9dc714e0d1088174feefb746432aea2090803bd135fcda77ffa │ │ └─30664 /pause │ ├─8c532a53884aef52f647385feaaf8ddc6169ac41b3449a7e848344ae70bf8fd5 │ │ └─4729 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 │ ├─50040042906ac64405c8d71b4aab3a7d0ec16ed1182a23925f6e3db6351796eb │ │ └─5080 /pause │ ├─548379e434adb2a2565fadf82c8fd128d3c61c69f58cf2e855d7705e3b1c6e2a │ │ └─11285 /pause │ ├─d77fd0d4fbdeee0fbdf899c02edf93d4f9eb80045b43a5b8888b3a9cbe4a7fca │ │ └─22576 /pause │ ├─f3963464957399a1369bea823313ab5abf61ea3f907be576c2b65322dc31b6e3 │ │ └─4711 etcd --advertise-client-urls=https://162.253.55.207:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://162.253.55.207:2380 --initial-cluster=instance=https://162.253.55.207:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://162.253.55.207:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://162.253.55.207:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ ├─497154b21b2d33382137fa3e2ed9b88276434f670d916fb4f82a03b574761542 │ │ └─15216 /bin/memcached_exporter │ ├─823547bad085d12aaf682715b081d360986d3e4abec76d58673cf61c3466d738 │ │ └─4761 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true │ └─41e305e605ab2b6c34300af5493057273ebc100c35ce758c59ae28ba78051d4a │ └─23612 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.162-253-55-207.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml ├─proc-sys-fs-binfmt_misc.mount └─dev-hugepages.mount