USER PID %CPU %MEM VSZ RSS TTY STAT START TIME COMMAND root 39276 18.5 1.9 2496352 1255336 ? Ssl 01:58 5:19 kube-apiserver --advertise-address=199.204.45.229 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key ubuntu 47901 9.8 1.5 8816272 988536 ? Ssl 02:02 2:24 mysqld --wsrep_start_position=5ee5014e-2b13-11f1-bb5a-2a443ae76054:20 167 12653 4.8 0.8 4218504 564328 ? Sl 01:51 1:41 /usr/bin/ceph-mgr -n mgr.instance.qtdepf -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug zuul 67432 3.3 0.8 1790508 537720 ? Ssl 02:10 0:33 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.199-204-45-229.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml zuul 50511 4.9 0.7 3838548 491332 ? Ssl 02:04 1:07 java -Dkc.config.built=true -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 -XX:FlightRecorderOptions=stackdepth=512 -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:MaxRAMPercentage=70 -XX:MinRAMPercentage=70 -XX:InitialRAMPercentage=50 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dpicocli.disable.closures=true -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true 42424 73415 0.0 0.3 289652 261196 ? Ss 02:12 0:00 apache2 -DFOREGROUND 10001 67140 1.3 0.3 1637636 204528 ? Ssl 02:10 0:13 /usr/bin/loki -config.file=/etc/loki/config/config.yaml -target=all root 39331 6.6 0.2 11359212 189064 ? Ssl 01:58 1:54 etcd --advertise-client-urls=https://199.204.45.229:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.229:2380 --initial-cluster=instance=https://199.204.45.229:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.229:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.229:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt 472 64435 1.0 0.2 1688220 176948 ? Ssl 02:09 0:10 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning 999 68512 1.6 0.2 4279676 173240 ? Sl 02:10 0:16 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-14.2.5.1/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false 999 94528 2.3 0.2 4278164 171876 ? Sl 02:18 0:11 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-14.2.5.1/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false root 39467 9.2 0.2 3179504 165960 ? Ssl 01:58 2:37 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.229 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 999 78983 1.6 0.2 4281660 165956 ? Sl 02:14 0:12 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-14.2.5.1/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false root 9795 11.3 0.2 3415384 164448 ? Ssl 01:50 4:11 /usr/bin/containerd --config /etc/containerd/config.toml root 66367 1.5 0.2 699348 160868 ? Ssl 02:10 0:15 /usr/bin/vector --config-dir /etc/vector/ root 39234 2.4 0.2 1412560 157784 ? Ssl 01:58 0:41 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true root 40882 2.2 0.2 1398168 157104 ? Ssl 02:00 0:36 cilium-agent --config-dir=/tmp/cilium/config-map 42424 108901 0.3 0.2 388720 149308 ? Sl 02:22 0:00 staffeln-conductor: Staffeln conductor backup controller worker(0) 42424 107122 1.5 0.2 1188876 148728 ? S 02:21 0:05 glance-api: uWSGI worker 2 42424 107121 1.3 0.2 250576 144204 ? S 02:21 0:04 glance-api: uWSGI worker 1 42424 108647 1.1 0.1 218104 131548 ? Ssl 02:21 0:03 staffeln-conductor: master process [/var/lib/openstack/bin/staffeln-conductor --config-file /etc/staffeln/staffeln.conf] 42424 73429 1.1 0.1 543040 124768 ? Sl 02:12 0:10 (wsgi:k -DFOREGROUND 42424 108904 0.0 0.1 374820 124528 ? Sl 02:22 0:00 staffeln-conductor: Staffeln conductor rotation controller worker(0) 42424 73432 1.4 0.1 542016 123172 ? Sl 02:12 0:12 (wsgi:k -DFOREGROUND 42424 73431 1.3 0.1 542016 122144 ? Sl 02:12 0:11 (wsgi:k -DFOREGROUND 42424 73430 1.1 0.1 540992 122104 ? Sl 02:12 0:09 (wsgi:k -DFOREGROUND 167 28813 0.6 0.1 626936 113376 ? Sl 01:55 0:11 /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug 65532 41770 0.2 0.1 1414804 111156 ? Ssl 02:00 0:04 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml 167 98084 0.4 0.1 5204496 106036 ? Ssl 02:19 0:02 radosgw --fsid=4837cbf8-4f90-4300-b3f6-726c9b9f89b4 --keyring=/etc/ceph/keyring-store/keyring --default-log-to-stderr=true --default-err-to-stderr=true --default-mon-cluster-log-to-stderr=true --default-log-stderr-prefix=debug --default-log-to-file=false --default-mon-cluster-log-to-file=false --mon-host=[v2:10.96.240.200:3300,v1:10.96.240.200:6789] --mon-initial-members=instance --id=rgw.ceph.a --setuser=ceph --setgroup=ceph --foreground --rgw-frontends=beast port=8080 --host=rook-ceph-rgw-ceph-a-699b8bdb59-f8t5n --rgw-mime-types-file=/etc/ceph/rgw/mime.types --rgw-realm=ceph --rgw-zonegroup=ceph --rgw-zone=ceph 167 24278 0.5 0.1 618732 102920 ? Sl 01:54 0:10 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug 167 33609 0.5 0.1 618728 102172 ? Sl 01:56 0:10 /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug 42424 85536 0.5 0.1 113804 93832 ? S 02:16 0:03 barbiacan-api: uWSGI worker 1 2016 88164 0.2 0.1 5492616 83704 ? Ssl 02:17 0:01 /usr/local/bin/rook ceph operator zuul 43440 0.2 0.1 1269728 80132 ? Ssl 02:00 0:04 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager root 40573 0.3 0.1 1306792 78584 ? Ssl 01:59 0:06 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false root 10165 2.5 0.1 2467412 75544 ? Ssl 01:50 0:56 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock zuul 43349 0.4 0.1 1291748 74292 ? Ssl 02:00 0:07 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.17 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 167 12285 0.9 0.1 288272 72152 ? Sl 01:51 0:19 /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true root 39301 0.5 0.1 1285140 68452 ? Ssl 01:58 0:09 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 bin 47163 1.2 0.1 1281628 68304 ? Ssl 02:01 0:19 percona-xtradb-cluster-operator nobody 61400 3.7 0.1 1287064 66524 ? Ssl 02:09 0:39 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.35.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 systemd+ 45869 0.3 0.1 1283008 66360 ? Ssl 02:01 0:05 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true 472 61990 0.2 0.0 70024 64136 ? Ss 02:09 0:02 python -u /app/sidecar.py 472 62015 0.2 0.0 70024 64112 ? Ss 02:09 0:02 python -u /app/sidecar.py 472 62428 0.1 0.0 72660 63708 ? S 02:09 0:01 python -u /app/sidecar.py 472 62430 0.0 0.0 70472 61496 ? S 02:09 0:00 python -u /app/sidecar.py 472 62429 0.0 0.0 70472 61156 ? S 02:09 0:00 python -u /app/sidecar.py 472 62427 0.0 0.0 70472 61104 ? S 02:09 0:00 python -u /app/sidecar.py root 42195 0.4 0.0 766704 57000 ? Ssl 02:00 0:07 /coredns -conf /etc/coredns/Corefile root 39613 0.0 0.0 1283460 56164 ? Ssl 01:58 0:01 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance root 42245 0.3 0.0 767216 55948 ? Ssl 02:00 0:05 /coredns -conf /etc/coredns/Corefile nobody 56097 0.1 0.0 1281608 50876 ? Ssl 02:07 0:01 nfd-master -crd-controller=true -metrics=8081 zuul 43395 0.1 0.0 1282112 46948 ? Ssl 02:00 0:02 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true ubuntu 46715 0.2 0.0 1271576 45952 ? Ssl 02:01 0:03 /manager --metrics-bind-address=:9782 nobody 62265 0.3 0.0 1271012 45024 ? Ssl 02:09 0:03 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments zuul 67474 0.0 0.0 1270924 44740 ? Ssl 02:10 0:00 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 zuul 66746 0.0 0.0 1268068 43280 ? Ssl 02:10 0:00 /goldpinger --static-file-path /static zuul 67240 0.0 0.0 1270924 41588 ? Ssl 02:10 0:00 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config zuul 67197 0.0 0.0 1257960 40904 ? Ssl 02:10 0:00 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.199-204-45-229.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml ubuntu 46861 0.4 0.0 1268572 39604 ? Ssl 02:01 0:06 /manager --metrics-bind-address=:8080 zuul 56212 0.1 0.0 1266300 39440 ? Ssl 02:07 0:01 /secretgen-controller nobody 56025 0.2 0.0 2592228 37592 ? Ssl 02:07 0:02 nfd-worker -metrics=8081 root 39412 0.3 0.0 1273180 37416 ? Ssl 01:58 0:05 /kube-vip manager nobody 56063 0.0 0.0 1266792 29716 ? Ssl 02:07 0:00 nfd-gc -gc-interval=1h nobody 55276 0.1 0.0 1262368 29532 ? Ssl 02:07 0:02 /kubernetes-entrypoint nobody 108273 0.1 0.0 1262116 29444 ? Ssl 02:21 0:00 kubernetes-entrypoint root 42876 0.0 0.0 735700 29052 ? Ssl 02:00 0:01 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config zuul 67519 0.0 0.0 1262788 28624 ? Ssl 02:10 0:00 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key nobody 62745 0.0 0.0 1262468 28408 ? Ssl 02:09 0:00 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key nobody 66859 0.0 0.0 1241188 24072 ? Ssl 02:10 0:00 /bin/pushgateway root 120187 0.0 0.0 106484 23064 ? Sl 02:27 0:00 /usr/bin/python3 nobody 64386 0.5 0.0 1241444 21732 ? Ssl 02:09 0:05 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml root 28759 0.0 0.0 736008 21568 ? Sl 01:55 0:00 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-1 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1:/var/lib/ceph/osd/ceph-1:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 12193 0.0 0.0 735496 21400 ? Sl 01:51 0:00 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mon-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance:/var/lib/ceph/mon/ceph-instance:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev quay.io/ceph/ceph:v18.2.7 -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true root 24224 0.0 0.0 736264 21396 ? Sl 01:54 0:00 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-0 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0:/var/lib/ceph/osd/ceph-0:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 19064 0.0 0.0 736148 21140 ? Sl 01:53 0:00 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-crash --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-crash-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/config:/etc/ceph/ceph.conf:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/keyring:/etc/ceph/ceph.client.crash.instance.keyring quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n client.crash.instance root 33549 0.0 0.0 735752 20976 ? Sl 01:56 0:00 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-2 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2:/var/lib/ceph/osd/ceph-2:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 12552 0.0 0.0 736148 20536 ? Sl 01:51 0:00 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mgr --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mgr-instance-qtdepf --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.qtdepf:/var/lib/ceph/mgr/ceph-instance.qtdepf:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.qtdepf/config:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v18.2.7 -n mgr.instance.qtdepf -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug zuul 68410 0.0 0.0 1248992 20312 ? Ssl 02:10 0:00 /bin/oauth2-proxy root 1187 0.0 0.0 296148 20220 ? Ssl 01:47 0:00 /usr/libexec/packagekitd zuul 68429 0.0 0.0 1249248 19776 ? Ssl 02:10 0:00 /bin/oauth2-proxy root 360 0.4 0.0 74028 19700 ? S /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi root 98188 0.0 0.0 4384 3396 ? Ss 02:19 0:00 /bin/bash -x -e -m -c CEPH_CLIENT_ID=ceph-client.rgw.ceph.a PERIODICITY=daily LOG_ROTATE_CEPH_FILE=/etc/logrotate.d/ceph LOG_MAX_SIZE=500M ROTATE=7 # edit the logrotate file to only rotate a specific daemon log # otherwise we will logrotate log files without reloading certain daemons # this might happen when multiple daemons run on the same machine sed -i "s|*.log|$CEPH_CLIENT_ID.log|" "$LOG_ROTATE_CEPH_FILE" # replace default daily with given user input sed --in-place "s/daily/$PERIODICITY/g" "$LOG_ROTATE_CEPH_FILE" # replace rotate count, default 7 for all ceph daemons other than rbd-mirror sed --in-place "s/rotate 7/rotate $ROTATE/g" "$LOG_ROTATE_CEPH_FILE" if [ "$LOG_MAX_SIZE" != "0" ]; then ?# adding maxsize $LOG_MAX_SIZE at the 4th line of the logrotate config file with 4 spaces to maintain indentation ?sed --in-place "4i \ \ \ \ maxsize $LOG_MAX_SIZE" "$LOG_ROTATE_CEPH_FILE" fi while true; do ?# we don't force the logrorate but we let the logrotate binary handle the rotation based on user's input for periodicity and size ?logrotate --verbose "$LOG_ROTATE_CEPH_FILE" ?sleep 15m done zuul 708 0.0 0.0 168464 2936 ? S 01:47 0:00 (sd-pam) ubuntu 48028 0.0 0.0 1226264 2404 ? Sl 02:02 0:00 /var/lib/mysql/mysql-state-monitor 999 78957 0.0 0.0 2900 1848 ? Ss 02:14 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 94511 0.0 0.0 2900 1836 ? Ss 02:18 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 68495 0.0 0.0 2896 1712 ? Ss 02:10 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server uuidd 1388 0.0 0.0 9620 1488 ? Ss 01:47 0:00 /usr/sbin/uuidd --socket-activation 999 79110 0.0 0.0 3740 1320 ? Ss 02:14 0:00 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/inet_gethost 4 root 98282 0.0 0.0 4940 1312 ? S+ 02:19 0:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 15m 999 68574 0.0 0.0 3736 1248 ? Ss 02:10 0:00 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/inet_gethost 4 999 94611 0.0 0.0 3740 1200 ? Ss 02:18 0:00 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/inet_gethost 4 root 629 0.0 0.0 6220 1096 ttyS0 Ss+ 01:45 0:00 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 999 79185 0.0 0.0 2900 1080 ? Ss 02:14 0:00 /bin/sh -s rabbit_disk_monitor root 628 0.0 0.0 6176 1052 tty1 Ss+ 01:45 0:00 /sbin/agetty -o -p -- \u --noclear tty1 linux 999 68518 0.0 0.0 2784 1052 ? Ss 02:10 0:00 erl_child_setup 1048576 999 79041 0.0 0.0 2784 1052 ? Ss 02:14 0:00 /opt/erlang/lib/erlang/lib/os_mon-2.9.1/priv/bin/cpu_sup 999 94609 0.0 0.0 2652 1036 ? Ss 02:18 0:00 /opt/erlang/lib/erlang/lib/os_mon-2.9.1/priv/bin/memsup 999 94867 0.0 0.0 2900 1024 ? Ss 02:18 0:00 /bin/sh -s rabbit_disk_monitor 999 68568 0.0 0.0 2780 996 ? Ss 02:10 0:00 /opt/erlang/lib/erlang/lib/os_mon-2.9.1/priv/bin/cpu_sup 999 68846 0.0 0.0 2896 988 ? Ss 02:10 0:00 /bin/sh -s rabbit_disk_monitor root 120186 0.0 0.0 2892 988 ? S 02:27 0:00 /bin/sh -c echo BECOME-SUCCESS-hxridrwlkewcsusjucybeidsvdweuath ; /usr/bin/python3 zuul 120183 0.0 0.0 2892 976 ? Ss 02:27 0:00 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-hxridrwlkewcsusjucybeidsvdweuath ; /usr/bin/python3'"'"' && sleep 0' 999 94610 0.0 0.0 2784 964 ? Ss 02:18 0:00 /opt/erlang/lib/erlang/lib/os_mon-2.9.1/priv/bin/cpu_sup 999 68565 0.0 0.0 2896 960 ? Ss 02:10 0:00 sh -s disksup 999 78989 0.0 0.0 2788 960 ? Ss 02:14 0:00 erl_child_setup 1048576 999 79038 0.0 0.0 2900 960 ? Ss 02:14 0:00 sh -s disksup 999 94607 0.0 0.0 2900 960 ? Ss 02:18 0:00 sh -s disksup zuul 120184 0.0 0.0 2892 960 ? S 02:27 0:00 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-hxridrwlkewcsusjucybeidsvdweuath ; /usr/bin/python3' && sleep 0 999 68567 0.0 0.0 2648 956 ? Ss 02:10 0:00 /opt/erlang/lib/erlang/lib/os_mon-2.9.1/priv/bin/memsup 999 79040 0.0 0.0 2652 956 ? Ss 02:14 0:00 /opt/erlang/lib/erlang/lib/os_mon-2.9.1/priv/bin/memsup 999 94534 0.0 0.0 2788 956 ? Ss 02:18 0:00 erl_child_setup 1048576 _chrony 10724 0.0 0.0 10580 544 ? S 01:50 0:00 /usr/sbin/chronyd -F 1 999 79111 0.0 0.0 3740 120 ? S 02:14 0:00 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/inet_gethost 4 999 94612 0.0 0.0 3740 120 ? S 02:18 0:00 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/inet_gethost 4 999 68575 0.0 0.0 3736 116 ? S 02:10 0:00 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/inet_gethost 4 999 94622 0.0 0.0 3748 112 ? S 02:18 0:00 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/epmd -daemon 999 68585 0.0 0.0 3744 108 ? S 02:10 0:00 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/epmd -daemon 999 79121 0.0 0.0 3748 108 ? S 02:14 0:00 /opt/erlang/lib/erlang/erts-14.2.5.1/bin/epmd -daemon root 12272 0.0 0.0 1048 4 ? Ss 01:51 0:00 /sbin/docker-init -- /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true root 12632 0.0 0.0 1048 4 ? Ss 01:51 0:00 /sbin/docker-init -- /usr/bin/ceph-mgr -n mgr.instance.qtdepf -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 19133 0.0 0.0 1048 4 ? Ss 01:53 0:00 /sbin/docker-init -- /usr/bin/ceph-crash -n client.crash.instance root 24265 0.0 0.0 1048 4 ? Ss 01:54 0:00 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 28801 0.0 0.0 1048 4 ? Ss 01:55 0:00 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 33595 0.0 0.0 1048 4 ? Ss 01:56 0:00 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug 65535 39029 0.0 0.0 1028 4 ? Ss 01:58 0:00 /pause 65535 39147 0.0 0.0 1028 4 ? Ss 01:58 0:00 /pause 65535 39152 0.0 0.0 1028 4 ? Ss 01:58 0:00 /pause 65535 39174 0.0 0.0 1028 4 ? Ss 01:58 0:00 /pause 65535 39187 0.0 0.0 1028 4 ? Ss 01:58 0:00 /pause 65535 39587 0.0 0.0 1028 4 ? Ss 01:58 0:00 /pause 65535 40413 0.0 0.0 1028 4 ? Ss 01:59 0:00 /pause 65535 40472 0.0 0.0 1028 4 ? Ss 01:59 0:00 /pause 65535 41743 0.0 0.0 1028 4 ? Ss 02:00 0:00 /pause 65535 42133 0.0 0.0 1028 4 ? Ss 02:00 0:00 /pause 65535 42177 0.0 0.0 1028 4 ? Ss 02:00 0:00 /pause 65535 42634 0.0 0.0 1028 4 ? Ss 02:00 0:00 /pause 65535 43158 0.0 0.0 1028 4 ? Ss 02:00 0:00 /pause 65535 43192 0.0 0.0 1028 4 ? Ss 02:00 0:00 /pause 65535 43244 0.0 0.0 1028 4 ? Ss 02:00 0:00 /pause 65535 45532 0.0 0.0 1028 4 ? Ss 02:01 0:00 /pause 65535 45619 0.0 0.0 1028 4 ? Ss 02:01 0:00 /pause systemd+ 45856 0.0 0.0 224 4 ? Ss 02:01 0:00 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true 65535 46561 0.0 0.0 1028 4 ? Ss 02:01 0:00 /pause 65535 46663 0.0 0.0 1028 4 ? Ss 02:01 0:00 /pause 65535 47049 0.0 0.0 1028 4 ? Ss 02:01 0:00 /pause 65535 47392 0.0 0.0 1028 4 ? Ss 02:02 0:00 /pause 65535 47716 0.0 0.0 1028 4 ? Ss 02:02 0:00 /pause 65535 50163 0.0 0.0 1028 4 ? Ss 02:04 0:00 /pause 65535 50340 0.0 0.0 1028 4 ? Ss 02:04 0:00 /pause 65535 55059 0.0 0.0 1028 4 ? Ss 02:07 0:00 /pause 65535 55659 0.0 0.0 1028 4 ? Ss 02:07 0:00 /pause 65535 55748 0.0 0.0 1028 4 ? Ss 02:07 0:00 /pause 65535 55797 0.0 0.0 1028 4 ? Ss 02:07 0:00 /pause 65535 55988 0.0 0.0 1028 4 ? Ss 02:07 0:00 /pause nobody 60778 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause nobody 60996 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause 472 61043 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause nobody 61087 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause zuul 62363 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause zuul 62716 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause systemd+ 62939 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause 65535 62988 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause 65535 63069 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause 65535 64050 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause zuul 64150 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause nobody 64868 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause 10001 64997 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause nobody 65755 0.0 0.0 1028 4 ? Ss 02:09 0:00 /pause 999 67034 0.0 0.0 1028 4 ? Ss 02:10 0:00 /pause 42424 69996 0.0 0.0 1028 4 ? Ss 02:11 0:00 /pause 999 78792 0.0 0.0 1028 4 ? Ss 02:14 0:00 /pause 42424 80131 0.0 0.0 1028 4 ? Ss 02:14 0:00 /pause 65535 86400 0.0 0.0 1028 4 ? Ss 02:16 0:00 /pause 999 94381 0.0 0.0 1028 4 ? Ss 02:18 0:00 /pause 65535 97814 0.0 0.0 1028 4 ? Ss 02:19 0:00 /pause 65535 97857 0.0 0.0 1028 4 ? Ss 02:19 0:00 /pause 42424 106740 0.0 0.0 1028 4 ? Ss 02:21 0:00 /pause 42424 108246 0.0 0.0 1028 4 ? Ss 02:21 0:00 /pause 42424 108389 0.0 0.0 1028 4 ? Ss 02:21 0:00 /pause 65535 109003 0.0 0.0 1028 4 ? Ss 02:22 0:00 /pause root 2 0.0 0.0 0 0 ? S 01:45 0:00 [kthreadd] root 3 0.0 0.0 0 0 ? I< 01:45 0:00 [rcu_gp] root 4 0.0 0.0 0 0 ? I< 01:45 0:00 [rcu_par_gp] root 5 0.0 0.0 0 0 ? I< 01:45 0:00 [slub_flushwq] root 6 0.0 0.0 0 0 ? I< 01:45 0:00 [netns] root 8 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/0:0H-events_highpri] root 9 0.1 0.0 0 0 ? I 01:45 0:02 [kworker/u32:0-flush-252:0] root 10 0.0 0.0 0 0 ? I< 01:45 0:00 [mm_percpu_wq] root 11 0.0 0.0 0 0 ? S 01:45 0:00 [rcu_tasks_rude_] root 12 0.0 0.0 0 0 ? S 01:45 0:00 [rcu_tasks_trace] root 13 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/0] root 14 0.1 0.0 0 0 ? I 01:45 0:04 [rcu_sched] root 15 0.0 0.0 0 0 ? S 01:45 0:00 [migration/0] root 16 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/0] root 18 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/0] root 19 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/1] root 20 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/1] root 21 0.0 0.0 0 0 ? S 01:45 0:00 [migration/1] root 22 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/1] root 24 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/1:0H-events_highpri] root 25 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/2] root 26 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/2] root 27 0.0 0.0 0 0 ? S 01:45 0:00 [migration/2] root 28 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/2] root 30 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/2:0H-kblockd] root 31 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/3] root 32 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/3] root 33 0.0 0.0 0 0 ? S 01:45 0:00 [migration/3] root 34 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/3] root 36 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/3:0H-events_highpri] root 37 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/4] root 38 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/4] root 39 0.0 0.0 0 0 ? S 01:45 0:00 [migration/4] root 40 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/4] root 42 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/4:0H-kblockd] root 43 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/5] root 44 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/5] root 45 0.0 0.0 0 0 ? S 01:45 0:00 [migration/5] root 46 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/5] root 48 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/5:0H-events_highpri] root 49 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/6] root 50 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/6] root 51 0.0 0.0 0 0 ? S 01:45 0:00 [migration/6] root 52 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/6] root 54 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/6:0H-events_highpri] root 55 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/7] root 56 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/7] root 57 0.0 0.0 0 0 ? S 01:45 0:00 [migration/7] root 58 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/7] root 60 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/7:0H-events_highpri] root 61 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/8] root 62 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/8] root 63 0.0 0.0 0 0 ? S 01:45 0:00 [migration/8] root 64 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/8] root 66 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/8:0H-kblockd] root 67 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/9] root 68 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/9] root 69 0.0 0.0 0 0 ? S 01:45 0:00 [migration/9] root 70 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/9] root 72 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/9:0H-events_highpri] root 73 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/10] root 74 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/10] root 75 0.0 0.0 0 0 ? S 01:45 0:00 [migration/10] root 76 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/10] root 78 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/10:0H-events_highpri] root 79 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/11] root 80 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/11] root 81 0.0 0.0 0 0 ? S 01:45 0:00 [migration/11] root 82 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/11] root 84 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/11:0H-events_highpri] root 85 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/12] root 86 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/12] root 87 0.0 0.0 0 0 ? S 01:45 0:00 [migration/12] root 88 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/12] root 90 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/12:0H-kblockd] root 91 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/13] root 92 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/13] root 93 0.0 0.0 0 0 ? S 01:45 0:00 [migration/13] root 94 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/13] root 96 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/13:0H-events_highpri] root 97 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/14] root 98 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/14] root 99 0.0 0.0 0 0 ? S 01:45 0:00 [migration/14] root 100 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/14] root 102 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/14:0H-kblockd] root 103 0.0 0.0 0 0 ? S 01:45 0:00 [cpuhp/15] root 104 0.0 0.0 0 0 ? S 01:45 0:00 [idle_inject/15] root 105 0.0 0.0 0 0 ? S 01:45 0:00 [migration/15] root 106 0.0 0.0 0 0 ? S 01:45 0:00 [ksoftirqd/15] root 108 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/15:0H-events_highpri] root 109 0.0 0.0 0 0 ? S 01:45 0:00 [kdevtmpfs] root 110 0.0 0.0 0 0 ? I< 01:45 0:00 [inet_frag_wq] root 111 0.0 0.0 0 0 ? S 01:45 0:00 [kauditd] root 112 0.0 0.0 0 0 ? S 01:45 0:00 [khungtaskd] root 113 0.0 0.0 0 0 ? S 01:45 0:00 [oom_reaper] root 114 0.0 0.0 0 0 ? I< 01:45 0:00 [writeback] root 115 0.0 0.0 0 0 ? S 01:45 0:00 [kcompactd0] root 116 0.0 0.0 0 0 ? SN 01:45 0:00 [ksmd] root 117 0.0 0.0 0 0 ? SN 01:45 0:00 [khugepaged] root 167 0.0 0.0 0 0 ? I< 01:45 0:00 [kintegrityd] root 168 0.0 0.0 0 0 ? I< 01:45 0:00 [kblockd] root 169 0.0 0.0 0 0 ? I< 01:45 0:00 [blkcg_punt_bio] root 170 0.0 0.0 0 0 ? I< 01:45 0:00 [tpm_dev_wq] root 171 0.0 0.0 0 0 ? I< 01:45 0:00 [ata_sff] root 172 0.0 0.0 0 0 ? I< 01:45 0:00 [md] root 173 0.0 0.0 0 0 ? I< 01:45 0:00 [edac-poller] root 174 0.0 0.0 0 0 ? I< 01:45 0:00 [devfreq_wq] root 175 0.0 0.0 0 0 ? S 01:45 0:00 [watchdogd] root 177 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/0:1H-kblockd] root 179 0.0 0.0 0 0 ? S 01:45 0:00 [kswapd0] root 180 0.0 0.0 0 0 ? S 01:45 0:00 [ecryptfs-kthrea] root 183 0.0 0.0 0 0 ? I< 01:45 0:00 [kthrotld] root 185 0.0 0.0 0 0 ? I< 01:45 0:00 [acpi_thermal_pm] root 186 0.1 0.0 0 0 ? I 01:45 0:03 [kworker/u32:2-events_unbound] root 187 0.0 0.0 0 0 ? S 01:45 0:00 [scsi_eh_0] root 188 0.0 0.0 0 0 ? I< 01:45 0:00 [scsi_tmf_0] root 189 0.0 0.0 0 0 ? S 01:45 0:00 [scsi_eh_1] root 190 0.0 0.0 0 0 ? I< 01:45 0:00 [scsi_tmf_1] root 192 0.0 0.0 0 0 ? I< 01:45 0:00 [vfio-irqfd-clea] root 193 0.0 0.0 0 0 ? I 01:45 0:00 [kworker/2:1-mm_percpu_wq] root 194 0.0 0.0 0 0 ? I< 01:45 0:00 [mld] root 195 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/3:1H-kblockd] root 196 0.0 0.0 0 0 ? I< 01:45 0:00 [ipv6_addrconf] root 199 0.0 0.0 0 0 ? I 01:45 0:02 [kworker/u32:5-ext4-rsv-conversion] root 202 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/2:1H-kblockd] root 217 0.0 0.0 0 0 ? I< 01:45 0:00 [kstrp] root 220 0.0 0.0 0 0 ? I< 01:45 0:00 [zswap-shrink] root 221 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/u33:0] root 226 0.0 0.0 0 0 ? I< 01:45 0:00 [charger_manager] root 268 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/15:1H-kblockd] root 283 0.0 0.0 0 0 ? S 01:45 0:00 [hwrng] root 288 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/11:1H-kblockd] root 291 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/6:1H-kblockd] root 292 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/13:1H-kblockd] root 293 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/7:1H-kblockd] root 294 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/10:1H-kblockd] root 313 0.2 0.0 0 0 ? S 01:45 0:06 [jbd2/vda1-8] root 314 0.0 0.0 0 0 ? I< 01:45 0:00 [ext4-rsv-conver] root 318 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/9:1H-kblockd] root 352 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/8:1H-kblockd] root 355 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/4:1H-kblockd] root 379 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/14:1H-kblockd] root 398 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/12:1H-kblockd] root 449 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/1:1H-kblockd] root 454 0.0 0.0 0 0 ? I< 01:45 0:00 [kworker/5:1H-kblockd] root 458 0.0 0.0 0 0 ? I< 01:45 0:00 [cryptd] root 2102 0.0 0.0 0 0 ? I 01:48 0:00 [kworker/15:2-events] root 9062 0.0 0.0 0 0 ? I< 01:49 0:00 [kdmflush] root 9079 0.0 0.0 0 0 ? I< 01:49 0:00 [kdmflush] root 9095 0.0 0.0 0 0 ? I< 01:49 0:00 [kdmflush] root 10237 0.0 0.0 0 0 ? I 01:50 0:00 [kworker/8:53-events] root 12128 0.1 0.0 0 0 ? I 01:51 0:02 [kworker/u32:4-flush-252:0] root 14621 0.0 0.0 0 0 ? I 01:52 0:01 [kworker/u32:6-flush-253:1] root 34388 0.0 0.0 0 0 ? I 01:56 0:01 [kworker/u32:7-flush-252:0] root 36605 0.0 0.0 0 0 ? I< 01:56 0:00 [ceph-msgr] root 36606 0.0 0.0 0 0 ? I< 01:56 0:00 [rbd] root 42693 0.0 0.0 0 0 ? I 02:00 0:00 [kworker/7:1-events] root 48143 0.0 0.0 0 0 ? I< 02:02 0:00 [dio/vda1] root 50038 0.0 0.0 0 0 ? I 02:04 0:00 [kworker/14:4-events] root 57063 0.0 0.0 0 0 ? I 02:07 0:00 [kworker/5:0-inode_switch_wbs] root 60642 0.0 0.0 0 0 ? I 02:09 0:00 [kworker/3:3-events] root 61160 0.0 0.0 0 0 ? I 02:09 0:00 [kworker/7:0-events] root 61870 0.0 0.0 0 0 ? I 02:09 0:00 [kworker/12:1-events] root 62106 0.0 0.0 0 0 ? I 02:09 0:00 [kworker/13:2-events] root 63697 0.0 0.0 0 0 ? I 02:09 0:00 [kworker/u32:9-ext4-rsv-conversion] root 64659 0.0 0.0 0 0 ? I 02:09 0:00 [kworker/9:1-events] root 65847 0.0 0.0 0 0 ? I 02:09 0:00 [kworker/u32:11-ext4-rsv-conversion] root 66652 0.1 0.0 0 0 ? I 02:10 0:01 [kworker/u32:12-flush-252:0] root 71564 0.0 0.0 0 0 ? I 02:11 0:00 [kworker/6:3-events] root 73957 0.0 0.0 0 0 ? I 02:12 0:00 [kworker/11:4-cgroup_free] root 80459 0.0 0.0 0 0 ? I 02:14 0:00 [kworker/12:2-events] root 81332 0.0 0.0 0 0 ? I 02:14 0:00 [kworker/1:6-events] root 81333 0.0 0.0 0 0 ? I 02:14 0:00 [kworker/1:7-events] root 82307 0.0 0.0 0 0 ? I 02:15 0:00 [kworker/5:1-inode_switch_wbs] root 83874 0.0 0.0 0 0 ? I 02:15 0:00 [kworker/u32:3-flush-252:0] root 83984 0.0 0.0 0 0 ? I 02:15 0:00 [kworker/0:63-events] root 85048 0.0 0.0 0 0 ? I 02:16 0:00 [kworker/3:0-events] root 85196 0.0 0.0 0 0 ? I 02:16 0:00 [kworker/10:1-events] root 85199 0.0 0.0 0 0 ? I 02:16 0:00 [kworker/14:2-cgroup_free] root 85816 0.0 0.0 0 0 ? I 02:16 0:00 [kworker/u32:13-flush-252:0] root 91277 0.0 0.0 0 0 ? I 02:18 0:00 [kworker/4:2-mm_percpu_wq] root 91395 0.0 0.0 0 0 ? I 02:18 0:00 [kworker/4:3-events] root 94241 0.0 0.0 0 0 ? I 02:18 0:00 [kworker/11:3-events] root 100883 0.0 0.0 0 0 ? I 02:19 0:00 [kworker/6:0-cgroup_free] root 101137 0.0 0.0 0 0 ? I 02:19 0:00 [kworker/15:0-events] root 102145 0.0 0.0 0 0 ? I 02:20 0:00 [kworker/13:0-events] root 102169 0.0 0.0 0 0 ? I 02:20 0:00 [kworker/2:3-events] root 104320 0.0 0.0 0 0 ? I 02:20 0:00 [kworker/9:3-events] root 104861 0.0 0.0 0 0 ? I 02:21 0:00 [kworker/5:2-inode_switch_wbs] root 106020 0.0 0.0 0 0 ? I 02:21 0:00 [kworker/8:2-mm_percpu_wq] root 106842 0.0 0.0 0 0 ? I 02:21 0:00 [kworker/10:2-events] root 108820 0.0 0.0 0 0 ? I 02:22 0:00 [kworker/0:2-events] root 109321 0.0 0.0 0 0 ? I 02:22 0:00 [kworker/3:1] root 109528 0.0 0.0 0 0 ? I 02:22 0:00 [kworker/5:3-inode_switch_wbs] root 109529 0.0 0.0 0 0 ? I 02:22 0:00 [kworker/5:4-inode_switch_wbs] root 109530 0.0 0.0 0 0 ? I 02:22 0:00 [kworker/5:5-inode_switch_wbs] root 109531 0.0 0.0 0 0 ? I 02:22 0:00 [kworker/5:6-inode_switch_wbs] root 109532 0.0 0.0 0 0 ? I 02:22 0:00 [kworker/5:7-events] root 109533 0.0 0.0 0 0 ? I 02:22 0:00 [kworker/5:8] root 117778 0.0 0.0 0 0 ? I 02:25 0:00 [kworker/9:0-mm_percpu_wq]