USER PID %CPU %MEM VSZ RSS TTY STAT START TIME COMMAND ubuntu 48428 16.6 3.1 13983476 2072680 ? Ssl 13:30 14:04 mysqld --wsrep_start_position=7b6eaafa-4306-11f1-a77c-df6819f9cfdd:20 root 39921 18.7 2.8 3106920 1853940 ? Ssl 13:28 16:19 kube-apiserver --advertise-address=199.204.45.109 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/12 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-min-version=VersionTLS13 --tls-private-key-file=/etc/kubernetes/pki/apiserver.key zuul 67303 4.5 1.4 2389484 946268 ? Ssl 13:38 3:30 /bin/prometheus --web.console.templates=/etc/prometheus/consoles --web.console.libraries=/etc/prometheus/console_libraries --config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.external-url=http://prometheus.199-204-45-109.nip.io/ --web.route-prefix=/ --storage.tsdb.retention.time=10d --storage.tsdb.path=/prometheus --storage.tsdb.wal-compression --web.config.file=/etc/prometheus/web_config/web-config.yaml 167 12610 2.8 0.8 4232332 578012 ? Sl 13:22 2:36 /usr/bin/ceph-mgr -n mgr.instance.csrfqj -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug 42424 159431 0.5 0.7 1942548 501240 ? SLl 13:59 0:19 /usr/sbin/ovs-vswitchd unix:/run/openvswitch/db.sock -vconsole:emer -vconsole:err -vconsole:info --pidfile=/run/openvswitch/ovs-vswitchd.pid --user=openvswitch:openvswitch --mlockall zuul 50827 1.7 0.7 3838548 473800 ? Ssl 13:32 1:29 java -Dkc.config.built=true -XX:MetaspaceSize=96M -XX:MaxMetaspaceSize=256m -Dfile.encoding=UTF-8 -Dsun.stdout.encoding=UTF-8 -Dsun.err.encoding=UTF-8 -Dstdout.encoding=UTF-8 -Dstderr.encoding=UTF-8 -XX:+ExitOnOutOfMemoryError -Djava.security.egd=file:/dev/urandom -XX:+UseParallelGC -XX:GCTimeRatio=4 -XX:AdaptiveSizePolicyWeight=90 -XX:FlightRecorderOptions=stackdepth=512 -XX:MinHeapFreeRatio=10 -XX:MaxHeapFreeRatio=20 -XX:MaxRAMPercentage=70 -XX:MinRAMPercentage=70 -XX:InitialRAMPercentage=50 --add-opens=java.base/java.util=ALL-UNNAMED --add-opens=java.base/java.util.concurrent=ALL-UNNAMED --add-opens=java.base/java.security=ALL-UNNAMED -Djgroups.dns.query=keycloak-headless.auth-system.svc.cluster.local -Dkc.home.dir=/opt/keycloak/bin/.. -Djboss.server.config.dir=/opt/keycloak/bin/../conf -Djava.util.logging.manager=org.jboss.logmanager.LogManager -Dpicocli.disable.closures=true -Dquarkus-log-max-startup-records=10000 -cp /opt/keycloak/bin/../lib/quarkus-run.jar io.quarkus.bootstrap.runner.QuarkusEntryPoint --verbose start --health-enabled=true --http-enabled=true --http-port=8080 --hostname-strict=false --spi-events-listener-jboss-logging-success-level=info --spi-events-listener-jboss-logging-error-level=warn --transaction-xa-enabled=false --metrics-enabled=true 42424 73534 0.0 0.5 607660 348796 ? Ss 13:39 0:00 apache2 -DFOREGROUND root 9797 19.6 0.4 4848332 318188 ? Ssl 13:21 18:24 /usr/bin/containerd --config /etc/containerd/config.toml 42424 193705 8.5 0.4 7405128 315040 ? Sl 14:07 4:01 neutron-api: uWSGI worker 1 42424 193706 8.7 0.4 7393996 304348 ? Sl 14:07 4:08 neutron-api: uWSGI worker 2 167 24771 0.8 0.4 827820 293136 ? Sl 13:25 0:46 /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug 42424 237071 1.0 0.4 1279548 285924 ? Sl 14:18 0:23 octavia-worker: ConsumerService worker(1) 167 34267 0.8 0.4 777568 277200 ? Sl 13:26 0:44 /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug 167 29500 0.8 0.3 768312 262620 ? Sl 13:26 0:47 /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 40093 16.6 0.3 3683260 249760 ? Ssl 13:28 14:27 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --cgroups-per-qos=false --container-runtime-endpoint=/run/containerd/containerd.sock --enforce-node-allocatable= --node-ip=199.204.45.109 --pod-infra-container-image=harbor.atmosphere.dev/registry.k8s.io/pause:3.9 10001 67938 1.2 0.3 1637796 227508 ? Ssl 13:38 0:59 /usr/bin/loki -config.file=/etc/loki/config/config.yaml -target=all 42424 202764 1.8 0.3 1844764 223600 ? Ssl 14:09 0:49 /var/lib/openstack/bin/python /var/lib/openstack/bin/nova-compute --config-file /etc/nova/nova.conf --config-file /tmp/pod-shared/nova-console.conf --config-file /tmp/pod-shared/nova-libvirt.conf --config-file /tmp/pod-shared/nova-compute-fqdn.conf --config-file /tmp/pod-shared/nova-hypervisor.conf 42424 237068 0.8 0.3 1162788 220984 ? Sl 14:18 0:18 octavia-worker: ConsumerService worker(0) 42424 178423 0.7 0.3 799504 197948 ? Sl 14:04 0:21 nova-api: uWSGI worker 1 42424 130473 0.3 0.2 3686440 197236 ? Sl 13:56 0:11 /var/lib/openstack/bin/python /var/lib/openstack/bin/cinder-volume --config-file /etc/cinder/cinder.conf --config-file /etc/cinder/conf/backends.conf --config-file /tmp/pod-shared/internal_tenant.conf root 39909 7.1 0.2 11426464 194948 ? Ssl 13:28 6:10 etcd --advertise-client-urls=https://199.204.45.109:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://199.204.45.109:2380 --initial-cluster=instance=https://199.204.45.109:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://199.204.45.109:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://199.204.45.109:2380 --name=instance --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --tls-min-version=TLS1.3 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt 42424 178424 0.6 0.2 795336 192700 ? Sl 14:04 0:21 nova-api: uWSGI worker 2 42424 193656 0.2 0.2 416652 182544 ? Ssl 14:07 0:07 neutron-rpc-server: master process [/var/lib/openstack/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /tmp/pod-shared/ovn.ini --config-file /etc/neutron/taas_plugin.ini --config-file /etc/neutron/plugins/ml2/ml2_conf.ini] 42424 127214 0.3 0.2 434432 181364 ? Sl 13:55 0:12 cinder-api: uWSGI worker 2 42424 127213 0.3 0.2 433276 180564 ? Sl 13:55 0:12 cinder-api: uWSGI worker 1 root 265475 2.3 0.2 1413360 179992 ? Ssl 14:25 0:42 kube-controller-manager --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=0.0.0.0 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-name=kubernetes --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key --controllers=*,bootstrapsigner,tokencleaner --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --tls-min-version=VersionTLS13 --use-service-account-credentials=true 42424 194057 1.3 0.2 1907024 179920 ? Sl 14:08 0:39 neutron-server: rpc worker (/var/lib/openstack/bin/python /var/lib/openstack/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /tmp/pod-shared/ovn.ini --config-file /etc/neutron/taas_plugin.ini --config-file /etc/neutron/plugins/ml2/ml2_conf.ini) 472 65263 0.5 0.2 1721244 178828 ? Ssl 13:37 0:25 grafana server --homepath=/usr/share/grafana --config=/etc/grafana/grafana.ini --packaging=docker cfg:default.log.mode=console cfg:default.paths.data=/var/lib/grafana/ cfg:default.paths.logs=/var/log/grafana cfg:default.paths.plugins=/var/lib/grafana/plugins cfg:default.paths.provisioning=/etc/grafana/provisioning 42424 194059 1.3 0.2 1831228 176128 ? Sl 14:08 0:38 neutron-server: rpc worker (/var/lib/openstack/bin/python /var/lib/openstack/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /tmp/pod-shared/ovn.ini --config-file /etc/neutron/taas_plugin.ini --config-file /etc/neutron/plugins/ml2/ml2_conf.ini) root 41491 2.7 0.2 1398904 175020 ? Ssl 13:28 2:20 cilium-agent --config-dir=/tmp/cilium/config-map 42424 111134 0.7 0.2 1217388 171336 ? S 13:50 0:28 glance-api: uWSGI worker 1 42424 194062 0.0 0.2 416684 166708 ? Sl 14:08 0:00 neutron-server: rpc reports worker (/var/lib/openstack/bin/python /var/lib/openstack/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /tmp/pod-shared/ovn.ini --config-file /etc/neutron/taas_plugin.ini --config-file /etc/neutron/plugins/ml2/ml2_conf.ini) 42424 195095 0.3 0.2 1210516 165788 ? Ssl 14:08 0:08 neutron-ovn-metadata-agent (/var/lib/openstack/bin/python /var/lib/openstack/bin/neutron-ovn-metadata-agent --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/ovn_metadata_agent.ini --config-file /tmp/pod-shared/neutron-agent.ini --config-file /tmp/pod-shared/ovn.ini) 42424 193737 0.2 0.2 180440 164912 ? S 14:07 0:07 neutron-policy-server: uWSGI worker 1 42424 193738 0.2 0.2 181464 164912 ? S 14:07 0:08 neutron-policy-server: uWSGI worker 2 42424 130191 1.1 0.2 193964 164516 ? Ss 13:56 0:40 /var/lib/openstack/bin/python /var/lib/openstack/bin/cinder-volume --config-file /etc/cinder/cinder.conf --config-file /etc/cinder/conf/backends.conf --config-file /tmp/pod-shared/internal_tenant.conf 42424 128807 0.2 0.2 193832 164184 ? Ss 13:56 0:09 /var/lib/openstack/bin/python /var/lib/openstack/bin/cinder-scheduler --config-file /etc/cinder/cinder.conf 42424 111135 0.1 0.2 265100 161076 ? S 13:50 0:07 glance-api: uWSGI worker 2 42424 195076 0.4 0.2 173072 159732 ? Ss 14:08 0:12 /var/lib/openstack/bin/python /var/lib/openstack/bin/neutron-ovn-vpn-agent --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/neutron_vpnaas.conf --config-file /etc/neutron/neutron_ovn_vpn_agent.ini --config-file /tmp/pod-shared/neutron-agent.ini --config-file /tmp/pod-shared/ovn.ini root 129824 0.2 0.2 5425124 157016 ? Ssl 13:56 0:07 /var/lib/openstack/bin/python /var/lib/openstack/bin/cinder-backup --config-file /etc/cinder/cinder.conf 999 163976 1.1 0.2 4291896 152284 ? Sl 14:00 0:38 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- 42424 73576 2.4 0.2 869436 152276 ? Sl 13:39 1:49 (wsgi:k -DFOREGROUND 42424 73578 2.6 0.2 869436 151728 ? Sl 13:39 1:57 (wsgi:k -DFOREGROUND root 66898 1.6 0.2 645596 150400 ? Ssl 13:38 1:18 /usr/bin/vector --config-dir /etc/vector/ 42424 211523 0.2 0.2 238244 149848 ? Ssl 14:11 0:05 heat-engine: master process [/var/lib/openstack/bin/heat-engine --config-file /etc/heat/heat.conf] 42424 112735 0.0 0.2 389096 149692 ? Sl 13:51 0:01 staffeln-conductor: Staffeln conductor backup controller worker(0) 42424 73579 2.6 0.2 868412 149608 ? Sl 13:39 2:00 (wsgi:k -DFOREGROUND 42424 73577 2.8 0.2 867388 148684 ? Sl 13:39 2:07 (wsgi:k -DFOREGROUND 42424 235446 0.4 0.2 238820 148520 ? Sl 14:17 0:10 octavia-api: uWSGI worker 1 42424 112737 0.0 0.2 387048 148344 ? Sl 13:51 0:01 staffeln-conductor: Staffeln conductor rotation controller worker(0) 42424 235447 0.4 0.2 238056 147784 ? Sl 14:17 0:09 octavia-api: uWSGI worker 2 root 195791 0.1 0.2 213420 146752 ? Sl 14:08 0:04 /var/lib/openstack/bin/python /var/lib/openstack/bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/ovn_metadata_agent.ini --config-file /tmp/pod-shared/neutron-agent.ini --config-file /tmp/pod-shared/ovn.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmp4wn8ur1r/privsep.sock root 196010 0.1 0.2 212424 146700 ? Sl 14:08 0:03 /var/lib/openstack/bin/python /var/lib/openstack/bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/neutron_vpnaas.conf --config-file /etc/neutron/neutron_ovn_vpn_agent.ini --config-file /tmp/pod-shared/neutron-agent.ini --config-file /tmp/pod-shared/ovn.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmpvv0k7mwz/privsep.sock 42424 178615 0.2 0.2 306636 143720 ? Sl 14:04 0:07 nova-metadata: uWSGI worker 2 42424 178614 0.2 0.2 304500 141612 ? Sl 14:04 0:06 nova-metadata: uWSGI worker 1 42424 173587 0.2 0.2 746456 141316 ? Ss 14:02 0:07 /var/lib/openstack/bin/python /var/lib/openstack/bin/nova-novncproxy --config-file /etc/nova/nova.conf --config-file /tmp/pod-shared/nova-vnc.ini 42424 211835 0.0 0.2 690960 140148 ? Sl 14:11 0:01 heat-engine: ServiceWrapper worker(0) 42424 211838 0.0 0.2 690960 139888 ? Sl 14:11 0:01 heat-engine: ServiceWrapper worker(1) 42424 179577 0.4 0.2 164804 139628 ? S 14:04 0:13 /var/lib/openstack/bin/python /var/lib/openstack/bin/nova-conductor --config-file /etc/nova/nova.conf 999 267433 0.8 0.2 4285136 139588 ? Sl 14:25 0:15 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- 42424 179576 0.4 0.2 163860 139380 ? S 14:04 0:13 /var/lib/openstack/bin/python /var/lib/openstack/bin/nova-conductor --config-file /etc/nova/nova.conf 42424 179183 1.1 0.2 150900 138416 ? Ss 14:04 0:35 /var/lib/openstack/bin/python /var/lib/openstack/bin/nova-conductor --config-file /etc/nova/nova.conf 42424 303284 0.2 0.2 163180 138308 ? S 14:28 0:04 /var/lib/openstack/bin/python /var/lib/openstack/bin/manila-share --config-file /etc/manila/manila.conf --config-file /tmp/pod-shared/manila-share-fqdn.conf 42424 179228 0.1 0.2 370528 137808 ? Ssl 14:04 0:04 nova-scheduler: master process [/var/lib/openstack/bin/nova-scheduler --config-file /etc/nova/nova.conf] 42424 310138 0.4 0.2 870316 136604 ? Sl 14:30 0:06 (wsgi:h -DFOREGROUND 999 114386 0.7 0.2 4280808 136416 ? Sl 13:51 0:28 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- 999 180519 0.8 0.2 4285792 135616 ? Sl 14:04 0:26 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- 42424 310140 0.4 0.2 871340 135104 ? Sl 14:30 0:06 (wsgi:h -DFOREGROUND 999 196428 0.8 0.2 4275468 134736 ? Sl 14:08 0:22 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- 42424 303097 0.3 0.2 222816 133756 ? Sl 14:28 0:05 manila-api: uWSGI worker 2 42424 303094 0.3 0.2 222816 133476 ? Sl 14:28 0:05 manila-api: uWSGI worker 1 999 68856 0.6 0.2 4275900 132612 ? Sl 13:38 0:28 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- root 188173 0.1 0.2 146984 132104 ? S 14:06 0:05 neutron-ovn-network-logging-parser: uWSGI worker 1 999 79538 0.6 0.2 4274108 131848 ? Sl 13:42 0:28 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- 999 93971 0.6 0.1 4268444 131348 ? Sl 13:46 0:26 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- 42424 112561 0.0 0.1 216372 130408 ? Ssl 13:51 0:03 staffeln-conductor: master process [/var/lib/openstack/bin/staffeln-conductor --config-file /etc/staffeln/staffeln.conf] 42424 236499 0.3 0.1 289416 130380 ? Ssl 14:18 0:06 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-housekeeping --config-file /etc/octavia/octavia.conf 999 238494 0.7 0.1 4272080 129904 ? Sl 14:18 0:16 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- 42424 179660 0.3 0.1 5764408 129880 ? Sl 14:04 0:11 nova-scheduler: ServiceWrapper worker(1) 999 215673 0.7 0.1 4271984 129756 ? Sl 14:12 0:18 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -B i -- -root /opt/erlang/lib/erlang -bindir /opt/erlang/lib/erlang/erts-15.2.7.2/bin -progname erl -- -home /var/lib/rabbitmq -- -pa -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger [] -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -- 42424 310137 0.2 0.1 642976 129428 ? Sl 14:30 0:04 (wsgi:h -DFOREGROUND 42424 310136 0.3 0.1 642976 129332 ? Sl 14:30 0:04 (wsgi:h -DFOREGROUND 42424 310139 0.3 0.1 644000 128996 ? Sl 14:30 0:04 (wsgi:h -DFOREGROUND 42424 179658 0.3 0.1 5764388 128920 ? Sl 14:04 0:11 nova-scheduler: ServiceWrapper worker(0) root 237394 0.2 0.1 139888 126828 ? Ss 14:18 0:06 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf 42424 236507 0.2 0.1 213572 126548 ? Ssl 14:18 0:06 octavia-worker: master process [/var/lib/openstack/bin/octavia-worker --config-file /etc/octavia/octavia.conf] 42424 303080 1.4 0.1 136828 124344 ? Ss 14:28 0:22 /var/lib/openstack/bin/python /var/lib/openstack/bin/manila-share --config-file /etc/manila/manila.conf --config-file /tmp/pod-shared/manila-share-fqdn.conf 167 98622 0.4 0.1 5243056 123512 ? Ssl 13:46 0:18 radosgw --fsid=4837cbf8-4f90-4300-b3f6-726c9b9f89b4 --keyring=/etc/ceph/keyring-store/keyring --default-log-to-stderr=true --default-err-to-stderr=true --default-mon-cluster-log-to-stderr=true --default-log-stderr-prefix=debug --default-log-to-file=false --default-mon-cluster-log-to-file=false --mon-host=[v2:10.96.240.200:3300,v1:10.96.240.200:6789] --mon-initial-members=instance --id=rgw.ceph.a --setuser=ceph --setgroup=ceph --foreground --rgw-frontends=beast port=8080 --host=rook-ceph-rgw-ceph-a-699b8bdb59-8l9pd --rgw-mime-types-file=/etc/ceph/rgw/mime.types --rgw-realm=ceph --rgw-zonegroup=ceph --rgw-zone=ceph 42424 302888 0.3 0.1 134432 121712 ? Ss 14:28 0:05 /var/lib/openstack/bin/python /var/lib/openstack/bin/manila-scheduler --config-file /etc/manila/manila.conf 42424 257231 0.3 0.1 284340 121576 ? Ssl 14:22 0:06 /var/lib/openstack/bin/python /var/lib/openstack/bin/magnum-cluster-api-proxy root 238124 0.2 0.1 143356 117664 ? S 14:18 0:04 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf 42424 257641 1.4 0.1 130032 117072 ? Ss 14:22 0:27 /var/lib/openstack/bin/python /var/lib/openstack/bin/magnum-conductor --config-file /etc/magnum/magnum.conf --config-file /tmp/pod-shared/magnum.conf root 217536 0.1 0.1 409816 116856 ? Sl 14:13 0:03 /var/lib/openstack/bin/python /var/lib/openstack/bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/ovn_metadata_agent.ini --config-file /tmp/pod-shared/neutron-agent.ini --config-file /tmp/pod-shared/ovn.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmpkyfk4kfg/privsep.sock root 411776 0.0 0.1 142016 116356 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 237367 0.0 0.1 188572 116332 ? Sl 14:18 0:02 /var/lib/openstack/bin/python /var/lib/openstack/bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/ovn_metadata_agent.ini --config-file /tmp/pod-shared/neutron-agent.ini --config-file /tmp/pod-shared/ovn.ini --privsep_context neutron.privileged.link_cmd --privsep_sock_path /tmp/tmpkf8rmfi4/privsep.sock root 411807 0.0 0.1 289512 116172 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf 42424 302886 0.3 0.1 129148 116156 ? Ss 14:28 0:05 /var/lib/openstack/bin/python /var/lib/openstack/bin/manila-data --config-file /etc/manila/manila.conf root 411809 0.0 0.1 289512 116152 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411808 0.0 0.1 289512 116148 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411801 0.0 0.1 289512 116144 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411778 0.0 0.1 142016 116140 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411781 0.0 0.1 142016 116140 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411804 0.0 0.1 289512 116140 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411806 0.0 0.1 289512 116140 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411773 0.0 0.1 142016 116136 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411775 0.0 0.1 142016 116136 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411779 0.0 0.1 142016 116136 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411802 0.0 0.1 289512 116136 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411803 0.0 0.1 289512 116136 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411777 0.0 0.1 142016 116132 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411782 0.0 0.1 142016 116132 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411785 0.0 0.1 142016 116132 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411788 0.0 0.1 142016 116132 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411805 0.0 0.1 289512 116132 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411786 0.0 0.1 142016 116128 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411774 0.0 0.1 142016 116124 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411787 0.0 0.1 142016 116124 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 399774 0.3 0.1 409816 115820 ? Sl 14:48 0:01 /var/lib/openstack/bin/python /var/lib/openstack/bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/neutron_vpnaas.conf --config-file /etc/neutron/neutron_ovn_vpn_agent.ini --config-file /tmp/pod-shared/neutron-agent.ini --config-file /tmp/pod-shared/ovn.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmpnbb2m1kx/privsep.sock root 399860 0.2 0.1 188572 115116 ? Sl 14:48 0:01 /var/lib/openstack/bin/python /var/lib/openstack/bin/privsep-helper --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/neutron_vpnaas.conf --config-file /etc/neutron/neutron_ovn_vpn_agent.ini --config-file /tmp/pod-shared/neutron-agent.ini --config-file /tmp/pod-shared/ovn.ini --privsep_context neutron.privileged.link_cmd --privsep_sock_path /tmp/tmp3z9w6y5c/privsep.sock 42424 257660 0.2 0.1 131052 115000 ? S 14:22 0:05 magnum-api: uWSGI worker 1 42424 257661 0.2 0.1 131052 115000 ? S 14:22 0:05 magnum-api: uWSGI worker 2 root 411794 0.0 0.1 288488 114456 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411800 0.0 0.1 288488 114280 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411810 0.0 0.1 288488 114280 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411799 0.0 0.1 288488 114268 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 238123 0.0 0.1 434920 114040 ? Sl 14:18 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf 65532 265331 0.2 0.1 1479900 113892 ? Ssl 14:25 0:03 /usr/local/bin/envoy-gateway server --config-path=/config/envoy-gateway.yaml 42424 86216 0.1 0.1 128816 113700 ? S 13:44 0:05 barbiacan-api: uWSGI worker 1 167 12241 0.9 0.1 331400 113284 ? Sl 13:22 0:51 /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true root 411811 0.0 0.1 287424 110624 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411812 0.0 0.1 287424 110624 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411813 0.0 0.1 287424 110624 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411789 0.0 0.1 139928 110504 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411790 0.0 0.1 139928 110504 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf root 411791 0.0 0.1 139928 110504 ? S 14:52 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/octavia-health-manager --config-file /etc/octavia/octavia.conf 42424 211417 0.1 0.1 119544 102036 ? S 14:11 0:04 heat-api: uWSGI worker 2 42424 211416 0.1 0.1 119544 102024 ? S 14:11 0:04 heat-api: uWSGI worker 1 42424 258014 0.0 0.1 127496 100932 ? S 14:22 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/magnum-conductor --config-file /etc/magnum/magnum.conf --config-file /tmp/pod-shared/magnum.conf 42424 258015 0.0 0.1 127496 100932 ? S 14:22 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/magnum-conductor --config-file /etc/magnum/magnum.conf --config-file /tmp/pod-shared/magnum.conf 42424 211567 0.1 0.1 118668 100776 ? S 14:11 0:04 heat-api-cfn: uWSGI worker 1 42424 211568 0.1 0.1 118668 100776 ? S 14:11 0:04 heat-api-cfn: uWSGI worker 2 zuul 43988 0.1 0.1 1337220 99004 ? Ssl 13:29 0:09 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=cert-manager root 149127 0.2 0.1 111024 96640 ? S 13:58 0:07 placement-api: uWSGI worker 1 root 149128 0.2 0.1 110936 96384 ? S 13:58 0:07 placement-api: uWSGI worker 2 42424 113967 0.1 0.1 239376 90684 ? Ss 13:51 0:04 /var/lib/openstack/bin/python /var/lib/openstack/bin/staffeln-api --config-file /etc/staffeln/staffeln.conf 2016 88108 0.1 0.1 5492616 90020 ? Ssl 13:44 0:07 /usr/local/bin/rook ceph operator bin 261657 1.1 0.1 1284636 86844 ? Ssl 14:24 0:21 percona-xtradb-cluster-operator root 387792 0.1 0.1 171212 81028 ? Sl 14:45 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/privsep-helper --config-file /etc/nova/nova.conf --config-file /tmp/pod-shared/nova-console.conf --config-file /tmp/pod-shared/nova-libvirt.conf --config-file /tmp/pod-shared/nova-compute-fqdn.conf --config-file /tmp/pod-shared/nova-hypervisor.conf --privsep_context vif_plug_ovs.privsep.vif_plug --privsep_sock_path /tmp/tmpj_d_ohox/privsep.sock 42424 420612 95.5 0.1 92924 81012 ? R 14:55 0:01 /var/lib/openstack/bin/python /var/lib/openstack/bin/heat-manage service clean zuul 43896 0.3 0.1 1292004 79888 ? Ssl 13:29 0:17 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=cert-manager --acme-http01-solver-image=harbor.atmosphere.dev/quay.io/jetstack/cert-manager-acmesolver:v1.12.17 --feature-gates=AdditionalCertificateOutputFormats=true --max-concurrent-challenges=60 root 265415 0.3 0.1 1305768 78180 ? Ssl 14:25 0:05 cilium-operator-generic --config-dir=/tmp/cilium/config-map --debug=false root 387259 0.0 0.1 168492 76192 ? Sl 14:45 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/privsep-helper --config-file /etc/nova/nova.conf --config-file /tmp/pod-shared/nova-console.conf --config-file /tmp/pod-shared/nova-libvirt.conf --config-file /tmp/pod-shared/nova-compute-fqdn.conf --config-file /tmp/pod-shared/nova-hypervisor.conf --privsep_context nova.privsep.sys_admin_pctxt --privsep_sock_path /tmp/tmpztjo6ueq/privsep.sock root 10171 1.0 0.1 2467348 75588 ? Ssl 13:21 1:01 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock root 257903 0.0 0.1 1296452 73712 ? Sl 14:22 0:00 haproxy -f /var/lib/magnum/.magnum-cluster-api-proxy/haproxy.cfg systemd+ 46397 0.3 0.1 1283008 72984 ? Ssl 13:29 0:16 /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true root 388781 0.0 0.1 241892 71936 ? Sl 14:45 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/privsep-helper --config-file /etc/nova/nova.conf --privsep_context os_brick.privileged.default --privsep_sock_path /tmp/tmp0zqgpiin/privsep.sock nobody 62985 2.5 0.1 1288280 70260 ? Ssl 13:37 1:58 /bin/operator --kubelet-service=kube-system/kube-prometheus-stack-kubelet --localhost=127.0.0.1 --prometheus-config-reloader=harbor.atmosphere.dev/quay.io/prometheus-operator/prometheus-config-reloader:v0.73.0 --config-reloader-cpu-request=0 --config-reloader-cpu-limit=0 --config-reloader-memory-request=0 --config-reloader-memory-limit=0 --thanos-default-base-image=quay.io/thanos/thanos:v0.35.1 --secret-field-selector=type!=kubernetes.io/dockercfg,type!=kubernetes.io/service-account-token,type!=helm.sh/release.v1 --web.enable-tls=true --web.cert-file=/cert/cert --web.key-file=/cert/key --web.listen-address=:10250 --web.tls-min-version=VersionTLS13 root 265539 0.4 0.1 1285908 67696 ? Ssl 14:25 0:08 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=0.0.0.0 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true --tls-min-version=VersionTLS13 65532 261747 0.5 0.1 1285460 66064 ? Ssl 14:24 0:09 /manager --metrics-bind-address=:8443 --leader-elect --health-probe-bind-address=:8081 472 63705 0.0 0.0 70024 64200 ? Ss 13:37 0:02 python -u /app/sidecar.py 472 63657 0.0 0.0 70024 64144 ? Ss 13:37 0:02 python -u /app/sidecar.py 472 63962 0.1 0.0 72668 63644 ? S 13:37 0:07 python -u /app/sidecar.py 65532 261729 0.4 0.0 1287632 63300 ? Ssl 14:24 0:07 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,ClusterResourceSet=true,ClusterTopology=true,RuntimeSDK=false,MachineSetPreflightChecks=true,MachineWaitForVolumeDetachConsiderVolumeAttachments=true,PriorityQueue=false 472 63971 0.0 0.0 70472 61508 ? S 13:37 0:00 python -u /app/sidecar.py 472 63970 0.0 0.0 70472 61152 ? S 13:37 0:00 python -u /app/sidecar.py 472 63961 0.0 0.0 70472 60944 ? S 13:37 0:00 python -u /app/sidecar.py 65532 243078 0.2 0.0 1287380 57888 ? Ssl 14:19 0:05 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,ClusterTopology=true,KubeadmBootstrapFormatIgnition=false,PriorityQueue=false root 40242 0.0 0.0 1284484 57428 ? Ssl 13:28 0:03 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=instance root 42862 0.5 0.0 767984 56688 ? Ssl 13:29 0:30 /coredns -conf /etc/coredns/Corefile 65532 265361 0.3 0.0 1283868 56524 ? Ssl 14:25 0:05 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,KubeadmBootstrapFormatIgnition=false,PriorityQueue=false --bootstrap-token-ttl=15m nobody 57615 0.1 0.0 1282120 56264 ? Ssl 13:35 0:05 nfd-master -crd-controller=true -metrics=8081 root 42819 0.6 0.0 767472 56220 ? Ssl 13:29 0:35 /coredns -conf /etc/coredns/Corefile 65532 265329 0.2 0.0 1284648 54204 ? Ssl 14:25 0:05 /manager --leader-elect --v=2 --diagnostics-address=127.0.0.1:8080 --insecure-diagnostics=true root 188018 0.0 0.0 271836 53676 ? Ssl 14:06 0:01 vector --config /etc/vector/vector.toml ubuntu 47034 0.1 0.0 1280876 53256 ? Ssl 13:30 0:09 /manager --metrics-bind-address=:9782 42424 420727 75.0 0.0 63156 50120 ? Rs 14:55 0:00 python /tmp/health-probe.py --config-file /etc/nova/nova.conf --service-queue-name scheduler 42424 420768 66.0 0.0 62000 49776 ? Rs 14:55 0:00 python /tmp/health-probe.py --config-file /etc/nova/nova.conf --service-queue-name conductor 42424 420769 66.0 0.0 62000 49708 ? Rs 14:55 0:00 python /tmp/health-probe.py --config-file /etc/nova/nova.conf --service-queue-name conductor --liveness-probe 42424 420725 75.0 0.0 62000 49520 ? Rs 14:55 0:00 python /tmp/health-probe.py --config-file /etc/nova/nova.conf --service-queue-name scheduler --liveness-probe nobody 63336 0.3 0.0 1271268 47188 ? Ssl 13:37 0:14 /kube-state-metrics --port=8080 --telemetry-port=8081 --port=8080 --resources=certificatesigningrequests,configmaps,cronjobs,daemonsets,deployments,endpoints,horizontalpodautoscalers,ingresses,jobs,leases,limitranges,mutatingwebhookconfigurations,namespaces,networkpolicies,nodes,persistentvolumeclaims,persistentvolumes,poddisruptionbudgets,pods,replicasets,replicationcontrollers,resourcequotas,secrets,services,statefulsets,storageclasses,validatingwebhookconfigurations,volumeattachments root 257872 0.0 0.0 145364 46896 ? Sl 14:22 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/privsep-helper --privsep_context magnum_cluster_api.privsep.haproxy_pctxt --privsep_sock_path /tmp/tmp_5bmej3z/privsep.sock zuul 43944 0.0 0.0 1282112 46064 ? Ssl 13:29 0:04 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --feature-gates=AdditionalCertificateOutputFormats=true --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc --feature-gates=AdditionalCertificateOutputFormats=true ubuntu 265485 0.2 0.0 1274208 45736 ? Ssl 14:25 0:04 /manager --metrics-bind-address=:8080 root 194906 0.0 0.0 2038788 44956 ? Ssl 14:08 0:00 /usr/bin/libvirt-tls-sidecar 42424 420812 0.0 0.0 56204 43944 ? Rs 14:55 0:00 python /tmp/health-probe.py --config-file /etc/nova/nova.conf --service-queue-name compute --use-fqdn zuul 58458 0.1 0.0 1267324 43892 ? Ssl 13:36 0:05 /secretgen-controller root 399901 0.0 0.0 350760 43392 ? Sl 14:48 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf 42424 420810 0.0 0.0 54708 42956 ? Rs 14:55 0:00 python /tmp/health-probe.py --config-file /etc/nova/nova.conf --service-queue-name compute --liveness-probe --use-fqdn root 388116 0.0 0.0 349460 42568 ? Sl 14:45 0:00 /var/lib/openstack/bin/python /var/lib/openstack/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf zuul 67010 0.0 0.0 1268836 42460 ? Ssl 13:38 0:02 /goldpinger --static-file-path /static nobody 57577 0.2 0.0 2592740 41992 ? Ssl 13:35 0:11 nfd-worker -metrics=8081 zuul 67354 0.0 0.0 1270412 40896 ? Ssl 13:38 0:00 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9090/-/reload --config-file=/etc/prometheus/config/prometheus.yaml.gz --config-envsubst-file=/etc/prometheus/config_out/prometheus.env.yaml --watched-dir=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0 zuul 67530 0.0 0.0 1256936 40440 ? Ssl 13:38 0:02 /bin/alertmanager --config.file=/etc/alertmanager/config_out/alertmanager.env.yaml --storage.path=/alertmanager --data.retention=120h --cluster.listen-address= --web.listen-address=:9093 --web.external-url=http://alertmanager.199-204-45-109.nip.io/ --web.route-prefix=/ --cluster.label=monitoring/kube-prometheus-stack-alertmanager --cluster.peer=alertmanager-kube-prometheus-stack-alertmanager-0.alertmanager-operated:9094 --cluster.reconnect-timeout=5m --web.config.file=/etc/alertmanager/web_config/web-config.yaml zuul 67651 0.0 0.0 1269900 39164 ? Ssl 13:38 0:00 /bin/prometheus-config-reloader --listen-address=:8080 --reload-url=http://127.0.0.1:9093/-/reload --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config root 265453 0.4 0.0 1272668 36832 ? Ssl 14:25 0:08 /kube-vip manager root 195792 0.5 0.0 1854364 35344 ? Sl 14:08 0:16 /usr/sbin/libvirtd --listen nobody 66917 0.2 0.0 439112 32224 ? Ssl 13:38 0:13 memcached -v -p 11211 -U 0 -c 8192 -m 1024 nobody 57656 0.0 0.0 1266792 31640 ? Ssl 13:35 0:00 nfd-gc -gc-interval=1h 42424 162529 0.0 0.0 128184 30376 ? Sl 14:00 0:02 ovsdb-server -vconsole:info -vfile:info --log-file=/var/log/ovn/ovsdb-server-sb.log --remote=punix:/var/run/ovn/ovnsb_db.sock --pidfile=/var/run/ovn/ovnsb_db.pid --unixctl=/var/run/ovn/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=db:OVN_Southbound,SSL,private_key --certificate=db:OVN_Southbound,SSL,certificate --ca-cert=db:OVN_Southbound,SSL,ca_cert --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers /etc/ovn/ovnsb_db.db root 43437 0.0 0.0 735700 29896 ? Ssl 13:29 0:03 local-path-provisioner --debug start --config /etc/config/config.json --service-account-name local-path-provisioner --provisioner-name cluster.local/local-path-provisioner --helper-image harbor.atmosphere.dev/ghcr.io/containerd/busybox:1.36 --configmap-name local-path-config 42424 162600 0.0 0.0 126160 28460 ? Sl 14:00 0:02 ovsdb-server -vconsole:info -vfile:info --log-file=/var/log/ovn/ovsdb-server-nb.log --remote=punix:/var/run/ovn/ovnnb_db.sock --pidfile=/var/run/ovn/ovnnb_db.pid --unixctl=/var/run/ovn/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=db:OVN_Northbound,SSL,private_key --certificate=db:OVN_Northbound,SSL,certificate --ca-cert=db:OVN_Northbound,SSL,ca_cert --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers /etc/ovn/ovnnb_db.db zuul 67389 0.0 0.0 1263300 27996 ? Ssl 13:38 0:00 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key nobody 63981 0.0 0.0 1263236 27884 ? Ssl 13:37 0:00 /usr/bin/pod-tls-sidecar --template=/config/certificate-template.yml --ca-path=/certs/ca.crt --cert-path=/certs/tls.crt --key-path=/certs/tls.key root 157631 0.0 0.0 223320 26620 ? SLsl 13:59 0:00 /sbin/multipathd -d -s 42424 159385 0.0 0.0 49400 26448 ? Ss 13:59 0:01 /usr/sbin/ovsdb-server /run/openvswitch/conf.db -vconsole:emer -vconsole:err -vconsole:info --pidfile=/run/openvswitch/ovsdb-server.pid --remote=punix:/run/openvswitch/db.sock --remote=db:Open_vSwitch,Open_vSwitch,manager_options --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert 11211 66406 0.0 0.0 432204 25236 ? Ssl 13:38 0:01 memcached -m 8192 --extended=modern,track_sizes -I 5m -c 16384 -v -u 11211 nobody 65995 0.7 0.0 1241700 23500 ? Ssl 13:38 0:33 /bin/node_exporter --path.procfs=/host/proc --path.sysfs=/host/sys --path.rootfs=/host/root --path.udev.data=/host/root/run/udev/data --web.listen-address=[0.0.0.0]:9100 --collector.diskstats.ignored-devices=^(ram|loop|nbd|fd|(h|s|v|xv)d[a-z]|nvme\\d+n\\d+p)\\d+$ --collector.filesystem.fs-types-exclude=^(autofs|binfmt_misc|bpf|cgroup2?|configfs|debugfs|devpts|devtmpfs|fusectl|fuse.squashfuse_ll|hugetlbfs|iso9660|mqueue|nsfs|overlay|proc|procfs|pstore|rpc_pipefs|securityfs|selinuxfs|squashfs|sysfs|tracefs)$ --collector.filesystem.mount-points-exclude=^/(dev|proc|run/credentials/.+|sys|var/lib/docker/.+|var/lib/kubelet/pods/.+|var/lib/kubelet/plugins/kubernetes.io/csi/.+|run/containerd/.+)($|/) --collector.netclass.ignored-devices=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.netdev.device-exclude=^(lxc|cilium_|qbr|qvb|qvo|tap|ovs-system|br|tbr|gre_sys|[0-9a-f]+_eth|vxlan).*$ --collector.processes --collector.systemd --collector.stat.softirq --web.config.file=/config/node-exporter.yml nobody 68125 0.0 0.0 1241956 23220 ? Ssl 13:38 0:01 /bin/pushgateway root 420828 0.0 0.0 106484 23076 ? Sl 14:55 0:00 /usr/bin/python3 systemd+ 307262 0.0 0.0 133208 22072 ? Sl 14:29 0:00 nginx: worker process systemd+ 307261 0.0 0.0 132952 21832 ? Sl 14:29 0:00 nginx: worker process root 24717 0.0 0.0 736288 21464 ? Sl 13:25 0:00 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-0 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0:/var/lib/ceph/osd/ceph-0:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 19526 0.0 0.0 736008 21288 ? Sl 13:24 0:00 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-crash --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-crash-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/config:/etc/ceph/ceph.conf:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/keyring:/etc/ceph/ceph.client.crash.instance.keyring quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n client.crash.instance root 12511 0.0 0.0 736404 21124 ? Sl 13:22 0:00 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mgr --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mgr-instance-csrfqj --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.csrfqj:/var/lib/ceph/mgr/ceph-instance.csrfqj:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.csrfqj/config:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v18.2.7 -n mgr.instance.csrfqj -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 34214 0.0 0.0 736404 20920 ? Sl 13:26 0:00 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-osd --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-osd-2 --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2:/var/lib/ceph/osd/ceph-2:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /:/rootfs quay.io/ceph/ceph@sha256:1b9158ce28975f95def6a0ad459fa19f1336506074267a4b47c1bd914a00fec0 -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 12147 0.0 0.0 736148 20908 ? Sl 13:22 0:01 /usr/bin/docker run --rm --ipc=host --stop-signal=SIGTERM --ulimit nofile=1048576 --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-4837cbf8-4f90-4300-b3f6-726c9b9f89b4-mon-instance --pids-limit=0 -e CONTAINER_IMAGE=quay.io/ceph/ceph:v18.2.7 -e NODE_NAME=instance -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/run/ceph:z -v /var/log/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4:/var/log/ceph:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash:/var/lib/ceph/crash:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance:/var/lib/ceph/mon/ceph-instance:z -v /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/config:/etc/ceph/ceph.conf:z -v /dev:/dev -v /run/udev:/run/udev quay.io/ceph/ceph:v18.2.7 -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true root 1182 0.0 0.0 296140 20492 ? Ssl 13:19 0:01 /usr/libexec/packagekitd root 356 0.3 0.0 74760 20468 ? S /tmp/logs/system/systemd-cgls.txt ip addr > /tmp/logs/system/ip-addr.txt ip route > /tmp/logs/system/ip-route.txt lsblk > /tmp/logs/system/lsblk.txt mount > /tmp/logs/system/mount.txt docker images > /tmp/logs/system/docker-images.txt brctl show > /tmp/logs/system/brctl-show.txt ps aux --sort=-%mem > /tmp/logs/system/ps.txt dpkg -l > /tmp/logs/system/packages.txt CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) if [ ! -z "$CONTAINERS" ]; then mkdir -p "/tmp/logs/system/containers" for CONTAINER in ${CONTAINERS}; do docker logs "${CONTAINER}" > "/tmp/logs/system/containers/${CONTAINER}.txt" done fi root 162582 0.0 0.0 5400 3400 ? S 14:00 0:00 /bin/bash /root/ovnkube.sh nb-ovsdb _chrony 10678 0.0 0.0 18908 3384 ? S 13:21 0:00 /usr/sbin/chronyd -F 1 root 162509 0.0 0.0 5400 3364 ? S 14:00 0:00 /bin/bash /root/ovnkube.sh sb-ovsdb root 98768 0.0 0.0 4384 3252 ? Ss 13:46 0:00 /bin/bash -x -e -m -c CEPH_CLIENT_ID=ceph-client.rgw.ceph.a PERIODICITY=daily LOG_ROTATE_CEPH_FILE=/etc/logrotate.d/ceph LOG_MAX_SIZE=500M ROTATE=7 # edit the logrotate file to only rotate a specific daemon log # otherwise we will logrotate log files without reloading certain daemons # this might happen when multiple daemons run on the same machine sed -i "s|*.log|$CEPH_CLIENT_ID.log|" "$LOG_ROTATE_CEPH_FILE" # replace default daily with given user input sed --in-place "s/daily/$PERIODICITY/g" "$LOG_ROTATE_CEPH_FILE" # replace rotate count, default 7 for all ceph daemons other than rbd-mirror sed --in-place "s/rotate 7/rotate $ROTATE/g" "$LOG_ROTATE_CEPH_FILE" if [ "$LOG_MAX_SIZE" != "0" ]; then ?# adding maxsize $LOG_MAX_SIZE at the 4th line of the logrotate config file with 4 spaces to maintain indentation ?sed --in-place "4i \ \ \ \ maxsize $LOG_MAX_SIZE" "$LOG_ROTATE_CEPH_FILE" fi while true; do ?# we don't force the logrorate but we let the logrotate binary handle the rotation based on user's input for periodicity and size ?logrotate --verbose "$LOG_ROTATE_CEPH_FILE" ?sleep 15m done zuul 700 0.0 0.0 168464 2976 ? S 13:18 0:00 (sd-pam) root 312469 0.0 0.0 27976 2948 ? S 14:30 0:00 keepalived -f /etc/keepalived/keepalived.conf --dont-fork --log-console --log-detail --dump-conf 42424 420599 1.0 0.0 4476 2772 ? Ss 14:55 0:00 /bin/bash /tmp/heat-engine-cleaner.sh root 19507 0.0 0.0 7764 2400 ? Ss 13:24 0:00 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/crash.instance/unit.run root 12118 0.0 0.0 7764 2396 ? Ss 13:22 0:00 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mon.instance/unit.run root 33938 0.0 0.0 7764 2388 ? Ss 13:26 0:00 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.2/unit.run root 12481 0.0 0.0 7764 2372 ? Ss 13:22 0:00 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/mgr.instance.csrfqj/unit.run root 24439 0.0 0.0 7764 2344 ? Ss 13:25 0:00 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.0/unit.run root 29171 0.0 0.0 7764 2336 ? Ss 13:26 0:00 /bin/bash /var/lib/ceph/4837cbf8-4f90-4300-b3f6-726c9b9f89b4/osd.1/unit.run 999 93952 0.0 0.0 2808 2020 ? Ss 13:46 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 180501 0.0 0.0 2812 2004 ? Ss 14:04 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 163960 0.0 0.0 2812 1992 ? Ss 14:00 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 196411 0.0 0.0 2816 1992 ? Ss 14:08 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 114369 0.0 0.0 2808 1964 ? Ss 13:51 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 215656 0.0 0.0 2816 1948 ? Ss 14:12 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 68840 0.0 0.0 2804 1920 ? Ss 13:38 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 238468 0.0 0.0 2816 1872 ? Ss 14:18 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 79520 0.0 0.0 2808 1844 ? Ss 13:42 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 267417 0.0 0.0 2820 1844 ? Ss 14:25 0:00 /bin/sh /opt/rabbitmq/sbin/rabbitmq-server 999 238587 0.0 0.0 3664 1532 ? Ss 14:18 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 79626 0.0 0.0 3656 1520 ? Ss 13:42 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 uuidd 1388 0.0 0.0 9620 1484 ? Ss 13:19 0:00 /usr/sbin/uuidd --socket-activation 999 267612 0.0 0.0 3668 1392 ? Ss 14:25 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 164064 0.0 0.0 3660 1364 ? Ss 14:00 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 196558 0.0 0.0 3664 1336 ? Ss 14:08 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 114544 0.0 0.0 3656 1332 ? Ss 13:51 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 root 393624 0.0 0.0 4940 1328 ? S+ 14:46 0:00 /usr/bin/coreutils --coreutils-prog-shebang=sleep /usr/bin/sleep 15m 999 94067 0.0 0.0 3656 1324 ? Ss 13:46 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 215741 0.0 0.0 3664 1308 ? Ss 14:12 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 180639 0.0 0.0 3660 1304 ? Ss 14:04 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 68936 0.0 0.0 3652 1296 ? Ss 13:38 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 215930 0.0 0.0 2816 1220 ? Ss 14:12 0:00 /bin/sh -s rabbit_disk_monitor 999 196758 0.0 0.0 2816 1204 ? Ss 14:08 0:00 /bin/sh -s rabbit_disk_monitor 999 238765 0.0 0.0 2816 1204 ? Ss 14:18 0:00 /bin/sh -s rabbit_disk_monitor 999 267802 0.0 0.0 2820 1196 ? Ss 14:25 0:00 /bin/sh -s rabbit_disk_monitor 999 79789 0.0 0.0 2808 1192 ? Ss 13:42 0:00 /bin/sh -s rabbit_disk_monitor root 159418 0.0 0.0 2572 1164 ? Ss 13:59 0:00 /tini -s -- /tmp/openvswitch-vswitchd.sh start root 420102 0.0 0.0 2600 1160 ? S 14:54 0:00 sleep 15 root 420692 0.0 0.0 2600 1160 ? S 14:55 0:00 sleep 15 root 162700 0.0 0.0 2636 1156 ? S 14:00 0:00 tail --follow=name /var/log/ovn/ovsdb-server-sb.log root 188252 0.0 0.0 2636 1136 ? S 14:06 0:00 tail --follow=name /var/log/ovn/ovn-controller.log 999 164197 0.0 0.0 2812 1120 ? Ss 14:00 0:00 /bin/sh -s rabbit_disk_monitor root 420488 0.0 0.0 2600 1116 ? S 14:55 0:00 sleep 15 root 163156 0.0 0.0 2636 1108 ? S 14:00 0:00 tail --follow=name /var/log/ovn/ovn-northd.log root 420099 0.0 0.0 2600 1108 ? S 14:54 0:00 sleep 15 999 69097 0.0 0.0 2804 1104 ? Ss 13:38 0:00 /bin/sh -s rabbit_disk_monitor 999 196434 0.0 0.0 2700 1092 ? Ss 14:08 0:00 erl_child_setup 1048576 999 267439 0.0 0.0 2704 1088 ? Ss 14:25 0:00 erl_child_setup 1048576 999 79544 0.0 0.0 2692 1084 ? Ss 13:42 0:00 erl_child_setup 1048576 999 93978 0.0 0.0 2692 1084 ? Ss 13:46 0:00 erl_child_setup 1048576 999 94297 0.0 0.0 2808 1084 ? Ss 13:46 0:00 /bin/sh -s rabbit_disk_monitor 999 180792 0.0 0.0 2812 1084 ? Ss 14:04 0:00 /bin/sh -s rabbit_disk_monitor 999 68862 0.0 0.0 2688 1080 ? Ss 13:38 0:00 erl_child_setup 1048576 999 114392 0.0 0.0 2692 1080 ? Ss 13:51 0:00 erl_child_setup 1048576 root 162723 0.0 0.0 2636 1080 ? S 14:00 0:00 tail --follow=name /var/log/ovn/ovsdb-server-nb.log 999 163982 0.0 0.0 2696 1080 ? Ss 14:00 0:00 erl_child_setup 1048576 999 114697 0.0 0.0 2808 1076 ? Ss 13:51 0:00 /bin/sh -s rabbit_disk_monitor root 627 0.0 0.0 6176 1072 tty1 Ss+ 13:18 0:00 /sbin/agetty -o -p -- \u --noclear tty1 linux root 628 0.0 0.0 6220 1072 ttyS0 Ss+ 13:18 0:00 /sbin/agetty -o -p -- \u --keep-baud 115200,57600,38400,9600 ttyS0 vt220 999 215679 0.0 0.0 2700 1068 ? Ss 14:12 0:00 erl_child_setup 1048576 999 238502 0.0 0.0 2700 1064 ? Ss 14:18 0:00 erl_child_setup 1048576 999 180525 0.0 0.0 2696 1056 ? Ss 14:04 0:00 erl_child_setup 1048576 42424 411299 0.0 0.0 2708 1036 ? S 14:52 0:00 sleep 300 root 420827 0.0 0.0 2892 988 ? S 14:55 0:00 /bin/sh -c echo BECOME-SUCCESS-tclfwndojrxucjzgfmcnavemirkeabkm ; /usr/bin/python3 zuul 420824 0.0 0.0 2892 980 ? Ss 14:55 0:00 sh -c /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-tclfwndojrxucjzgfmcnavemirkeabkm ; /usr/bin/python3'"'"' && sleep 0' zuul 420825 0.0 0.0 2892 972 ? S 14:55 0:00 /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-tclfwndojrxucjzgfmcnavemirkeabkm ; /usr/bin/python3' && sleep 0 root 157522 0.0 0.0 12924 764 ? Ss 13:59 0:00 /sbin/iscsid ubuntu 48575 0.0 0.0 1226068 656 ? Sl 13:30 0:00 /var/lib/mysql/mysql-state-monitor _chrony 10679 0.0 0.0 10580 544 ? S 13:21 0:00 /usr/sbin/chronyd -F 1 999 196559 0.0 0.0 3664 128 ? S 14:08 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 215742 0.0 0.0 3664 128 ? S 14:12 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 238588 0.0 0.0 3664 128 ? S 14:18 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 267613 0.0 0.0 3668 128 ? S 14:25 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 164065 0.0 0.0 3660 124 ? S 14:00 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 180640 0.0 0.0 3660 124 ? S 14:04 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 267626 0.0 0.0 3672 124 ? S 14:25 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon 999 94068 0.0 0.0 3656 120 ? S 13:46 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 114545 0.0 0.0 3656 120 ? S 13:51 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 196569 0.0 0.0 3668 120 ? S 14:08 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon 999 215755 0.0 0.0 3668 120 ? S 14:12 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon 999 238598 0.0 0.0 3668 120 ? S 14:18 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon 999 68937 0.0 0.0 3652 116 ? S 13:38 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 79627 0.0 0.0 3656 116 ? S 13:42 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/inet_gethost 4 999 164075 0.0 0.0 3664 116 ? S 14:00 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon 999 180671 0.0 0.0 3664 116 ? S 14:04 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon 999 94078 0.0 0.0 3660 112 ? S 13:46 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon 999 114555 0.0 0.0 3660 112 ? S 13:51 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon 999 68947 0.0 0.0 3656 108 ? S 13:38 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon 999 79637 0.0 0.0 3660 108 ? S 13:42 0:00 /opt/erlang/lib/erlang/erts-15.2.7.2/bin/epmd -daemon root 12221 0.0 0.0 1048 4 ? Ss 13:22 0:00 /sbin/docker-init -- /usr/bin/ceph-mon -n mon.instance -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true root 12589 0.0 0.0 1048 4 ? Ss 13:22 0:00 /sbin/docker-init -- /usr/bin/ceph-mgr -n mgr.instance.csrfqj -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 19577 0.0 0.0 1048 4 ? Ss 13:24 0:00 /sbin/docker-init -- /usr/bin/ceph-crash -n client.crash.instance root 24758 0.0 0.0 1048 4 ? Ss 13:25 0:00 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 29487 0.0 0.0 1048 4 ? Ss 13:26 0:00 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug root 34254 0.0 0.0 1048 4 ? Ss 13:26 0:00 /sbin/docker-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug 65535 39723 0.0 0.0 1028 4 ? Ss 13:28 0:00 /pause 65535 39744 0.0 0.0 1028 4 ? Ss 13:28 0:00 /pause 65535 39770 0.0 0.0 1028 4 ? Ss 13:28 0:00 /pause 65535 39791 0.0 0.0 1028 4 ? Ss 13:28 0:00 /pause 65535 39805 0.0 0.0 1028 4 ? Ss 13:28 0:00 /pause 65535 40217 0.0 0.0 1028 4 ? Ss 13:28 0:00 /pause 65535 40994 0.0 0.0 1028 4 ? Ss 13:28 0:00 /pause 65535 41038 0.0 0.0 1028 4 ? Ss 13:28 0:00 /pause 65535 42354 0.0 0.0 1028 4 ? Ss 13:29 0:00 /pause 65535 42756 0.0 0.0 1028 4 ? Ss 13:29 0:00 /pause 65535 42800 0.0 0.0 1028 4 ? Ss 13:29 0:00 /pause 65535 43223 0.0 0.0 1028 4 ? Ss 13:29 0:00 /pause 65535 43693 0.0 0.0 1028 4 ? Ss 13:29 0:00 /pause 65535 43749 0.0 0.0 1028 4 ? Ss 13:29 0:00 /pause 65535 43783 0.0 0.0 1028 4 ? Ss 13:29 0:00 /pause 65535 46076 0.0 0.0 1028 4 ? Ss 13:29 0:00 /pause 65535 46154 0.0 0.0 1028 4 ? Ss 13:29 0:00 /pause systemd+ 46384 0.0 0.0 224 4 ? Ss 13:29 0:00 /usr/bin/dumb-init -- /nginx-ingress-controller --default-backend-service=ingress-nginx/ingress-nginx-defaultbackend --publish-service=ingress-nginx/ingress-nginx-controller --election-id=ingress-nginx-leader --controller-class=k8s.io/ingress-nginx --ingress-class=atmosphere --configmap=ingress-nginx/ingress-nginx-controller --tcp-services-configmap=ingress-nginx/ingress-nginx-tcp --udp-services-configmap=ingress-nginx/ingress-nginx-udp --validating-webhook=:7443 --validating-webhook-certificate=/usr/local/certificates/cert --validating-webhook-key=/usr/local/certificates/key --enable-metrics=true 65535 46890 0.0 0.0 1028 4 ? Ss 13:30 0:00 /pause 65535 46989 0.0 0.0 1028 4 ? Ss 13:30 0:00 /pause 65535 47334 0.0 0.0 1028 4 ? Ss 13:30 0:00 /pause 65535 47841 0.0 0.0 1028 4 ? Ss 13:30 0:00 /pause 65535 48246 0.0 0.0 1028 4 ? Ss 13:30 0:00 /pause 65535 50375 0.0 0.0 1028 4 ? Ss 13:31 0:00 /pause 65535 50639 0.0 0.0 1028 4 ? Ss 13:31 0:00 /pause 65535 55787 0.0 0.0 1028 4 ? Ss 13:35 0:00 /pause 65535 56990 0.0 0.0 1028 4 ? Ss 13:35 0:00 /pause 65535 57069 0.0 0.0 1028 4 ? Ss 13:35 0:00 /pause 65535 57196 0.0 0.0 1028 4 ? Ss 13:35 0:00 /pause 65535 57472 0.0 0.0 1028 4 ? Ss 13:35 0:00 /pause nobody 62535 0.0 0.0 1028 4 ? Ss 13:37 0:00 /pause nobody 62614 0.0 0.0 1028 4 ? Ss 13:37 0:00 /pause nobody 62657 0.0 0.0 1028 4 ? Ss 13:37 0:00 /pause 472 62708 0.0 0.0 1028 4 ? Ss 13:37 0:00 /pause zuul 63919 0.0 0.0 1028 4 ? Ss 13:37 0:00 /pause zuul 64057 0.0 0.0 1028 4 ? Ss 13:37 0:00 /pause systemd+ 64576 0.0 0.0 1028 4 ? Ss 13:37 0:00 /pause 65535 64742 0.0 0.0 1028 4 ? Ss 13:37 0:00 /pause 65535 64789 0.0 0.0 1028 4 ? Ss 13:37 0:00 /pause 65535 65139 0.0 0.0 1028 4 ? Ss 13:37 0:00 /pause zuul 65386 0.0 0.0 1028 4 ? Ss 13:37 0:00 /pause 10001 65692 0.0 0.0 1028 4 ? Ss 13:38 0:00 /pause nobody 65969 0.0 0.0 1028 4 ? Ss 13:38 0:00 /pause nobody 66736 0.0 0.0 1028 4 ? Ss 13:38 0:00 /pause 999 68436 0.0 0.0 1028 4 ? Ss 13:38 0:00 /pause 42424 70186 0.0 0.0 1028 4 ? Ss 13:38 0:00 /pause 999 79427 0.0 0.0 1028 4 ? Ss 13:42 0:00 /pause 42424 80965 0.0 0.0 1028 4 ? Ss 13:42 0:00 /pause 65535 87155 0.0 0.0 1028 4 ? Ss 13:44 0:00 /pause 999 93739 0.0 0.0 1028 4 ? Ss 13:46 0:00 /pause 65535 98347 0.0 0.0 1028 4 ? Ss 13:46 0:00 /pause 65535 98349 0.0 0.0 1028 4 ? Ss 13:46 0:00 /pause 42424 110764 0.0 0.0 1028 4 ? Ss 13:50 0:00 /pause 42424 112202 0.0 0.0 1028 4 ? Ss 13:51 0:00 /pause 42424 112326 0.0 0.0 1028 4 ? Ss 13:51 0:00 /pause 999 114254 0.0 0.0 1028 4 ? Ss 13:51 0:00 /pause 42424 116102 0.0 0.0 1028 4 ? Ss 13:52 0:00 /pause 42424 116187 0.0 0.0 1028 4 ? Ss 13:52 0:00 /pause 42424 129203 0.0 0.0 1028 4 ? Ss 13:56 0:00 /pause 42424 129210 0.0 0.0 1028 4 ? Ss 13:56 0:00 /pause 42424 131769 0.0 0.0 1028 4 ? Ss 13:56 0:00 /pause 42424 158894 0.0 0.0 1028 4 ? Ss 13:59 0:00 /pause 65535 160813 0.0 0.0 1028 4 ? Ss 14:00 0:00 /pause 65535 162147 0.0 0.0 1028 4 ? Ss 14:00 0:00 /pause 65535 162184 0.0 0.0 1028 4 ? Ss 14:00 0:00 /pause root 162400 0.0 0.0 1028 4 ? Ss 14:00 0:00 /pause 999 163755 0.0 0.0 1028 4 ? Ss 14:00 0:00 /pause 42424 165917 0.0 0.0 1028 4 ? Ss 14:01 0:00 /pause 42424 166003 0.0 0.0 1028 4 ? Ss 14:01 0:00 /pause 42424 166319 0.0 0.0 1028 4 ? Ss 14:01 0:00 /pause 42424 166461 0.0 0.0 1028 4 ? Ss 14:01 0:00 /pause 42424 166513 0.0 0.0 1028 4 ? Ss 14:01 0:00 /pause 42424 166825 0.0 0.0 1028 4 ? Ss 14:01 0:00 /pause 999 180326 0.0 0.0 1028 4 ? Ss 14:04 0:00 /pause 42424 182739 0.0 0.0 1028 4 ? Ss 14:05 0:00 /pause 42424 182768 0.0 0.0 1028 4 ? Ss 14:05 0:00 /pause 42424 182769 0.0 0.0 1028 4 ? Ss 14:05 0:00 /pause 42424 182984 0.0 0.0 1028 4 ? Ss 14:05 0:00 /pause 42424 183053 0.0 0.0 1028 4 ? Ss 14:05 0:00 /pause 65535 187191 0.0 0.0 1028 4 ? Ss 14:06 0:00 /pause 999 196197 0.0 0.0 1028 4 ? Ss 14:08 0:00 /pause 42424 198413 0.0 0.0 1028 4 ? Ss 14:08 0:00 /pause 42424 198558 0.0 0.0 1028 4 ? Ss 14:08 0:00 /pause 42424 198593 0.0 0.0 1028 4 ? Ss 14:08 0:00 /pause 999 215465 0.0 0.0 1028 4 ? Ss 14:12 0:00 /pause 65535 226249 0.0 0.0 1028 4 ? Ss 14:15 0:00 /pause 65535 226285 0.0 0.0 1028 4 ? Ss 14:15 0:00 /pause 65535 226368 0.0 0.0 1028 4 ? Ss 14:15 0:00 /pause 65535 226570 0.0 0.0 1028 4 ? Ss 14:15 0:00 /pause 999 238273 0.0 0.0 1028 4 ? Ss 14:18 0:00 /pause 65535 240810 0.0 0.0 1028 4 ? Ss 14:19 0:00 /pause 65535 242009 0.0 0.0 1028 4 ? Ss 14:19 0:00 /pause 65535 243048 0.0 0.0 1028 4 ? Ss 14:19 0:00 /pause 65535 244016 0.0 0.0 1028 4 ? Ss 14:19 0:00 /pause 65535 245723 0.0 0.0 1028 4 ? Ss 14:20 0:00 /pause 65535 247575 0.0 0.0 1028 4 ? Ss 14:20 0:00 /pause 65535 247679 0.0 0.0 1028 4 ? Ss 14:20 0:00 /pause 42424 257196 0.0 0.0 1028 4 ? Ss 14:22 0:00 /pause 65535 257707 0.0 0.0 1028 4 ? Ss 14:22 0:00 /pause 999 267238 0.0 0.0 1028 4 ? Ss 14:25 0:00 /pause 42424 271222 0.0 0.0 1028 4 ? Ss 14:26 0:00 /pause 42424 271363 0.0 0.0 1028 4 ? Ss 14:26 0:00 /pause 42424 271420 0.0 0.0 1028 4 ? Ss 14:26 0:00 /pause 42424 271500 0.0 0.0 1028 4 ? Ss 14:26 0:00 /pause 42424 304910 0.0 0.0 1028 4 ? Ss 14:29 0:00 /pause 65535 307726 0.0 0.0 1028 4 ? Ss 14:29 0:00 /pause 65535 309142 0.0 0.0 1028 4 ? Ss 14:29 0:00 /pause 42424 420355 1.0 0.0 1028 4 ? Ss 14:55 0:00 /pause root 2 0.0 0.0 0 0 ? S 13:17 0:00 [kthreadd] root 3 0.0 0.0 0 0 ? I< 13:17 0:00 [rcu_gp] root 4 0.0 0.0 0 0 ? I< 13:17 0:00 [rcu_par_gp] root 5 0.0 0.0 0 0 ? I< 13:17 0:00 [slub_flushwq] root 6 0.0 0.0 0 0 ? I< 13:17 0:00 [netns] root 8 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/0:0H-events_highpri] root 10 0.0 0.0 0 0 ? I< 13:17 0:00 [mm_percpu_wq] root 11 0.0 0.0 0 0 ? S 13:17 0:00 [rcu_tasks_rude_] root 12 0.0 0.0 0 0 ? S 13:17 0:00 [rcu_tasks_trace] root 13 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/0] root 14 0.1 0.0 0 0 ? I 13:17 0:09 [rcu_sched] root 15 0.0 0.0 0 0 ? S 13:17 0:00 [migration/0] root 16 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/0] root 18 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/0] root 19 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/1] root 20 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/1] root 21 0.0 0.0 0 0 ? S 13:17 0:01 [migration/1] root 22 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/1] root 24 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/1:0H-events_highpri] root 25 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/2] root 26 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/2] root 27 0.0 0.0 0 0 ? S 13:17 0:01 [migration/2] root 28 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/2] root 30 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/2:0H-events_highpri] root 31 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/3] root 32 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/3] root 33 0.0 0.0 0 0 ? S 13:17 0:01 [migration/3] root 34 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/3] root 36 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/3:0H-events_highpri] root 37 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/4] root 38 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/4] root 39 0.0 0.0 0 0 ? S 13:17 0:01 [migration/4] root 40 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/4] root 42 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/4:0H-events_highpri] root 43 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/5] root 44 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/5] root 45 0.0 0.0 0 0 ? S 13:17 0:01 [migration/5] root 46 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/5] root 48 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/5:0H] root 49 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/6] root 50 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/6] root 51 0.0 0.0 0 0 ? S 13:17 0:01 [migration/6] root 52 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/6] root 54 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/6:0H-events_highpri] root 55 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/7] root 56 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/7] root 57 0.0 0.0 0 0 ? S 13:17 0:01 [migration/7] root 58 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/7] root 60 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/7:0H-events_highpri] root 61 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/8] root 62 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/8] root 63 0.0 0.0 0 0 ? S 13:17 0:01 [migration/8] root 64 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/8] root 66 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/8:0H-events_highpri] root 67 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/9] root 68 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/9] root 69 0.0 0.0 0 0 ? S 13:17 0:01 [migration/9] root 70 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/9] root 72 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/9:0H-events_highpri] root 73 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/10] root 74 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/10] root 75 0.0 0.0 0 0 ? S 13:17 0:01 [migration/10] root 76 0.0 0.0 0 0 ? S 13:17 0:00 [ksoftirqd/10] root 78 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/10:0H-events_highpri] root 79 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/11] root 80 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/11] root 81 0.0 0.0 0 0 ? S 13:17 0:01 [migration/11] root 82 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/11] root 84 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/11:0H-events_highpri] root 85 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/12] root 86 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/12] root 87 0.0 0.0 0 0 ? S 13:17 0:01 [migration/12] root 88 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/12] root 90 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/12:0H-events_highpri] root 91 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/13] root 92 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/13] root 93 0.0 0.0 0 0 ? S 13:17 0:01 [migration/13] root 94 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/13] root 96 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/13:0H-events_highpri] root 97 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/14] root 98 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/14] root 99 0.0 0.0 0 0 ? S 13:17 0:01 [migration/14] root 100 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/14] root 102 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/14:0H-events_highpri] root 103 0.0 0.0 0 0 ? S 13:17 0:00 [cpuhp/15] root 104 0.0 0.0 0 0 ? S 13:17 0:00 [idle_inject/15] root 105 0.0 0.0 0 0 ? S 13:17 0:01 [migration/15] root 106 0.0 0.0 0 0 ? S 13:17 0:01 [ksoftirqd/15] root 108 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/15:0H-events_highpri] root 109 0.0 0.0 0 0 ? S 13:17 0:00 [kdevtmpfs] root 110 0.0 0.0 0 0 ? I< 13:17 0:00 [inet_frag_wq] root 111 0.0 0.0 0 0 ? S 13:17 0:00 [kauditd] root 112 0.0 0.0 0 0 ? S 13:17 0:00 [khungtaskd] root 113 0.0 0.0 0 0 ? S 13:17 0:00 [oom_reaper] root 114 0.0 0.0 0 0 ? I< 13:17 0:00 [writeback] root 115 0.0 0.0 0 0 ? S 13:17 0:00 [kcompactd0] root 116 0.0 0.0 0 0 ? SN 13:17 0:00 [ksmd] root 117 0.0 0.0 0 0 ? SN 13:17 0:00 [khugepaged] root 166 0.0 0.0 0 0 ? I< 13:17 0:00 [kintegrityd] root 167 0.0 0.0 0 0 ? I< 13:17 0:00 [kblockd] root 168 0.0 0.0 0 0 ? I< 13:17 0:00 [blkcg_punt_bio] root 169 0.0 0.0 0 0 ? I< 13:17 0:00 [tpm_dev_wq] root 170 0.0 0.0 0 0 ? I< 13:17 0:00 [ata_sff] root 171 0.0 0.0 0 0 ? I< 13:17 0:00 [md] root 172 0.0 0.0 0 0 ? I< 13:17 0:00 [edac-poller] root 173 0.0 0.0 0 0 ? I< 13:17 0:00 [devfreq_wq] root 174 0.0 0.0 0 0 ? S 13:17 0:00 [watchdogd] root 176 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/0:1H-kblockd] root 178 0.1 0.0 0 0 ? S 13:17 0:06 [kswapd0] root 179 0.0 0.0 0 0 ? S 13:17 0:00 [ecryptfs-kthrea] root 181 0.0 0.0 0 0 ? I< 13:17 0:00 [kthrotld] root 183 0.0 0.0 0 0 ? I< 13:17 0:00 [acpi_thermal_pm] root 185 0.0 0.0 0 0 ? S 13:17 0:00 [scsi_eh_0] root 186 0.0 0.0 0 0 ? I< 13:17 0:00 [scsi_tmf_0] root 187 0.0 0.0 0 0 ? S 13:17 0:00 [scsi_eh_1] root 188 0.0 0.0 0 0 ? I< 13:17 0:00 [scsi_tmf_1] root 190 0.0 0.0 0 0 ? I< 13:17 0:00 [vfio-irqfd-clea] root 191 0.0 0.0 0 0 ? I< 13:17 0:00 [mld] root 192 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/2:1H-kblockd] root 193 0.0 0.0 0 0 ? I< 13:17 0:00 [ipv6_addrconf] root 194 0.1 0.0 0 0 ? I 13:17 0:06 [kworker/u32:4-flush-252:0] root 215 0.0 0.0 0 0 ? I< 13:17 0:00 [kstrp] root 219 0.0 0.0 0 0 ? I< 13:17 0:00 [zswap-shrink] root 220 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/u33:0] root 225 0.0 0.0 0 0 ? I< 13:17 0:00 [charger_manager] root 247 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/15:1H-kblockd] root 268 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/14:1H-kblockd] root 284 0.0 0.0 0 0 ? S 13:17 0:00 [hwrng] root 285 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/11:1H-kblockd] root 286 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/7:1H-kblockd] root 288 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/10:1H-kblockd] root 308 0.3 0.0 0 0 ? S 13:17 0:17 [jbd2/vda1-8] root 309 0.0 0.0 0 0 ? I< 13:17 0:00 [ext4-rsv-conver] root 313 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/4:1H-kblockd] root 314 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/13:1H-kblockd] root 323 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/12:1H-kblockd] root 348 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/5:1H-kblockd] root 351 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/3:1H-kblockd] root 386 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/1:1H-kblockd] root 399 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/8:1H-kblockd] root 451 0.0 0.0 0 0 ? I< 13:17 0:00 [cryptd] root 452 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/9:1H-kblockd] root 489 0.0 0.0 0 0 ? I< 13:17 0:00 [kworker/6:1H-kblockd] root 9062 0.0 0.0 0 0 ? I< 13:20 0:00 [kdmflush] root 9079 0.0 0.0 0 0 ? I< 13:20 0:00 [kdmflush] root 9095 0.0 0.0 0 0 ? I< 13:20 0:00 [kdmflush] root 10193 0.0 0.0 0 0 ? I 13:21 0:04 [kworker/u32:7-flush-252:0] root 35841 0.0 0.0 0 0 ? I< 13:26 0:00 [ceph-msgr] root 35843 0.0 0.0 0 0 ? I< 13:26 0:00 [rbd] root 48661 0.0 0.0 0 0 ? I< 13:30 0:00 [dio/vda1] root 50132 0.0 0.0 0 0 ? I 13:31 0:03 [kworker/u32:9-ext4-rsv-conversion] root 62786 0.0 0.0 0 0 ? I 13:37 0:04 [kworker/u32:1-flush-252:0] root 62788 0.0 0.0 0 0 ? I 13:37 0:03 [kworker/u32:3-ext4-rsv-conversion] root 131075 0.0 0.0 0 0 ? I 13:56 0:00 [kworker/11:5-rcu_gp] root 150308 0.0 0.0 0 0 ? I< 13:58 0:00 [kmpathd] root 150309 0.0 0.0 0 0 ? I< 13:58 0:00 [kmpath_handlerd] root 157525 0.0 0.0 0 0 ? I< 13:59 0:00 [iscsi_eh] root 157526 0.0 0.0 0 0 ? I< 13:59 0:00 [iscsi_conn_clea] root 159463 0.0 0.0 0 0 ? I< 13:59 0:00 [ib-comp-wq] root 159464 0.0 0.0 0 0 ? I< 13:59 0:00 [ib-comp-unb-wq] root 159465 0.0 0.0 0 0 ? I< 13:59 0:00 [ib_mcast] root 159466 0.0 0.0 0 0 ? I< 13:59 0:00 [ib_nl_sa_wq] root 162903 0.1 0.0 0 0 ? I 14:00 0:03 [kworker/u32:6-ext4-rsv-conversion] root 189951 0.0 0.0 0 0 ? I 14:06 0:02 [kworker/u32:8-flush-252:0] root 189955 0.0 0.0 0 0 ? I 14:06 0:00 [kworker/u32:16-flush-252:0] root 207230 0.0 0.0 0 0 ? I 14:10 0:00 [kworker/13:0-cgroup_release] root 210859 0.0 0.0 0 0 ? I 14:11 0:00 [kworker/7:0-events] root 238120 0.0 0.0 0 0 ? I 14:18 0:00 [kworker/14:2-events] root 249967 0.0 0.0 0 0 ? I 14:20 0:00 [kworker/4:11-cgroup_free] root 250189 0.0 0.0 0 0 ? I 14:21 0:00 [kworker/15:3-events] root 251008 0.0 0.0 0 0 ? I 14:21 0:00 [kworker/2:3-rcu_gp] root 264152 0.0 0.0 0 0 ? I 14:24 0:00 [kworker/u32:12-events_unbound] root 272286 0.0 0.0 0 0 ? I 14:26 0:00 [kworker/12:3-cgroup_release] root 277812 0.0 0.0 0 0 ? I 14:28 0:01 [kworker/u32:15-flush-252:0] root 278786 0.0 0.0 0 0 ? I 14:28 0:00 [kworker/3:0-events] root 336865 0.0 0.0 0 0 ? I 14:34 0:00 [kworker/8:1-cgroup_free] root 339151 0.0 0.0 0 0 ? I 14:35 0:00 [kworker/7:2-events] root 339222 0.0 0.0 0 0 ? I 14:35 0:00 [kworker/0:0-events] root 339316 0.0 0.0 0 0 ? I 14:35 0:00 [kworker/1:2-events] root 341713 0.0 0.0 0 0 ? I 14:35 0:00 [kworker/6:0-events] root 353582 0.0 0.0 0 0 ? I 14:38 0:00 [kworker/4:1-cgroup_release] root 364962 0.0 0.0 0 0 ? I 14:40 0:00 [kworker/5:3-cgroup_release] root 370105 0.0 0.0 0 0 ? I 14:41 0:00 [kworker/3:2-rcu_gp] root 372984 0.0 0.0 0 0 ? I 14:41 0:00 [kworker/15:1-cgroup_free] root 382869 0.0 0.0 0 0 ? I 14:43 0:00 [kworker/u32:0-writeback] root 382870 0.0 0.0 0 0 ? I 14:43 0:00 [kworker/9:1-events] root 383595 0.0 0.0 0 0 ? I 14:44 0:00 [kworker/u32:5-flush-253:0] root 383831 0.0 0.0 0 0 ? I 14:44 0:00 [kworker/14:4-rcu_gp] root 387188 0.0 0.0 0 0 ? I 14:45 0:00 [kworker/4:0-events] root 387363 0.0 0.0 0 0 ? I 14:45 0:00 [kworker/7:4-events] root 388301 0.0 0.0 0 0 ? I 14:45 0:00 [kworker/11:1-events] root 388577 0.0 0.0 0 0 ? I 14:45 0:00 [kworker/13:1-events] root 388607 0.0 0.0 0 0 ? I 14:45 0:00 [kworker/1:0-rcu_gp] root 388865 0.0 0.0 0 0 ? I 14:45 0:00 [kworker/10:0-inet_frag_wq] root 390701 0.0 0.0 0 0 ? I 14:46 0:00 [kworker/2:2-rcu_gp] root 393693 0.0 0.0 0 0 ? I 14:46 0:00 [kworker/10:1-events] root 393965 0.0 0.0 0 0 ? I 14:46 0:00 [kworker/9:3-events] root 394441 0.0 0.0 0 0 ? I 14:47 0:00 [kworker/5:0-events] root 394873 0.0 0.0 0 0 ? I 14:47 0:00 [kworker/0:2-events] root 394939 0.0 0.0 0 0 ? I 14:47 0:00 [kworker/13:3-events] root 399018 0.0 0.0 0 0 ? I 14:48 0:00 [kworker/8:3-events] root 399540 0.0 0.0 0 0 ? I 14:48 0:00 [kworker/12:2-events] root 402840 0.0 0.0 0 0 ? I 14:49 0:00 [kworker/6:2-events] root 404716 0.0 0.0 0 0 ? I 14:50 0:00 [kworker/3:1-events] root 405093 0.0 0.0 0 0 ? I 14:50 0:00 [kworker/4:2-events] root 405105 0.0 0.0 0 0 ? I 14:50 0:00 [kworker/4:3-events] root 411512 0.0 0.0 0 0 ? I 14:52 0:00 [kworker/15:0-events] root 414999 0.0 0.0 0 0 ? I 14:53 0:00 [kworker/1:1-events] root 417634 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/10:2-rcu_gp] root 417833 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/11:0-events] root 417866 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/5:1-rcu_gp] root 418360 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/0:1-events] root 418828 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/2:0-events] root 418917 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/12:0-events] root 418918 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/9:0-events] root 418919 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/14:0-events] root 420110 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/8:0-events] root 420158 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/9:2-events] root 420165 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/6:1] root 420173 0.0 0.0 0 0 ? I 14:54 0:00 [kworker/9:4-events]