● np0000163014 State: running Units: 506 loaded (incl. loaded aliases) Jobs: 0 queued Failed: 0 units Since: Thu 2026-04-02 14:21:03 UTC; 30min ago systemd: 255.4-1ubuntu8.14 CGroup: / ├─init.scope │ └─1 /sbin/init nofb ├─system.slice │ ├─apache-htcacheclean.service │ │ └─13911 /usr/bin/htcacheclean -d 120 -p /var/cache/apache2/mod_cache_disk -l 300M -n │ ├─apache2.service │ │ ├─121459 /usr/sbin/apache2 -k start │ │ ├─121463 /usr/sbin/apache2 -k start │ │ └─121465 /usr/sbin/apache2 -k start │ ├─containerd.service │ │ ├─20255 /usr/bin/containerd │ │ └─21112 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8 -address /run/containerd/containerd.sock │ ├─dbus.service │ │ └─699 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─dm-event.service │ │ └─113941 /usr/sbin/dmeventd -f │ ├─docker-54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8.scope │ │ ├─init.scope │ │ │ └─21135 /sbin/init │ │ ├─kubelet.slice │ │ │ ├─kubelet-kubepods.slice │ │ │ │ ├─kubelet-kubepods-besteffort.slice │ │ │ │ │ ├─kubelet-kubepods-besteffort-pod0252a4bd_02fc_4e45_b96b_87fd60aae118.slice │ │ │ │ │ │ ├─cri-containerd-38b467f0d9256afa3b38e32675958de195df595af5b33c93ef4ad091fd7ca2ce.scope │ │ │ │ │ │ │ └─24819 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,KubeadmBootstrapFormatIgnition=true,PriorityQueue=false --bootstrap-token-ttl=15m │ │ │ │ │ │ └─cri-containerd-cab203df0fa479277115976d13436c61979f71791765859f21b6ef4d545e0a28.scope │ │ │ │ │ │ └─24432 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-pod0b3e7747_7f3a_4a37_880c_7609beb2bec2.slice │ │ │ │ │ │ ├─cri-containerd-0d1c8a0dc4444ab42de6f1edc1a3d7f9d15e8257c27b683b19c06074509a75fe.scope │ │ │ │ │ │ │ └─25117 /manager --leader-elect --v=2 --diagnostics-address=127.0.0.1:8080 --insecure-diagnostics=true │ │ │ │ │ │ └─cri-containerd-feb0b8b6bb17ac3fdd5946e4a75f306d531c347d9923d5f36691d6d272a7bb3f.scope │ │ │ │ │ │ └─24759 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-pod19708082_a0bb_4e1a_8b9f_8b55bade2a97.slice │ │ │ │ │ │ ├─cri-containerd-267658bf363448825ad8effc3144e2a05b7e99f48f9ef266949239dadf0e647f.scope │ │ │ │ │ │ │ └─24642 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,ClusterResourceSet=true,ClusterTopology=true,RuntimeSDK=false,MachineSetPreflightChecks=true,MachineWaitForVolumeDetachConsiderVolumeAttachments=true,PriorityQueue=false │ │ │ │ │ │ └─cri-containerd-f15e92eac3c699ac4091b22d11aebac4ddda90272e826d9430f8393eb08e6290.scope │ │ │ │ │ │ └─24363 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-pod2b65498f_1697_43c6_a06c_b95cb04050e9.slice │ │ │ │ │ │ ├─cri-containerd-61bac5c5b3476d4d45e83534a006e7e92b17979efcd672630913702862a801da.scope │ │ │ │ │ │ │ └─23224 /pause │ │ │ │ │ │ └─cri-containerd-7f8394ff1f96657dbbd8821483c8bd4d1c2cdd2a378ea35fe8be66375af9edd2.scope │ │ │ │ │ │ └─23605 local-path-provisioner --debug start --helper-image docker.io/kindest/local-path-helper:v20220607-9a4d8d2a --config /etc/config/config.json │ │ │ │ │ ├─kubelet-kubepods-besteffort-pod3c159520_1948_489d_96ad_e4753123a872.slice │ │ │ │ │ │ ├─cri-containerd-c43573aa54912bff9c0d74b6ef47458ab4fbfcd9e25f44f5250dba0ace166133.scope │ │ │ │ │ │ │ └─24161 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=kube-system │ │ │ │ │ │ └─cri-containerd-ccc37416b9aa55b0d540fe8325a804eb567d99c4e67550ef009f367d5d735c4a.scope │ │ │ │ │ │ └─23568 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-podcfd58e14_dd69_4f38_9746_b9c5983f57b6.slice │ │ │ │ │ │ ├─cri-containerd-2c4b62468f8906532289d834d3d20d746566a61df2d31dc105f87c9c243ba045.scope │ │ │ │ │ │ │ └─23866 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --acme-http01-solver-image=quay.io/jetstack/cert-manager-acmesolver:v1.18.1 --max-concurrent-challenges=60 │ │ │ │ │ │ └─cri-containerd-fd50372de23838501cfd1f9593f95c913e7d080f10f61d6628f9f6f454cca9d5.scope │ │ │ │ │ │ └─23400 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-podd87bcd89_9c5c_4cfb_a125_66fbb1d9c47d.slice │ │ │ │ │ │ ├─cri-containerd-88e90d76e68188293a25b8b8c51be0f3ccf15c5f821aa794f1b0cefabe0a94a3.scope │ │ │ │ │ │ │ └─22896 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=kind-control-plane │ │ │ │ │ │ └─cri-containerd-ff188e61a32f5f07cc3bf88092974b3b46868bbc83fdde14a2d301c6228e65e7.scope │ │ │ │ │ │ └─22841 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-pode2c03ea4_9bb3_4184_b5e4_4e142c6fb653.slice │ │ │ │ │ │ ├─cri-containerd-0e3f59113c1c6e6512e47da5dbba7302a32112b2f9c352d5e9f73f4ab1595ca9.scope │ │ │ │ │ │ │ └─23560 /pause │ │ │ │ │ │ └─cri-containerd-68ceb73a5cea74ca45f26405a3bd47783a761566d102a7097661496ba37a5647.scope │ │ │ │ │ │ └─24073 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc │ │ │ │ │ └─kubelet-kubepods-besteffort-podf874c790_24e9_4b31_a814_05df2252c78f.slice │ │ │ │ │ ├─cri-containerd-be887502786ee7935d2add369d9825d38b897c67c3b50c15b89b6710ecae3e6c.scope │ │ │ │ │ │ └─24579 /pause │ │ │ │ │ └─cri-containerd-d1af59187bb078e68792bc9bfcee67bd66f3249aee55bbf56c703b117ecff25e.scope │ │ │ │ │ └─24953 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,ClusterTopology=true,KubeadmBootstrapFormatIgnition=true,PriorityQueue=false │ │ │ │ ├─kubelet-kubepods-burstable.slice │ │ │ │ │ ├─kubelet-kubepods-burstable-pod0656ab70da313d6449b17f099a2a3110.slice │ │ │ │ │ │ ├─cri-containerd-2ab87108ea1e6b8b64d509ef7c1c9342d77a7d272ccb9d4240e562e03b234ab1.scope │ │ │ │ │ │ │ └─22131 etcd --advertise-client-urls=https://172.18.0.2:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://172.18.0.2:2380 --initial-cluster=kind-control-plane=https://172.18.0.2:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://172.18.0.2:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://172.18.0.2:2380 --name=kind-control-plane --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ │ │ │ │ │ └─cri-containerd-6208997ab8fc91afafd91db67859c3b91025856a1f308a614267a5a3d0c37366.scope │ │ │ │ │ │ └─21804 /pause │ │ │ │ │ ├─kubelet-kubepods-burstable-pod53ff6c8abd472f64bc9a9afbd3a471a9.slice │ │ │ │ │ │ ├─cri-containerd-870df6abbc20c8d337ecc4c2bce2a50cd533c2789b8639678ed04c6f1d760e4c.scope │ │ │ │ │ │ │ └─21917 /pause │ │ │ │ │ │ └─cri-containerd-e8f0ee1d61dfec222520c595b1deb5719cdeebd6ece0c11d1ae8d92037973a09.scope │ │ │ │ │ │ └─22041 kube-controller-manager --allocate-node-cidrs=true --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=127.0.0.1 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-cidr=10.244.0.0/16 --cluster-name=kind --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key "--controllers=*,bootstrapsigner,tokencleaner" --enable-hostpath-provisioner=true --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/16 --use-service-account-credentials=true │ │ │ │ │ ├─kubelet-kubepods-burstable-pod7afc15c3_ead2_46e7_9a64_97d9b2aeff9e.slice │ │ │ │ │ │ ├─cri-containerd-19e68a60cb0b1f7aaff877e9aa2ce9e215bb133217cb6c6bec51e589245f9657.scope │ │ │ │ │ │ │ └─23978 /manager --metrics-bind-address=:8443 --leader-elect --health-probe-bind-address=:8081 │ │ │ │ │ │ └─cri-containerd-24a89f3f6c95c059079539dd5111752929e7c7739c4f76340c3501a5beba8343.scope │ │ │ │ │ │ └─23468 /pause │ │ │ │ │ ├─kubelet-kubepods-burstable-podaf8e1906_1a5b_4f97_902c_ca7d36e47926.slice │ │ │ │ │ │ ├─cri-containerd-48b1c569323512f743b5d77a02134ea29982e2225d3d1e74fb3f1f47fc7e0b24.scope │ │ │ │ │ │ │ └─23654 /coredns -conf /etc/coredns/Corefile │ │ │ │ │ │ └─cri-containerd-6472d35a6a9b786a04324c9f49d7f145bd3246b78321220d11fae0c5e43970c5.scope │ │ │ │ │ │ └─23217 /pause │ │ │ │ │ ├─kubelet-kubepods-burstable-podbee69ab63b6471d4da666ee970746eae.slice │ │ │ │ │ │ ├─cri-containerd-31d336081f7b38682b523a2347ccf8f412fd51b43947bcb6cc3201680c38d863.scope │ │ │ │ │ │ │ └─21975 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=127.0.0.1 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true │ │ │ │ │ │ └─cri-containerd-7c89215ac2295c29f7bd0d888c165a63c94d11eeb8c6a7292afd33a2f3cdd899.scope │ │ │ │ │ │ └─21907 /pause │ │ │ │ │ ├─kubelet-kubepods-burstable-podd25193d9_141b_4581_a1b0_bb10a32dc523.slice │ │ │ │ │ │ ├─cri-containerd-c21138d45276460b2a29141bd3bb697e08f97fec20a617b33be15c71bc61f54e.scope │ │ │ │ │ │ │ └─23679 /coredns -conf /etc/coredns/Corefile │ │ │ │ │ │ └─cri-containerd-f459ee43bf1a88a56ea2d80d0152986c09363cfb35460b03ee8f3a0f4424f49c.scope │ │ │ │ │ │ └─23475 /pause │ │ │ │ │ └─kubelet-kubepods-burstable-podef6ebc9842be361e05ebdb6790c540b6.slice │ │ │ │ │ ├─cri-containerd-1926238e4fdf69a50346ffe73b66ea67b7f75760bf715fa5fff86efedc94f171.scope │ │ │ │ │ │ └─21844 /pause │ │ │ │ │ └─cri-containerd-3c7a2ea72ddf5756767d9d3c9628ca18ef585748147673c7e3ac210cdce69837.scope │ │ │ │ │ └─22017 kube-apiserver --advertise-address=172.18.0.2 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --runtime-config= --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/16 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ │ │ │ └─kubelet-kubepods-pod30514ae6_c3de_43c7_85f3_bc4f3ac7154d.slice │ │ │ │ ├─cri-containerd-522d017c8e1da3feb0777e65b66d2de23ebc9a86b25a0adc7a5587008b794a64.scope │ │ │ │ │ └─22935 /bin/kindnetd │ │ │ │ └─cri-containerd-d01f22860fc3f4da6c86ca2a85f5326a74793725f56e3e7216be56ae80ef1f44.scope │ │ │ │ └─22849 /pause │ │ │ └─kubelet.service │ │ │ └─22229 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --container-runtime=remote --container-runtime-endpoint=unix:///run/containerd/containerd.sock --node-ip=172.18.0.2 --node-labels= --pod-infra-container-image=registry.k8s.io/pause:3.8 --provider-id=kind://docker/kind/kind-control-plane --fail-swap-on=false --cgroup-root=/kubelet │ │ └─system.slice │ │ ├─containerd.service │ │ │ ├─21331 /usr/local/bin/containerd │ │ │ ├─21782 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6208997ab8fc91afafd91db67859c3b91025856a1f308a614267a5a3d0c37366 -address /run/containerd/containerd.sock │ │ │ ├─21818 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1926238e4fdf69a50346ffe73b66ea67b7f75760bf715fa5fff86efedc94f171 -address /run/containerd/containerd.sock │ │ │ ├─21863 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7c89215ac2295c29f7bd0d888c165a63c94d11eeb8c6a7292afd33a2f3cdd899 -address /run/containerd/containerd.sock │ │ │ ├─21890 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 870df6abbc20c8d337ecc4c2bce2a50cd533c2789b8639678ed04c6f1d760e4c -address /run/containerd/containerd.sock │ │ │ ├─22800 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id ff188e61a32f5f07cc3bf88092974b3b46868bbc83fdde14a2d301c6228e65e7 -address /run/containerd/containerd.sock │ │ │ ├─22808 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id d01f22860fc3f4da6c86ca2a85f5326a74793725f56e3e7216be56ae80ef1f44 -address /run/containerd/containerd.sock │ │ │ ├─23169 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6472d35a6a9b786a04324c9f49d7f145bd3246b78321220d11fae0c5e43970c5 -address /run/containerd/containerd.sock │ │ │ ├─23188 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 61bac5c5b3476d4d45e83534a006e7e92b17979efcd672630913702862a801da -address /run/containerd/containerd.sock │ │ │ ├─23380 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id fd50372de23838501cfd1f9593f95c913e7d080f10f61d6628f9f6f454cca9d5 -address /run/containerd/containerd.sock │ │ │ ├─23427 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 24a89f3f6c95c059079539dd5111752929e7c7739c4f76340c3501a5beba8343 -address /run/containerd/containerd.sock │ │ │ ├─23435 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id f459ee43bf1a88a56ea2d80d0152986c09363cfb35460b03ee8f3a0f4424f49c -address /run/containerd/containerd.sock │ │ │ ├─23506 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0e3f59113c1c6e6512e47da5dbba7302a32112b2f9c352d5e9f73f4ab1595ca9 -address /run/containerd/containerd.sock │ │ │ ├─23523 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id ccc37416b9aa55b0d540fe8325a804eb567d99c4e67550ef009f367d5d735c4a -address /run/containerd/containerd.sock │ │ │ ├─24342 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id f15e92eac3c699ac4091b22d11aebac4ddda90272e826d9430f8393eb08e6290 -address /run/containerd/containerd.sock │ │ │ ├─24412 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id cab203df0fa479277115976d13436c61979f71791765859f21b6ef4d545e0a28 -address /run/containerd/containerd.sock │ │ │ ├─24559 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id be887502786ee7935d2add369d9825d38b897c67c3b50c15b89b6710ecae3e6c -address /run/containerd/containerd.sock │ │ │ └─24739 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id feb0b8b6bb17ac3fdd5946e4a75f306d531c347d9923d5f36691d6d272a7bb3f -address /run/containerd/containerd.sock │ │ └─systemd-journald.service │ │ └─21316 /lib/systemd/systemd-journald │ ├─docker.service │ │ ├─20379 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock │ │ └─21206 /usr/bin/docker-proxy -proto tcp -host-ip 127.0.0.1 -host-port 42841 -container-ip 172.18.0.2 -container-port 6443 -use-listen-fd │ ├─epmd.service │ │ └─25717 /usr/bin/epmd -systemd │ ├─fsidd.service │ │ └─53941 /usr/sbin/fsidd │ ├─haproxy.service │ │ ├─12955 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock │ │ └─12957 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock │ ├─iscsid.service │ │ ├─43375 /usr/sbin/iscsid │ │ └─43376 /usr/sbin/iscsid │ ├─ksmtuned.service │ │ ├─ 4941 /bin/bash /usr/sbin/ksmtuned │ │ └─128547 sleep 60 │ ├─libvirtd.service │ │ ├─ 42323 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/lib/libvirt/libvirt_leaseshelper │ │ ├─ 42324 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/lib/libvirt/libvirt_leaseshelper │ │ └─110870 /usr/sbin/libvirtd --timeout 120 │ ├─memcached.service │ │ └─65584 /usr/bin/memcached -m 64 -p 11211 -u memcache -l 127.0.0.1 -l ::1 -P /var/run/memcached/memcached.pid │ ├─mysql.service │ │ └─62164 /usr/sbin/mysqld │ ├─nfs-blkmap.service │ │ └─53946 /usr/sbin/blkmapd │ ├─nfs-idmapd.service │ │ └─53949 /usr/sbin/rpc.idmapd │ ├─nfs-mountd.service │ │ └─53960 /usr/sbin/rpc.mountd │ ├─nfsdcld.service │ │ └─53961 /usr/sbin/nfsdcld │ ├─nmbd.service │ │ └─54587 /usr/sbin/nmbd --foreground --no-process-group │ ├─ovn-controller-vtep.service │ │ └─99715 ovn-controller-vtep -vconsole:emer -vsyslog:err -vfile:info --vtep-db=/var/run/openvswitch/db.sock --ovnsb-db=/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-controller-vtep.log --pidfile=/var/run/ovn/ovn-controller-vtep.pid --detach │ ├─ovn-controller.service │ │ └─100402 ovn-controller unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --no-chdir --log-file=/var/log/ovn/ovn-controller.log --pidfile=/var/run/ovn/ovn-controller.pid --detach │ ├─ovn-northd.service │ │ └─100136 ovn-northd -vconsole:emer -vsyslog:err -vfile:info --ovnnb-db=unix:/var/run/ovn/ovnnb_db.sock --ovnsb-db=unix:/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-northd.log --pidfile=/var/run/ovn/ovn-northd.pid --detach │ ├─ovn-ovsdb-server-nb.service │ │ └─100056 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-nb.log --remote=punix:/var/run/ovn/ovnnb_db.sock --pidfile=/var/run/ovn/ovnnb_db.pid --unixctl=/var/run/ovn/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=db:OVN_Northbound,SSL,private_key --certificate=db:OVN_Northbound,SSL,certificate --ca-cert=db:OVN_Northbound,SSL,ca_cert --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers /var/lib/ovn/ovnnb_db.db │ ├─ovn-ovsdb-server-sb.service │ │ └─100058 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-sb.log --remote=punix:/var/run/ovn/ovnsb_db.sock --pidfile=/var/run/ovn/ovnsb_db.pid --unixctl=/var/run/ovn/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=db:OVN_Southbound,SSL,private_key --certificate=db:OVN_Southbound,SSL,certificate --ca-cert=db:OVN_Southbound,SSL,ca_cert --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers /var/lib/ovn/ovnsb_db.db │ ├─ovs-vswitchd.service │ │ └─99627 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach │ ├─ovsdb-server.service │ │ └─99578 ovsdb-server /etc/openvswitch/conf.db -vconsole:emer -vsyslog:err -vfile:info --remote=punix:/var/run/openvswitch/db.sock --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --no-chdir --log-file=/var/log/openvswitch/ovsdb-server.log --pidfile=/var/run/openvswitch/ovsdb-server.pid --detach │ ├─polkit.service │ │ └─742 /usr/lib/polkit-1/polkitd --no-debug │ ├─rabbitmq-server.service │ │ ├─25837 /usr/lib/erlang/erts-13.2.2.5/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -- -root /usr/lib/erlang -bindir /usr/lib/erlang/erts-13.2.2.5/bin -progname erl -- -home /var/lib/rabbitmq -- -pa "" -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger "[]" -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -enable-feature maybe_expr │ │ ├─25847 erl_child_setup 65536 │ │ ├─25955 /usr/lib/erlang/erts-13.2.2.5/bin/inet_gethost 4 │ │ ├─25956 /usr/lib/erlang/erts-13.2.2.5/bin/inet_gethost 4 │ │ └─25969 /bin/sh -s rabbit_disk_monitor │ ├─rpc-statd.service │ │ └─53951 /usr/sbin/rpc.statd │ ├─rpcbind.service │ │ └─53314 /sbin/rpcbind -f -w │ ├─rsyslog.service │ │ └─124176 /usr/sbin/rsyslogd -n -iNONE │ ├─smbd.service │ │ ├─54520 /usr/sbin/smbd --foreground --no-process-group │ │ ├─54523 "smbd: notifyd" . │ │ └─54524 "smbd: cleanupd " │ ├─ssh.service │ │ ├─ 744 "sshd: /usr/sbin/sshd -D [listener] 25 of 10-100 startups" │ │ ├─127825 "sshd: root [priv]" │ │ ├─127826 "sshd: root [net]" │ │ ├─127848 "sshd: [accepted]" │ │ ├─127849 "sshd: [net]" │ │ ├─127850 "sshd: [accepted]" │ │ ├─127851 "sshd: [net]" │ │ ├─127852 "sshd: [accepted]" │ │ ├─127853 "sshd: [net]" │ │ ├─127854 "sshd: [accepted]" │ │ ├─127855 "sshd: [net]" │ │ ├─127891 "sshd: [accepted]" │ │ ├─127893 "sshd: [net]" │ │ ├─127897 "sshd: [accepted]" │ │ ├─127898 "sshd: [net]" │ │ ├─127899 "sshd: [accepted]" │ │ ├─127900 "sshd: [net]" │ │ ├─127904 "sshd: [accepted]" │ │ ├─127905 "sshd: [net]" │ │ ├─127928 "sshd: [accepted]" │ │ ├─127988 "sshd: [accepted]" │ │ ├─127999 "sshd: [accepted]" │ │ ├─128090 "sshd: [accepted]" │ │ ├─128134 "sshd: [accepted]" │ │ ├─128135 "sshd: [net]" │ │ ├─128136 "sshd: [accepted]" │ │ ├─128137 "sshd: [accepted]" │ │ ├─128138 "sshd: [net]" │ │ ├─128141 "sshd: [accepted]" │ │ ├─128292 "sshd: [accepted]" │ │ ├─128364 "sshd: [accepted]" │ │ ├─128365 "sshd: [net]" │ │ ├─128604 "sshd: [accepted]" │ │ ├─129705 "sshd: [accepted]" │ │ ├─129861 "sshd: [accepted]" │ │ ├─129864 "sshd: [net]" │ │ ├─129915 "sshd: [accepted]" │ │ ├─130024 "sshd: [accepted]" │ │ ├─130095 "sshd: [accepted]" │ │ └─130096 "sshd: [net]" │ ├─system-devstack.slice │ │ ├─devstack@barbican-keystone-listener.service │ │ │ ├─117066 "barbican-keystone-listener: master process [/opt/stack/data/venv/bin/barbican-keystone-listener --config-file=/etc/barbican/barbican.conf]" │ │ │ └─117314 "barbican-keystone-listener: ServiceWrapper worker(0)" │ │ ├─devstack@barbican-retry.service │ │ │ ├─116544 "barbican-retry: master process [/opt/stack/data/venv/bin/barbican-retry --config-file=/etc/barbican/barbican.conf]" │ │ │ └─116844 "barbican-retry: ServiceWrapper worker(0)" │ │ ├─devstack@barbican-svc.service │ │ │ ├─115994 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─115995 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─115996 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─115997 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ │ └─115998 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ ├─devstack@c-api.service │ │ │ ├─111318 "cinder-apiuWSGI master" │ │ │ ├─111320 "cinder-apiuWSGI worker 1" │ │ │ ├─111321 "cinder-apiuWSGI worker 2" │ │ │ ├─111322 "cinder-apiuWSGI worker 3" │ │ │ └─111323 "cinder-apiuWSGI worker 4" │ │ ├─devstack@c-bak.service │ │ │ └─112602 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-backup --config-file /etc/cinder/cinder.conf │ │ ├─devstack@c-sch.service │ │ │ └─112032 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-scheduler --config-file /etc/cinder/cinder.conf │ │ ├─devstack@c-vol.service │ │ │ ├─113189 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ │ │ └─113469 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ │ ├─devstack@etcd.service │ │ │ └─63970 /opt/stack/bin/etcd --name np0000163014 --data-dir /opt/stack/data/etcd --initial-cluster-state new --initial-cluster-token etcd-cluster-01 --initial-cluster np0000163014=http://199.204.45.109:2380 --initial-advertise-peer-urls http://199.204.45.109:2380 --advertise-client-urls http://199.204.45.109:2379 --listen-peer-urls http://0.0.0.0:2380 --listen-client-urls http://199.204.45.109:2379 --log-level=debug │ │ ├─devstack@file_tracker.service │ │ │ ├─ 63325 /bin/bash /opt/stack/devstack/tools/file_tracker.sh │ │ │ └─129995 sleep 20 │ │ ├─devstack@g-api.service │ │ │ ├─114027 "glance-apiuWSGI master" │ │ │ ├─114029 "glance-apiuWSGI worker 1" │ │ │ ├─114030 "glance-apiuWSGI worker 2" │ │ │ ├─114031 "glance-apiuWSGI worker 3" │ │ │ └─114032 "glance-apiuWSGI worker 4" │ │ ├─devstack@keystone.service │ │ │ ├─65167 "keystoneuWSGI master" │ │ │ ├─65168 "keystoneuWSGI worker 1" │ │ │ ├─65169 "keystoneuWSGI worker 2" │ │ │ ├─65170 "keystoneuWSGI worker 3" │ │ │ └─65171 "keystoneuWSGI worker 4" │ │ ├─devstack@m-api.service │ │ │ ├─121134 "manila-apiuWSGI master" │ │ │ ├─121135 "manila-apiuWSGI worker 1" │ │ │ ├─121136 "manila-apiuWSGI worker 2" │ │ │ ├─121137 "manila-apiuWSGI worker 3" │ │ │ └─121138 "manila-apiuWSGI worker 4" │ │ ├─devstack@m-dat.service │ │ │ └─127317 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-data --config-file /etc/manila/manila.conf │ │ ├─devstack@m-sch.service │ │ │ └─126741 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-scheduler --config-file /etc/manila/manila.conf │ │ ├─devstack@m-shr.service │ │ │ ├─126208 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-share --config-file /etc/manila/manila.conf │ │ │ └─126626 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-share --config-file /etc/manila/manila.conf │ │ ├─devstack@magnum-api.service │ │ │ ├─118635 "magnum-apiuWSGI master" │ │ │ ├─118636 "magnum-apiuWSGI worker 1" │ │ │ ├─118637 "magnum-apiuWSGI worker 2" │ │ │ ├─118638 "magnum-apiuWSGI worker 3" │ │ │ └─118639 "magnum-apiuWSGI worker 4" │ │ ├─devstack@magnum-cond.service │ │ │ ├─119227 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119467 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119468 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119469 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119470 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119471 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119472 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119473 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119474 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119475 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119476 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119477 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119478 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119479 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119480 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119481 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ └─119482 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─devstack@memory_tracker.service │ │ │ ├─ 62827 /bin/bash /opt/stack/devstack/tools/memory_tracker.sh │ │ │ └─130003 sleep 20 │ │ ├─devstack@n-api-meta.service │ │ │ ├─107210 "nova-api-metauWSGI master" │ │ │ ├─107211 "nova-api-metauWSGI worker 1" │ │ │ ├─107212 "nova-api-metauWSGI worker 2" │ │ │ ├─107213 "nova-api-metauWSGI worker 3" │ │ │ ├─107214 "nova-api-metauWSGI worker 4" │ │ │ └─107215 "nova-api-metauWSGI http 1" │ │ ├─devstack@n-api.service │ │ │ ├─98732 "nova-apiuWSGI master" │ │ │ ├─98733 "nova-apiuWSGI worker 1" │ │ │ ├─98734 "nova-apiuWSGI worker 2" │ │ │ ├─98735 "nova-apiuWSGI worker 3" │ │ │ └─98736 "nova-apiuWSGI worker 4" │ │ ├─devstack@n-cond-cell1.service │ │ │ ├─109228 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova_cell1.conf]" │ │ │ ├─109927 "nova-conductor: ServiceWrapper worker(0)" │ │ │ ├─109935 "nova-conductor: ServiceWrapper worker(1)" │ │ │ ├─109944 "nova-conductor: ServiceWrapper worker(2)" │ │ │ └─109954 "nova-conductor: ServiceWrapper worker(3)" │ │ ├─devstack@n-cpu.service │ │ │ └─110381 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/nova-compute --config-file /etc/nova/nova-cpu.conf │ │ ├─devstack@n-novnc-cell1.service │ │ │ └─107911 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/nova-novncproxy --config-file /etc/nova/nova_cell1.conf --web /opt/stack/novnc │ │ ├─devstack@n-sch.service │ │ │ ├─106605 "nova-scheduler: master process [/opt/stack/data/venv/bin/nova-scheduler --config-file /etc/nova/nova.conf]" │ │ │ ├─107339 "nova-scheduler: ServiceWrapper worker(0)" │ │ │ ├─107345 "nova-scheduler: ServiceWrapper worker(1)" │ │ │ ├─107352 "nova-scheduler: ServiceWrapper worker(2)" │ │ │ └─107362 "nova-scheduler: ServiceWrapper worker(3)" │ │ ├─devstack@n-super-cond.service │ │ │ ├─108690 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova.conf]" │ │ │ ├─109512 "nova-conductor: ServiceWrapper worker(0)" │ │ │ ├─109517 "nova-conductor: ServiceWrapper worker(1)" │ │ │ ├─109523 "nova-conductor: ServiceWrapper worker(2)" │ │ │ └─109534 "nova-conductor: ServiceWrapper worker(3)" │ │ ├─devstack@neutron-api.service │ │ │ ├─102068 "neutron-apiuWSGI master" │ │ │ ├─102069 "neutron-apiuWSGI worker 1" │ │ │ ├─102070 "neutron-apiuWSGI worker 2" │ │ │ ├─102071 "neutron-apiuWSGI worker 3" │ │ │ └─102072 "neutron-apiuWSGI worker 4" │ │ ├─devstack@neutron-ovn-maintenance-worker.service │ │ │ ├─103560 "neutron-ovn-maintenance-worker: master process [/opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ └─104364 "neutron-server: maintenance worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@neutron-periodic-workers.service │ │ │ ├─103064 "neutron-periodic-workers: master process [/opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ ├─103784 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ ├─103796 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ ├─103807 "neutron-server: Periodic worker for \"DbQuotaNoLockDriver\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ └─103819 "neutron-server: Periodic worker for \"L3_NAT_dbonly_mixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@neutron-rpc-server.service │ │ │ ├─102556 "neutron-rpc-server: master process [/opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ ├─103752 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ └─103760 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@o-api.service │ │ │ ├─122913 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─122915 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─122916 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─122917 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ │ └─122918 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ ├─devstack@o-da.service │ │ │ ├─123448 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/octavia-driver-agent --config-file /etc/octavia/octavia.conf │ │ │ ├─124243 "octavia-driver-agent - status_listener" │ │ │ ├─124245 "octavia-driver-agent - stats_listener" │ │ │ ├─124248 "octavia-driver-agent - get_listener" │ │ │ └─124363 "octavia-driver-agent - provider_agent -- ovn" │ │ ├─devstack@o-hk.service │ │ │ └─124055 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/octavia-housekeeping --config-file /etc/octavia/octavia.conf │ │ ├─devstack@openstack-cli-server.service │ │ │ └─61420 /opt/stack/data/venv/bin/python3 /opt/stack/devstack/files/openstack-cli-server/openstack-cli-server │ │ ├─devstack@placement-api.service │ │ │ ├─104337 "placementuWSGI master" │ │ │ ├─104338 "placementuWSGI worker 1" │ │ │ ├─104339 "placementuWSGI worker 2" │ │ │ ├─104340 "placementuWSGI worker 3" │ │ │ └─104341 "placementuWSGI worker 4" │ │ └─devstack@q-ovn-agent.service │ │ ├─100946 "neutron-ovn-agent: master process [/opt/stack/data/venv/bin/neutron-ovn-agent --config-file /etc/neutron/plugins/ml2/ovn_agent.ini]" │ │ ├─101518 "neutron-ovn-agent: ServiceWrapper worker(0)" │ │ ├─101743 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmpzexbsa_i/privsep.sock │ │ ├─105256 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmp2clb52x1/privsep.sock │ │ ├─127213 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.link_cmd --privsep_sock_path /tmp/tmpypayx3b_/privsep.sock │ │ ├─127750 sudo /opt/stack/data/venv/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf │ │ ├─127751 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf │ │ └─127791 haproxy -f /opt/stack/data/neutron/ovn-metadata-proxy/6fccf94a-88ef-49fd-a0c3-4d8e39aeea0c.conf │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─714 /sbin/agetty -o "-p -- \\u" --noclear - linux │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─715 /sbin/agetty -o "-p -- \\u" --keep-baud 115200,57600,38400,9600 - vt220 │ ├─systemd-journald.service │ │ └─18695 /usr/lib/systemd/systemd-journald │ ├─systemd-logind.service │ │ └─704 /usr/lib/systemd/systemd-logind │ ├─systemd-machined.service │ │ └─42225 /usr/lib/systemd/systemd-machined │ ├─systemd-networkd.service │ │ └─598 /usr/lib/systemd/systemd-networkd │ ├─systemd-resolved.service │ │ └─464 /usr/lib/systemd/systemd-resolved │ ├─systemd-timesyncd.service │ │ └─465 /usr/lib/systemd/systemd-timesyncd │ ├─systemd-udevd.service │ │ └─udev │ │ └─453 /usr/lib/systemd/systemd-udevd │ ├─virtlockd.service │ │ └─42437 /usr/sbin/virtlockd │ └─virtlogd.service │ └─48046 /usr/sbin/virtlogd └─user.slice └─user-1000.slice ├─session-1.scope │ ├─ 828 "sshd: zuul [priv]" │ ├─ 853 "sshd: zuul@notty" │ ├─ 1054 /usr/bin/python3 │ ├─130107 sh -c "/bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '\"'\"'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3'\"'\"' && sleep 0'" │ ├─130108 /bin/sh -c "sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3' && sleep 0" │ ├─130109 sudo -H -S -n -u root /bin/sh -c "echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3" │ ├─130110 /bin/sh -c "echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3" │ ├─130111 /usr/bin/python3 │ ├─130112 /bin/bash -c "sudo iptables-save > /home/zuul/iptables.txt\n\n# NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from\n# stale NFS mounts.\ntimeout -s 9 60s df -h > /home/zuul/df.txt || true\n# If 'df' times out, the mount output helps debug which NFS share\n# is unresponsive.\nmount > /home/zuul/mount.txt\n\nfor py_ver in 2 3; do\n if [[ \`which python\${py_ver}\` ]]; then\n python\${py_ver} -m pip freeze > /home/zuul/pip\${py_ver}-freeze.txt\n fi\ndone\n\nif [ \`command -v dpkg\` ]; then\n dpkg -l> /home/zuul/dpkg-l.txt\nfi\nif [ \`command -v rpm\` ]; then\n rpm -qa | sort > /home/zuul/rpm-qa.txt\nfi\n\n# Services status\nsudo systemctl status --all > services.txt 2>/dev/null\n\n# NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU\n# failed to start due to denials from SELinux — useful for CentOS\n# and Fedora machines. For Ubuntu (which runs AppArmor), DevStack\n# already captures the contents of /var/log/kern.log (via\n# \`journalctl -t kernel\` redirected into syslog.txt.gz), which\n# contains AppArmor-related messages.\nif [ -f /var/log/audit/audit.log ] ; then\n sudo cp /var/log/audit/audit.log /home/zuul/audit.log &&\n chmod +r /home/zuul/audit.log;\nfi\n\n# gzip and save any coredumps in /var/core\nif [ -d /var/core ]; then\n sudo gzip -r /var/core\n sudo cp -r /var/core /home/zuul/\nfi\n\nsudo ss -lntup | grep ':53' > /home/zuul/listen53.txt\n\n# NOTE(andreaf) Service logs are already in logs/ thanks for the\n# export-devstack-journal log. Apache logs are under apache/ thans to the\n# apache-logs-conf role.\ngrep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}\\.[0-9]{1,3}/ /g' | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}/ /g' | \\\n sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' |\n sed -r 's/\\[.*\\]/ /g' | \\\n sed -r 's/\\s[0-9]+\\s/ /g' | \\\n awk '{if (\$0 in seen) {seen[\$0]++} else {out[++n]=\$0;seen[\$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]\" :: \" out[i] }' > /home/zuul/deprecations.log\n" │ ├─130125 sudo systemctl status --all │ └─130126 systemctl status --all └─user@1000.service └─init.scope ├─833 /usr/lib/systemd/systemd --user └─834 "(sd-pam)" ● proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point Loaded: loaded (/usr/lib/systemd/system/proc-sys-fs-binfmt_misc.automount; static) Active: active (running) since Thu 2026-04-02 14:21:03 UTC; 30min ago Triggers: ● proc-sys-fs-binfmt_misc.mount Where: /proc/sys/fs/binfmt_misc Docs: https://docs.kernel.org/admin-guide/binfmt-misc.html https://www.freedesktop.org/wiki/Software/systemd/APIFileSystems Apr 02 14:21:04 ubuntu systemd[1]: proc-sys-fs-binfmt_misc.automount: Got automount request for /proc/sys/fs/binfmt_misc, triggered by 451 (systemd-binfmt) Notice: journal has been rotated since unit was started, output may be incomplete. ● dev-cdrom.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2ddiskseq-11.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2ddiskseq-20.device - /dev/disk/by-diskseq/20 Follows: unit currently follows state of sys-devices-virtual-block-loop0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:41:37 UTC; 10min ago Device: /sys/devices/virtual/block/loop0 ● dev-disk-by\x2ddiskseq-9.device - /dev/disk/by-diskseq/9 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda ● dev-disk-by\x2ddiskseq-9\x2dpart1.device - /dev/disk/by-diskseq/9-part1 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda/vda1 ● dev-disk-by\x2did-ata\x2dQEMU_DVD\x2dROM_QM00001.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2did-lvm\x2dpv\x2duuid\x2daVQ2aM\x2dpREI\x2d4vUM\x2dxejG\x2d5auI\x2d5tpZ\x2dVFSToF.device - /dev/disk/by-id/lvm-pv-uuid-aVQ2aM-pREI-4vUM-xejG-5auI-5tpZ-VFSToF Follows: unit currently follows state of sys-devices-virtual-block-loop0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:41:37 UTC; 10min ago Device: /sys/devices/virtual/block/loop0 ● dev-disk-by\x2dlabel-cloudimg\x2drootfs.device - /dev/disk/by-label/cloudimg-rootfs Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda/vda1 ● dev-disk-by\x2dlabel-config\x2d2.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2dloop\x2dinode-253:1\x2d4697696.device - /dev/disk/by-loop-inode/253:1-4697696 Follows: unit currently follows state of sys-devices-virtual-block-loop0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:41:37 UTC; 10min ago Device: /sys/devices/virtual/block/loop0 ● dev-disk-by\x2dloop\x2dref-\x5cx2fopt\x5cx2fstack\x5cx2fdata\x5cx2fstack\x2dvolumes\x2dlvmdriver\x2d1\x2dbacking\x2dfile.device - /dev/disk/by-loop-ref/\x2fopt\x2fstack\x2fdata\x2fstack-volumes-lvmdriver-1-backing-file Follows: unit currently follows state of sys-devices-virtual-block-loop0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:41:37 UTC; 10min ago Device: /sys/devices/virtual/block/loop0 ● dev-disk-by\x2dpartuuid-4780bdb8\x2d01.device - /dev/disk/by-partuuid/4780bdb8-01 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda/vda1 ● dev-disk-by\x2dpath-pci\x2d0000:00:01.1\x2data\x2d1.0.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2dpath-pci\x2d0000:00:01.1\x2data\x2d1.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2dpath-pci\x2d0000:00:04.0.device - /dev/disk/by-path/pci-0000:00:04.0 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda ● dev-disk-by\x2dpath-pci\x2d0000:00:04.0\x2dpart1.device - /dev/disk/by-path/pci-0000:00:04.0-part1 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda/vda1 ● dev-disk-by\x2dpath-virtio\x2dpci\x2d0000:00:04.0.device - /dev/disk/by-path/virtio-pci-0000:00:04.0 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda ● dev-disk-by\x2dpath-virtio\x2dpci\x2d0000:00:04.0\x2dpart1.device - /dev/disk/by-path/virtio-pci-0000:00:04.0-part1 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda/vda1 ● dev-disk-by\x2duuid-2026\x2d04\x2d02\x2d14\x2d20\x2d57\x2d00.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-disk-by\x2duuid-2ea01307\x2d91a6\x2d4cac\x2da2e5\x2d756fcf145094.device - /dev/disk/by-uuid/2ea01307-91a6-4cac-a2e5-756fcf145094 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda/vda1 ● dev-dm\x2d0.device - /dev/dm-0 Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:46:29 UTC; 5min ago Device: /sys/devices/virtual/block/dm-0 ● dev-dm\x2d1.device - /dev/dm-1 Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d1.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:46:29 UTC; 5min ago Device: /sys/devices/virtual/block/dm-1 ● dev-dm\x2d2.device - /dev/dm-2 Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d2.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:46:29 UTC; 5min ago Device: /sys/devices/virtual/block/dm-2 ● dev-loop0.device - /dev/loop0 Follows: unit currently follows state of sys-devices-virtual-block-loop0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:41:37 UTC; 10min ago Device: /sys/devices/virtual/block/loop0 ● dev-mapper-stack\x2d\x2dvolumes\x2d\x2dlvmdriver\x2d\x2d1\x2dstack\x2d\x2dvolumes\x2d\x2dlvmdriver\x2d\x2d1\x2d\x2dpool.device - /dev/mapper/stack--volumes--lvmdriver--1-stack--volumes--lvmdriver--1--pool Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d2.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:46:29 UTC; 5min ago Device: /sys/devices/virtual/block/dm-2 ● dev-mapper-stack\x2d\x2dvolumes\x2d\x2dlvmdriver\x2d\x2d1\x2dstack\x2d\x2dvolumes\x2d\x2dlvmdriver\x2d\x2d1\x2d\x2dpool_tdata.device - /dev/mapper/stack--volumes--lvmdriver--1-stack--volumes--lvmdriver--1--pool_tdata Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d1.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:46:29 UTC; 5min ago Device: /sys/devices/virtual/block/dm-1 ● dev-mapper-stack\x2d\x2dvolumes\x2d\x2dlvmdriver\x2d\x2d1\x2dstack\x2d\x2dvolumes\x2d\x2dlvmdriver\x2d\x2d1\x2d\x2dpool_tmeta.device - /dev/mapper/stack--volumes--lvmdriver--1-stack--volumes--lvmdriver--1--pool_tmeta Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:46:29 UTC; 5min ago Device: /sys/devices/virtual/block/dm-0 ● dev-rfkill.device - /dev/rfkill Follows: unit currently follows state of sys-devices-virtual-misc-rfkill.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/virtual/misc/rfkill ● dev-sr0.device - QEMU_DVD-ROM config-2 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● dev-stack\x2dvolumes\x2dlvmdriver\x2d1-stack\x2dvolumes\x2dlvmdriver\x2d1\x2dpool.device - /dev/stack-volumes-lvmdriver-1/stack-volumes-lvmdriver-1-pool Follows: unit currently follows state of sys-devices-virtual-block-dm\x2d2.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:46:29 UTC; 5min ago Device: /sys/devices/virtual/block/dm-2 ● dev-ttyprintk.device - /dev/ttyprintk Follows: unit currently follows state of sys-devices-virtual-tty-ttyprintk.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/virtual/tty/ttyprintk ● dev-ttyS0.device - /dev/ttyS0 Follows: unit currently follows state of sys-devices-pnp0-00:00-00:00:0-00:00:0.0-tty-ttyS0.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pnp0/00:00/00:00:0/00:00:0.0/tty/ttyS0 Apr 02 14:21:04 ubuntu systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. ● dev-ttyS1.device - /dev/ttyS1 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.1-tty-ttyS1.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.1/tty/ttyS1 ● dev-ttyS10.device - /dev/ttyS10 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.10-tty-ttyS10.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.10/tty/ttyS10 ● dev-ttyS11.device - /dev/ttyS11 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.11-tty-ttyS11.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.11/tty/ttyS11 ● dev-ttyS12.device - /dev/ttyS12 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.12-tty-ttyS12.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.12/tty/ttyS12 ● dev-ttyS13.device - /dev/ttyS13 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.13-tty-ttyS13.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.13/tty/ttyS13 ● dev-ttyS14.device - /dev/ttyS14 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.14-tty-ttyS14.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.14/tty/ttyS14 ● dev-ttyS15.device - /dev/ttyS15 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.15-tty-ttyS15.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.15/tty/ttyS15 ● dev-ttyS16.device - /dev/ttyS16 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.16-tty-ttyS16.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.16/tty/ttyS16 ● dev-ttyS17.device - /dev/ttyS17 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.17-tty-ttyS17.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.17/tty/ttyS17 ● dev-ttyS18.device - /dev/ttyS18 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.18-tty-ttyS18.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.18/tty/ttyS18 ● dev-ttyS19.device - /dev/ttyS19 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.19-tty-ttyS19.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.19/tty/ttyS19 ● dev-ttyS2.device - /dev/ttyS2 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.2-tty-ttyS2.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.2/tty/ttyS2 ● dev-ttyS20.device - /dev/ttyS20 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.20-tty-ttyS20.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.20/tty/ttyS20 ● dev-ttyS21.device - /dev/ttyS21 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.21-tty-ttyS21.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.21/tty/ttyS21 ● dev-ttyS22.device - /dev/ttyS22 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.22-tty-ttyS22.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.22/tty/ttyS22 ● dev-ttyS23.device - /dev/ttyS23 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.23-tty-ttyS23.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.23/tty/ttyS23 ● dev-ttyS24.device - /dev/ttyS24 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.24-tty-ttyS24.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.24/tty/ttyS24 ● dev-ttyS25.device - /dev/ttyS25 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.25-tty-ttyS25.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.25/tty/ttyS25 ● dev-ttyS26.device - /dev/ttyS26 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.26-tty-ttyS26.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.26/tty/ttyS26 ● dev-ttyS27.device - /dev/ttyS27 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.27-tty-ttyS27.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.27/tty/ttyS27 ● dev-ttyS28.device - /dev/ttyS28 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.28-tty-ttyS28.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.28/tty/ttyS28 ● dev-ttyS29.device - /dev/ttyS29 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.29-tty-ttyS29.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.29/tty/ttyS29 ● dev-ttyS3.device - /dev/ttyS3 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.3-tty-ttyS3.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.3/tty/ttyS3 ● dev-ttyS30.device - /dev/ttyS30 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.30-tty-ttyS30.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.30/tty/ttyS30 ● dev-ttyS31.device - /dev/ttyS31 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.31-tty-ttyS31.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.31/tty/ttyS31 ● dev-ttyS4.device - /dev/ttyS4 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.4-tty-ttyS4.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.4/tty/ttyS4 ● dev-ttyS5.device - /dev/ttyS5 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.5-tty-ttyS5.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.5/tty/ttyS5 ● dev-ttyS6.device - /dev/ttyS6 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.6-tty-ttyS6.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.6/tty/ttyS6 ● dev-ttyS7.device - /dev/ttyS7 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.7-tty-ttyS7.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.7/tty/ttyS7 ● dev-ttyS8.device - /dev/ttyS8 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.8-tty-ttyS8.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.8/tty/ttyS8 ● dev-ttyS9.device - /dev/ttyS9 Follows: unit currently follows state of sys-devices-platform-serial8250-serial8250:0-serial8250:0.9-tty-ttyS9.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.9/tty/ttyS9 ● dev-vda.device - /dev/vda Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda ● dev-vda1.device - /dev/vda1 Follows: unit currently follows state of sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda-vda1.device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda/vda1 Notice: journal has been rotated since unit was started, output may be incomplete. ● sys-devices-pci0000:00-0000:00:01.1-ata1-host0-target0:0:0-0:0:0:0-block-sr0.device - QEMU_DVD-ROM config-2 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:01.1/ata1/host0/target0:0:0/0:0:0:0/block/sr0 ● sys-devices-pci0000:00-0000:00:03.0-virtio1-net-ens3.device - Virtio network device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:03.0/virtio1/net/ens3 ● sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda-vda1.device - /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda/vda1 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda/vda1 ● sys-devices-pci0000:00-0000:00:04.0-virtio2-block-vda.device - /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:04.0/virtio2/block/vda ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.1-tty-ttyS1.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.1/tty/ttyS1 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.1/tty/ttyS1 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.10-tty-ttyS10.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.10/tty/ttyS10 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.10/tty/ttyS10 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.11-tty-ttyS11.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.11/tty/ttyS11 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.11/tty/ttyS11 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.12-tty-ttyS12.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.12/tty/ttyS12 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.12/tty/ttyS12 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.13-tty-ttyS13.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.13/tty/ttyS13 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.13/tty/ttyS13 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.14-tty-ttyS14.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.14/tty/ttyS14 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.14/tty/ttyS14 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.15-tty-ttyS15.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.15/tty/ttyS15 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.15/tty/ttyS15 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.16-tty-ttyS16.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.16/tty/ttyS16 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.16/tty/ttyS16 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.17-tty-ttyS17.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.17/tty/ttyS17 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.17/tty/ttyS17 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.18-tty-ttyS18.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.18/tty/ttyS18 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.18/tty/ttyS18 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.19-tty-ttyS19.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.19/tty/ttyS19 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.19/tty/ttyS19 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.2-tty-ttyS2.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.2/tty/ttyS2 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.2/tty/ttyS2 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.20-tty-ttyS20.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.20/tty/ttyS20 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.20/tty/ttyS20 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.21-tty-ttyS21.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.21/tty/ttyS21 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.21/tty/ttyS21 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.22-tty-ttyS22.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.22/tty/ttyS22 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.22/tty/ttyS22 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.23-tty-ttyS23.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.23/tty/ttyS23 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.23/tty/ttyS23 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.24-tty-ttyS24.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.24/tty/ttyS24 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.24/tty/ttyS24 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.25-tty-ttyS25.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.25/tty/ttyS25 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.25/tty/ttyS25 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.26-tty-ttyS26.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.26/tty/ttyS26 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.26/tty/ttyS26 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.27-tty-ttyS27.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.27/tty/ttyS27 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.27/tty/ttyS27 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.28-tty-ttyS28.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.28/tty/ttyS28 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.28/tty/ttyS28 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.29-tty-ttyS29.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.29/tty/ttyS29 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.29/tty/ttyS29 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.3-tty-ttyS3.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.3/tty/ttyS3 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.3/tty/ttyS3 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.30-tty-ttyS30.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.30/tty/ttyS30 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.30/tty/ttyS30 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.31-tty-ttyS31.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.31/tty/ttyS31 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.31/tty/ttyS31 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.4-tty-ttyS4.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.4/tty/ttyS4 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.4/tty/ttyS4 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.5-tty-ttyS5.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.5/tty/ttyS5 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.5/tty/ttyS5 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.6-tty-ttyS6.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.6/tty/ttyS6 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.6/tty/ttyS6 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.7-tty-ttyS7.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.7/tty/ttyS7 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.7/tty/ttyS7 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.8-tty-ttyS8.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.8/tty/ttyS8 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.8/tty/ttyS8 ● sys-devices-platform-serial8250-serial8250:0-serial8250:0.9-tty-ttyS9.device - /sys/devices/platform/serial8250/serial8250:0/serial8250:0.9/tty/ttyS9 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/platform/serial8250/serial8250:0/serial8250:0.9/tty/ttyS9 ● sys-devices-pnp0-00:00-00:00:0-00:00:0.0-tty-ttyS0.device - /sys/devices/pnp0/00:00/00:00:0/00:00:0.0/tty/ttyS0 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pnp0/00:00/00:00:0/00:00:0.0/tty/ttyS0 ● sys-devices-virtual-block-dm\x2d0.device - /sys/devices/virtual/block/dm-0 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:46:29 UTC; 5min ago Device: /sys/devices/virtual/block/dm-0 ● sys-devices-virtual-block-dm\x2d1.device - /sys/devices/virtual/block/dm-1 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:46:29 UTC; 5min ago Device: /sys/devices/virtual/block/dm-1 ● sys-devices-virtual-block-dm\x2d2.device - /sys/devices/virtual/block/dm-2 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:46:29 UTC; 5min ago Device: /sys/devices/virtual/block/dm-2 ● sys-devices-virtual-block-loop0.device - /sys/devices/virtual/block/loop0 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:41:37 UTC; 10min ago Device: /sys/devices/virtual/block/loop0 ● sys-devices-virtual-misc-rfkill.device - /sys/devices/virtual/misc/rfkill Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/virtual/misc/rfkill ● sys-devices-virtual-net-br\x2d94c6ce52c320.device - /sys/devices/virtual/net/br-94c6ce52c320 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:29:47 UTC; 22min ago Device: /sys/devices/virtual/net/br-94c6ce52c320 ● sys-devices-virtual-net-br\x2dex.device - /sys/devices/virtual/net/br-ex Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:45:13 UTC; 6min ago Device: /sys/devices/virtual/net/br-ex ● sys-devices-virtual-net-br\x2dint.device - /sys/devices/virtual/net/br-int Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:34:16 UTC; 17min ago Device: /sys/devices/virtual/net/br-int ● sys-devices-virtual-net-docker0.device - /sys/devices/virtual/net/docker0 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:29:34 UTC; 22min ago Device: /sys/devices/virtual/net/docker0 ● sys-devices-virtual-net-ovs\x2dsystem.device - /sys/devices/virtual/net/ovs-system Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:34:16 UTC; 17min ago Device: /sys/devices/virtual/net/ovs-system ● sys-devices-virtual-net-tap6fccf94a\x2d80.device - /sys/devices/virtual/net/tap6fccf94a-80 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:50:42 UTC; 1min 11s ago Device: /sys/devices/virtual/net/tap6fccf94a-80 ● sys-devices-virtual-net-tapec2ec1a6\x2d67.device - /sys/devices/virtual/net/tapec2ec1a6-67 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:50:40 UTC; 1min 13s ago Device: /sys/devices/virtual/net/tapec2ec1a6-67 ● sys-devices-virtual-net-vethd9f9d99.device - /sys/devices/virtual/net/vethd9f9d99 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:29:56 UTC; 21min ago Device: /sys/devices/virtual/net/vethd9f9d99 ● sys-devices-virtual-net-virbr0.device - /sys/devices/virtual/net/virbr0 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:34:41 UTC; 17min ago Device: /sys/devices/virtual/net/virbr0 ● sys-devices-virtual-tty-ttyprintk.device - /sys/devices/virtual/tty/ttyprintk Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/virtual/tty/ttyprintk ● sys-module-configfs.device - /sys/module/configfs Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/module/configfs ● sys-module-fuse.device - /sys/module/fuse Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/module/fuse ● sys-subsystem-net-devices-br\x2d94c6ce52c320.device - /sys/subsystem/net/devices/br-94c6ce52c320 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:29:47 UTC; 22min ago Device: /sys/devices/virtual/net/br-94c6ce52c320 ● sys-subsystem-net-devices-br\x2dex.device - /sys/subsystem/net/devices/br-ex Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:45:13 UTC; 6min ago Device: /sys/devices/virtual/net/br-ex ● sys-subsystem-net-devices-br\x2dint.device - /sys/subsystem/net/devices/br-int Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:34:16 UTC; 17min ago Device: /sys/devices/virtual/net/br-int ● sys-subsystem-net-devices-docker0.device - /sys/subsystem/net/devices/docker0 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:29:34 UTC; 22min ago Device: /sys/devices/virtual/net/docker0 ● sys-subsystem-net-devices-ens3.device - Virtio network device Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:21:04 UTC; 30min ago Device: /sys/devices/pci0000:00/0000:00:03.0/virtio1/net/ens3 ● sys-subsystem-net-devices-ovs\x2dsystem.device - /sys/subsystem/net/devices/ovs-system Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:34:16 UTC; 17min ago Device: /sys/devices/virtual/net/ovs-system ● sys-subsystem-net-devices-tap6fccf94a\x2d80.device - /sys/subsystem/net/devices/tap6fccf94a-80 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:50:42 UTC; 1min 11s ago Device: /sys/devices/virtual/net/tap6fccf94a-80 ● sys-subsystem-net-devices-tapec2ec1a6\x2d67.device - /sys/subsystem/net/devices/tapec2ec1a6-67 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:50:40 UTC; 1min 13s ago Device: /sys/devices/virtual/net/tapec2ec1a6-67 ● sys-subsystem-net-devices-vethd9f9d99.device - /sys/subsystem/net/devices/vethd9f9d99 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:29:56 UTC; 21min ago Device: /sys/devices/virtual/net/vethd9f9d99 ● sys-subsystem-net-devices-virbr0.device - /sys/subsystem/net/devices/virbr0 Loaded: loaded Active: active (plugged) since Thu 2026-04-02 14:34:41 UTC; 17min ago Device: /sys/devices/virtual/net/virbr0 ● -.mount - Root Mount Loaded: loaded (/etc/fstab; generated) Active: active (mounted) since Thu 2026-04-02 14:21:03 UTC; 30min ago Where: / What: /dev/vda1 Docs: man:fstab(5) man:systemd-fstab-generator(8) Notice: journal has been rotated since unit was started, output may be incomplete. ● dev-hugepages.mount - Huge Pages File System Loaded: loaded (/proc/self/mountinfo; static) Active: active (mounted) since Thu 2026-04-02 14:21:04 UTC; 30min ago Where: /dev/hugepages What: hugetlbfs Docs: https://docs.kernel.org/admin-guide/mm/hugetlbpage.html https://www.freedesktop.org/wiki/Software/systemd/APIFileSystems Tasks: 0 (limit: 77075) Memory: 64.0K (peak: 592.0K) CPU: 3ms CGroup: /dev-hugepages.mount Apr 02 14:21:04 ubuntu systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Notice: journal has been rotated since unit was started, output may be incomplete. ● dev-mqueue.mount - POSIX Message Queue File System Loaded: loaded (/proc/self/mountinfo; static) Active: active (mounted) since Thu 2026-04-02 14:21:04 UTC; 30min ago Where: /dev/mqueue What: mqueue Docs: man:mq_overview(7) https://www.freedesktop.org/wiki/Software/systemd/APIFileSystems Tasks: 0 (limit: 77075) Memory: 4.0K (peak: 1.7M) CPU: 10ms CGroup: /dev-mqueue.mount Apr 02 14:21:04 ubuntu systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Notice: journal has been rotated since unit was started, output may be incomplete. ● opt-stack-data-etcd.mount - /opt/stack/data/etcd Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Thu 2026-04-02 14:32:10 UTC; 19min ago Where: /opt/stack/data/etcd What: tmpfs ● proc-fs-nfsd.mount - NFSD configuration filesystem Loaded: loaded (/usr/lib/systemd/system/proc-fs-nfsd.mount; static) Active: active (mounted) since Thu 2026-04-02 14:36:36 UTC; 15min ago Where: /proc/fs/nfsd What: nfsd Tasks: 0 (limit: 77075) Memory: 20.0K (peak: 1.5M) CPU: 11ms CGroup: /proc-fs-nfsd.mount Apr 02 14:36:36 np0000163014 systemd[1]: Mounting proc-fs-nfsd.mount - NFSD configuration filesystem... Apr 02 14:36:36 np0000163014 systemd[1]: Mounted proc-fs-nfsd.mount - NFSD configuration filesystem. ● proc-sys-fs-binfmt_misc.mount - Arbitrary Executable File Formats File System Loaded: loaded (/proc/self/mountinfo; disabled; preset: disabled) Active: active (mounted) since Thu 2026-04-02 14:21:04 UTC; 30min ago TriggeredBy: ● proc-sys-fs-binfmt_misc.automount Where: /proc/sys/fs/binfmt_misc What: binfmt_misc Docs: https://docs.kernel.org/admin-guide/binfmt-misc.html https://www.freedesktop.org/wiki/Software/systemd/APIFileSystems Tasks: 0 (limit: 77075) Memory: 8.0K (peak: 1.5M) CPU: 14ms CGroup: /proc-sys-fs-binfmt_misc.mount Apr 02 14:21:04 ubuntu systemd[1]: Mounting proc-sys-fs-binfmt_misc.mount - Arbitrary Executable File Formats File System... Apr 02 14:21:04 ubuntu systemd[1]: Mounted proc-sys-fs-binfmt_misc.mount - Arbitrary Executable File Formats File System. ● run-docker-netns-de88040ede10.mount - /run/docker/netns/de88040ede10 Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Thu 2026-04-02 14:29:56 UTC; 21min ago Where: /run/docker/netns/de88040ede10 What: nsfs ● run-netns-ovnmeta\x2d6fccf94a\x2d88ef\x2d49fd\x2da0c3\x2d4d8e39aeea0c.mount - /run/netns/ovnmeta-6fccf94a-88ef-49fd-a0c3-4d8e39aeea0c Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Thu 2026-04-02 14:50:40 UTC; 1min 13s ago Where: /run/netns/ovnmeta-6fccf94a-88ef-49fd-a0c3-4d8e39aeea0c What: nsfs ● run-netns.mount - /run/netns Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Thu 2026-04-02 14:50:40 UTC; 1min 13s ago Where: /run/netns What: tmpfs ● run-rpc_pipefs.mount - RPC Pipe File System Loaded: loaded (/run/systemd/generator/run-rpc_pipefs.mount; generated) Active: active (mounted) since Thu 2026-04-02 14:36:35 UTC; 15min ago Where: /run/rpc_pipefs What: sunrpc Tasks: 0 (limit: 77075) Memory: 20.0K (peak: 552.0K) CPU: 5ms CGroup: /system.slice/run-rpc_pipefs.mount Apr 02 14:36:35 np0000163014 systemd[1]: Mounting run-rpc_pipefs.mount - RPC Pipe File System... Apr 02 14:36:35 np0000163014 systemd[1]: Mounted run-rpc_pipefs.mount - RPC Pipe File System. ● run-user-1000.mount - /run/user/1000 Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Thu 2026-04-02 14:21:59 UTC; 29min ago Where: /run/user/1000 What: tmpfs ● sys-fs-fuse-connections.mount - FUSE Control File System Loaded: loaded (/proc/self/mountinfo; static) Active: active (mounted) since Thu 2026-04-02 14:21:04 UTC; 30min ago Where: /sys/fs/fuse/connections What: fusectl Docs: https://docs.kernel.org/filesystems/fuse.html https://www.freedesktop.org/wiki/Software/systemd/APIFileSystems Tasks: 0 (limit: 77075) Memory: 4.0K (peak: 1.5M) CPU: 11ms CGroup: /sys-fs-fuse-connections.mount Apr 02 14:21:04 ubuntu systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 02 14:21:04 ubuntu systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. ● sys-kernel-config.mount - Kernel Configuration File System Loaded: loaded (/proc/self/mountinfo; static) Active: active (mounted) since Thu 2026-04-02 14:21:04 UTC; 30min ago Where: /sys/kernel/config What: configfs Docs: https://docs.kernel.org/filesystems/configfs.html https://www.freedesktop.org/wiki/Software/systemd/APIFileSystems Tasks: 0 (limit: 77075) Memory: 8.0K (peak: 1.7M) CPU: 12ms CGroup: /sys-kernel-config.mount Apr 02 14:21:04 ubuntu systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 02 14:21:04 ubuntu systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. ● sys-kernel-debug.mount - Kernel Debug File System Loaded: loaded (/proc/self/mountinfo; static) Active: active (mounted) since Thu 2026-04-02 14:21:04 UTC; 30min ago Where: /sys/kernel/debug What: debugfs Docs: https://docs.kernel.org/filesystems/debugfs.html https://www.freedesktop.org/wiki/Software/systemd/APIFileSystems Tasks: 0 (limit: 77075) Memory: 4.0K (peak: 1.5M) CPU: 12ms CGroup: /sys-kernel-debug.mount Apr 02 14:21:04 ubuntu systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Notice: journal has been rotated since unit was started, output may be incomplete. ● sys-kernel-tracing.mount - Kernel Trace File System Loaded: loaded (/proc/self/mountinfo; static) Active: active (mounted) since Thu 2026-04-02 14:21:04 UTC; 30min ago Where: /sys/kernel/tracing What: tracefs Docs: https://docs.kernel.org/trace/ftrace.html https://www.freedesktop.org/wiki/Software/systemd/APIFileSystems Tasks: 0 (limit: 77075) Memory: 4.0K (peak: 1.5M) CPU: 12ms CGroup: /sys-kernel-tracing.mount Apr 02 14:21:04 ubuntu systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Notice: journal has been rotated since unit was started, output may be incomplete. ● var-lib-docker-rootfs-overlayfs-54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8.mount - /var/lib/docker/rootfs/overlayfs/54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8 Loaded: loaded (/proc/self/mountinfo) Active: active (mounted) since Thu 2026-04-02 14:29:56 UTC; 21min ago Where: /var/lib/docker/rootfs/overlayfs/54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8 What: overlay ○ var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) Loaded: loaded (/usr/lib/systemd/system/var-lib-machines.mount; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:34:41 UTC; 17min ago Where: /var/lib/machines What: /var/lib/machines.raw Apr 02 14:34:37 np0000163014 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 02 14:34:41 np0000163014 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). ● systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch Loaded: loaded (/usr/lib/systemd/system/systemd-ask-password-console.path; static) Active: active (waiting) since Thu 2026-04-02 14:21:03 UTC; 30min ago Triggers: ● systemd-ask-password-console.service Docs: man:systemd-ask-password-console.path(8) Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch Loaded: loaded (/usr/lib/systemd/system/systemd-ask-password-wall.path; static) Active: active (waiting) since Thu 2026-04-02 14:21:03 UTC; 30min ago Triggers: ● systemd-ask-password-wall.service Docs: man:systemd-ask-password-wall.path(8) Notice: journal has been rotated since unit was started, output may be incomplete. ● docker-54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8.scope - libcontainer container 54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8 Loaded: loaded (/run/systemd/transient/docker-54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8.scope; transient) Transient: yes Drop-In: /run/systemd/transient/docker-54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8.scope.d └─50-DeviceAllow.conf, 50-DevicePolicy.conf Active: active (running) since Thu 2026-04-02 14:29:56 UTC; 21min ago IO: 400.0K read, 1.5G written Tasks: 603 (limit: 77075) Memory: 2.3G (peak: 2.4G) CPU: 7min 18.866s CGroup: /system.slice/docker-54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8.scope ├─init.scope │ └─21135 /sbin/init ├─kubelet.slice │ ├─kubelet-kubepods.slice │ │ ├─kubelet-kubepods-besteffort.slice │ │ │ ├─kubelet-kubepods-besteffort-pod0252a4bd_02fc_4e45_b96b_87fd60aae118.slice │ │ │ │ ├─cri-containerd-38b467f0d9256afa3b38e32675958de195df595af5b33c93ef4ad091fd7ca2ce.scope │ │ │ │ │ └─24819 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,KubeadmBootstrapFormatIgnition=true,PriorityQueue=false --bootstrap-token-ttl=15m │ │ │ │ └─cri-containerd-cab203df0fa479277115976d13436c61979f71791765859f21b6ef4d545e0a28.scope │ │ │ │ └─24432 /pause │ │ │ ├─kubelet-kubepods-besteffort-pod0b3e7747_7f3a_4a37_880c_7609beb2bec2.slice │ │ │ │ ├─cri-containerd-0d1c8a0dc4444ab42de6f1edc1a3d7f9d15e8257c27b683b19c06074509a75fe.scope │ │ │ │ │ └─25117 /manager --leader-elect --v=2 --diagnostics-address=127.0.0.1:8080 --insecure-diagnostics=true │ │ │ │ └─cri-containerd-feb0b8b6bb17ac3fdd5946e4a75f306d531c347d9923d5f36691d6d272a7bb3f.scope │ │ │ │ └─24759 /pause │ │ │ ├─kubelet-kubepods-besteffort-pod19708082_a0bb_4e1a_8b9f_8b55bade2a97.slice │ │ │ │ ├─cri-containerd-267658bf363448825ad8effc3144e2a05b7e99f48f9ef266949239dadf0e647f.scope │ │ │ │ │ └─24642 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,ClusterResourceSet=true,ClusterTopology=true,RuntimeSDK=false,MachineSetPreflightChecks=true,MachineWaitForVolumeDetachConsiderVolumeAttachments=true,PriorityQueue=false │ │ │ │ └─cri-containerd-f15e92eac3c699ac4091b22d11aebac4ddda90272e826d9430f8393eb08e6290.scope │ │ │ │ └─24363 /pause │ │ │ ├─kubelet-kubepods-besteffort-pod2b65498f_1697_43c6_a06c_b95cb04050e9.slice │ │ │ │ ├─cri-containerd-61bac5c5b3476d4d45e83534a006e7e92b17979efcd672630913702862a801da.scope │ │ │ │ │ └─23224 /pause │ │ │ │ └─cri-containerd-7f8394ff1f96657dbbd8821483c8bd4d1c2cdd2a378ea35fe8be66375af9edd2.scope │ │ │ │ └─23605 local-path-provisioner --debug start --helper-image docker.io/kindest/local-path-helper:v20220607-9a4d8d2a --config /etc/config/config.json │ │ │ ├─kubelet-kubepods-besteffort-pod3c159520_1948_489d_96ad_e4753123a872.slice │ │ │ │ ├─cri-containerd-c43573aa54912bff9c0d74b6ef47458ab4fbfcd9e25f44f5250dba0ace166133.scope │ │ │ │ │ └─24161 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=kube-system │ │ │ │ └─cri-containerd-ccc37416b9aa55b0d540fe8325a804eb567d99c4e67550ef009f367d5d735c4a.scope │ │ │ │ └─23568 /pause │ │ │ ├─kubelet-kubepods-besteffort-podcfd58e14_dd69_4f38_9746_b9c5983f57b6.slice │ │ │ │ ├─cri-containerd-2c4b62468f8906532289d834d3d20d746566a61df2d31dc105f87c9c243ba045.scope │ │ │ │ │ └─23866 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --acme-http01-solver-image=quay.io/jetstack/cert-manager-acmesolver:v1.18.1 --max-concurrent-challenges=60 │ │ │ │ └─cri-containerd-fd50372de23838501cfd1f9593f95c913e7d080f10f61d6628f9f6f454cca9d5.scope │ │ │ │ └─23400 /pause │ │ │ ├─kubelet-kubepods-besteffort-podd87bcd89_9c5c_4cfb_a125_66fbb1d9c47d.slice │ │ │ │ ├─cri-containerd-88e90d76e68188293a25b8b8c51be0f3ccf15c5f821aa794f1b0cefabe0a94a3.scope │ │ │ │ │ └─22896 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=kind-control-plane │ │ │ │ └─cri-containerd-ff188e61a32f5f07cc3bf88092974b3b46868bbc83fdde14a2d301c6228e65e7.scope │ │ │ │ └─22841 /pause │ │ │ ├─kubelet-kubepods-besteffort-pode2c03ea4_9bb3_4184_b5e4_4e142c6fb653.slice │ │ │ │ ├─cri-containerd-0e3f59113c1c6e6512e47da5dbba7302a32112b2f9c352d5e9f73f4ab1595ca9.scope │ │ │ │ │ └─23560 /pause │ │ │ │ └─cri-containerd-68ceb73a5cea74ca45f26405a3bd47783a761566d102a7097661496ba37a5647.scope │ │ │ │ └─24073 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc │ │ │ └─kubelet-kubepods-besteffort-podf874c790_24e9_4b31_a814_05df2252c78f.slice │ │ │ ├─cri-containerd-be887502786ee7935d2add369d9825d38b897c67c3b50c15b89b6710ecae3e6c.scope │ │ │ │ └─24579 /pause │ │ │ └─cri-containerd-d1af59187bb078e68792bc9bfcee67bd66f3249aee55bbf56c703b117ecff25e.scope │ │ │ └─24953 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,ClusterTopology=true,KubeadmBootstrapFormatIgnition=true,PriorityQueue=false │ │ ├─kubelet-kubepods-burstable.slice │ │ │ ├─kubelet-kubepods-burstable-pod0656ab70da313d6449b17f099a2a3110.slice │ │ │ │ ├─cri-containerd-2ab87108ea1e6b8b64d509ef7c1c9342d77a7d272ccb9d4240e562e03b234ab1.scope │ │ │ │ │ └─22131 etcd --advertise-client-urls=https://172.18.0.2:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://172.18.0.2:2380 --initial-cluster=kind-control-plane=https://172.18.0.2:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://172.18.0.2:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://172.18.0.2:2380 --name=kind-control-plane --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ │ │ │ └─cri-containerd-6208997ab8fc91afafd91db67859c3b91025856a1f308a614267a5a3d0c37366.scope │ │ │ │ └─21804 /pause │ │ │ ├─kubelet-kubepods-burstable-pod53ff6c8abd472f64bc9a9afbd3a471a9.slice │ │ │ │ ├─cri-containerd-870df6abbc20c8d337ecc4c2bce2a50cd533c2789b8639678ed04c6f1d760e4c.scope │ │ │ │ │ └─21917 /pause │ │ │ │ └─cri-containerd-e8f0ee1d61dfec222520c595b1deb5719cdeebd6ece0c11d1ae8d92037973a09.scope │ │ │ │ └─22041 kube-controller-manager --allocate-node-cidrs=true --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=127.0.0.1 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-cidr=10.244.0.0/16 --cluster-name=kind --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key "--controllers=*,bootstrapsigner,tokencleaner" --enable-hostpath-provisioner=true --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/16 --use-service-account-credentials=true │ │ │ ├─kubelet-kubepods-burstable-pod7afc15c3_ead2_46e7_9a64_97d9b2aeff9e.slice │ │ │ │ ├─cri-containerd-19e68a60cb0b1f7aaff877e9aa2ce9e215bb133217cb6c6bec51e589245f9657.scope │ │ │ │ │ └─23978 /manager --metrics-bind-address=:8443 --leader-elect --health-probe-bind-address=:8081 │ │ │ │ └─cri-containerd-24a89f3f6c95c059079539dd5111752929e7c7739c4f76340c3501a5beba8343.scope │ │ │ │ └─23468 /pause │ │ │ ├─kubelet-kubepods-burstable-podaf8e1906_1a5b_4f97_902c_ca7d36e47926.slice │ │ │ │ ├─cri-containerd-48b1c569323512f743b5d77a02134ea29982e2225d3d1e74fb3f1f47fc7e0b24.scope │ │ │ │ │ └─23654 /coredns -conf /etc/coredns/Corefile │ │ │ │ └─cri-containerd-6472d35a6a9b786a04324c9f49d7f145bd3246b78321220d11fae0c5e43970c5.scope │ │ │ │ └─23217 /pause │ │ │ ├─kubelet-kubepods-burstable-podbee69ab63b6471d4da666ee970746eae.slice │ │ │ │ ├─cri-containerd-31d336081f7b38682b523a2347ccf8f412fd51b43947bcb6cc3201680c38d863.scope │ │ │ │ │ └─21975 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=127.0.0.1 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true │ │ │ │ └─cri-containerd-7c89215ac2295c29f7bd0d888c165a63c94d11eeb8c6a7292afd33a2f3cdd899.scope │ │ │ │ └─21907 /pause │ │ │ ├─kubelet-kubepods-burstable-podd25193d9_141b_4581_a1b0_bb10a32dc523.slice │ │ │ │ ├─cri-containerd-c21138d45276460b2a29141bd3bb697e08f97fec20a617b33be15c71bc61f54e.scope │ │ │ │ │ └─23679 /coredns -conf /etc/coredns/Corefile │ │ │ │ └─cri-containerd-f459ee43bf1a88a56ea2d80d0152986c09363cfb35460b03ee8f3a0f4424f49c.scope │ │ │ │ └─23475 /pause │ │ │ └─kubelet-kubepods-burstable-podef6ebc9842be361e05ebdb6790c540b6.slice │ │ │ ├─cri-containerd-1926238e4fdf69a50346ffe73b66ea67b7f75760bf715fa5fff86efedc94f171.scope │ │ │ │ └─21844 /pause │ │ │ └─cri-containerd-3c7a2ea72ddf5756767d9d3c9628ca18ef585748147673c7e3ac210cdce69837.scope │ │ │ └─22017 kube-apiserver --advertise-address=172.18.0.2 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --runtime-config= --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/16 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ │ └─kubelet-kubepods-pod30514ae6_c3de_43c7_85f3_bc4f3ac7154d.slice │ │ ├─cri-containerd-522d017c8e1da3feb0777e65b66d2de23ebc9a86b25a0adc7a5587008b794a64.scope │ │ │ └─22935 /bin/kindnetd │ │ └─cri-containerd-d01f22860fc3f4da6c86ca2a85f5326a74793725f56e3e7216be56ae80ef1f44.scope │ │ └─22849 /pause │ └─kubelet.service │ └─22229 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --container-runtime=remote --container-runtime-endpoint=unix:///run/containerd/containerd.sock --node-ip=172.18.0.2 --node-labels= --pod-infra-container-image=registry.k8s.io/pause:3.8 --provider-id=kind://docker/kind/kind-control-plane --fail-swap-on=false --cgroup-root=/kubelet └─system.slice ├─containerd.service │ ├─21331 /usr/local/bin/containerd │ ├─21782 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6208997ab8fc91afafd91db67859c3b91025856a1f308a614267a5a3d0c37366 -address /run/containerd/containerd.sock │ ├─21818 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1926238e4fdf69a50346ffe73b66ea67b7f75760bf715fa5fff86efedc94f171 -address /run/containerd/containerd.sock │ ├─21863 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7c89215ac2295c29f7bd0d888c165a63c94d11eeb8c6a7292afd33a2f3cdd899 -address /run/containerd/containerd.sock │ ├─21890 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 870df6abbc20c8d337ecc4c2bce2a50cd533c2789b8639678ed04c6f1d760e4c -address /run/containerd/containerd.sock │ ├─22800 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id ff188e61a32f5f07cc3bf88092974b3b46868bbc83fdde14a2d301c6228e65e7 -address /run/containerd/containerd.sock │ ├─22808 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id d01f22860fc3f4da6c86ca2a85f5326a74793725f56e3e7216be56ae80ef1f44 -address /run/containerd/containerd.sock │ ├─23169 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6472d35a6a9b786a04324c9f49d7f145bd3246b78321220d11fae0c5e43970c5 -address /run/containerd/containerd.sock │ ├─23188 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 61bac5c5b3476d4d45e83534a006e7e92b17979efcd672630913702862a801da -address /run/containerd/containerd.sock │ ├─23380 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id fd50372de23838501cfd1f9593f95c913e7d080f10f61d6628f9f6f454cca9d5 -address /run/containerd/containerd.sock │ ├─23427 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 24a89f3f6c95c059079539dd5111752929e7c7739c4f76340c3501a5beba8343 -address /run/containerd/containerd.sock │ ├─23435 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id f459ee43bf1a88a56ea2d80d0152986c09363cfb35460b03ee8f3a0f4424f49c -address /run/containerd/containerd.sock │ ├─23506 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0e3f59113c1c6e6512e47da5dbba7302a32112b2f9c352d5e9f73f4ab1595ca9 -address /run/containerd/containerd.sock │ ├─23523 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id ccc37416b9aa55b0d540fe8325a804eb567d99c4e67550ef009f367d5d735c4a -address /run/containerd/containerd.sock │ ├─24342 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id f15e92eac3c699ac4091b22d11aebac4ddda90272e826d9430f8393eb08e6290 -address /run/containerd/containerd.sock │ ├─24412 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id cab203df0fa479277115976d13436c61979f71791765859f21b6ef4d545e0a28 -address /run/containerd/containerd.sock │ ├─24559 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id be887502786ee7935d2add369d9825d38b897c67c3b50c15b89b6710ecae3e6c -address /run/containerd/containerd.sock │ └─24739 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id feb0b8b6bb17ac3fdd5946e4a75f306d531c347d9923d5f36691d6d272a7bb3f -address /run/containerd/containerd.sock └─systemd-journald.service └─21316 /lib/systemd/systemd-journald Apr 02 14:29:56 np0000163014 systemd[1]: Started docker-54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8.scope - libcontainer container 54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8. ● init.scope - System and Service Manager Loaded: loaded Transient: yes Active: active (running) since Thu 2026-04-02 14:21:03 UTC; 30min ago Docs: man:systemd(1) Tasks: 1 (limit: 77075) Memory: 10.5M (peak: 24.5M) CPU: 1min 13.311s CGroup: /init.scope └─1 /sbin/init nofb Apr 02 14:50:39 np0000163014 systemd[1]: Reloading... Apr 02 14:50:39 np0000163014 systemd[1]: Reloading finished in 321 ms. Apr 02 14:50:39 np0000163014 systemd[1]: Started devstack@m-sch.service - Devstack devstack@m-sch.service. Apr 02 14:50:41 np0000163014 systemd[1]: Reloading requested from client PID 127127 ('systemctl') (unit session-1.scope)... Apr 02 14:50:41 np0000163014 systemd[1]: Reloading... Apr 02 14:50:41 np0000163014 systemd[1]: Reloading finished in 294 ms. Apr 02 14:50:41 np0000163014 systemd[1]: Reloading requested from client PID 127219 ('systemctl') (unit session-1.scope)... Apr 02 14:50:41 np0000163014 systemd[1]: Reloading... Apr 02 14:50:41 np0000163014 systemd[1]: Reloading finished in 302 ms. Apr 02 14:50:42 np0000163014 systemd[1]: Started devstack@m-dat.service - Devstack devstack@m-dat.service. ● session-1.scope - Session 1 of User zuul Loaded: loaded (/run/systemd/transient/session-1.scope; transient) Transient: yes Active: active (running) since Thu 2026-04-02 14:22:00 UTC; 29min ago Tasks: 13 Memory: 20.7G (peak: 20.8G) CPU: 23min 15.200s CGroup: /user.slice/user-1000.slice/session-1.scope ├─ 828 "sshd: zuul [priv]" ├─ 853 "sshd: zuul@notty" ├─ 1054 /usr/bin/python3 ├─130107 sh -c "/bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '\"'\"'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3'\"'\"' && sleep 0'" ├─130108 /bin/sh -c "sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3' && sleep 0" ├─130109 sudo -H -S -n -u root /bin/sh -c "echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3" ├─130110 /bin/sh -c "echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3" ├─130111 /usr/bin/python3 ├─130112 /bin/bash -c "sudo iptables-save > /home/zuul/iptables.txt\n\n# NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from\n# stale NFS mounts.\ntimeout -s 9 60s df -h > /home/zuul/df.txt || true\n# If 'df' times out, the mount output helps debug which NFS share\n# is unresponsive.\nmount > /home/zuul/mount.txt\n\nfor py_ver in 2 3; do\n if [[ \`which python\${py_ver}\` ]]; then\n python\${py_ver} -m pip freeze > /home/zuul/pip\${py_ver}-freeze.txt\n fi\ndone\n\nif [ \`command -v dpkg\` ]; then\n dpkg -l> /home/zuul/dpkg-l.txt\nfi\nif [ \`command -v rpm\` ]; then\n rpm -qa | sort > /home/zuul/rpm-qa.txt\nfi\n\n# Services status\nsudo systemctl status --all > services.txt 2>/dev/null\n\n# NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU\n# failed to start due to denials from SELinux — useful for CentOS\n# and Fedora machines. For Ubuntu (which runs AppArmor), DevStack\n# already captures the contents of /var/log/kern.log (via\n# \`journalctl -t kernel\` redirected into syslog.txt.gz), which\n# contains AppArmor-related messages.\nif [ -f /var/log/audit/audit.log ] ; then\n sudo cp /var/log/audit/audit.log /home/zuul/audit.log &&\n chmod +r /home/zuul/audit.log;\nfi\n\n# gzip and save any coredumps in /var/core\nif [ -d /var/core ]; then\n sudo gzip -r /var/core\n sudo cp -r /var/core /home/zuul/\nfi\n\nsudo ss -lntup | grep ':53' > /home/zuul/listen53.txt\n\n# NOTE(andreaf) Service logs are already in logs/ thanks for the\n# export-devstack-journal log. Apache logs are under apache/ thans to the\n# apache-logs-conf role.\ngrep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}\\.[0-9]{1,3}/ /g' | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}/ /g' | \\\n sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' |\n sed -r 's/\\[.*\\]/ /g' | \\\n sed -r 's/\\s[0-9]+\\s/ /g' | \\\n awk '{if (\$0 in seen) {seen[\$0]++} else {out[++n]=\$0;seen[\$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]\" :: \" out[i] }' > /home/zuul/deprecations.log\n" ├─130125 sudo systemctl status --all └─130126 systemctl status --all Apr 02 14:51:52 np0000163014 python3[130101]: ansible-ansible.legacy.command Invoked with _raw_params=cp -pRL /etc/openstack /home/zuul/etc/ zuul_no_log=False zuul_log_id=0242ac17-0011-42b5-8cef-00000000002f-1-controller zuul_output_max_bytes=1073741824 zuul_ansible_split_streams=False _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Apr 02 14:51:52 np0000163014 sudo[130099]: pam_unix(sudo:session): session closed for user root Apr 02 14:51:53 np0000163014 sudo[130109]: zuul : PWD=/home/zuul ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3' Apr 02 14:51:53 np0000163014 sudo[130109]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=1000) Apr 02 14:51:53 np0000163014 python3[130111]: ansible-ansible.legacy.command Invoked with executable=/bin/bash _raw_params=sudo iptables-save > /home/zuul/iptables.txt # NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from # stale NFS mounts. timeout -s 9 60s df -h > /home/zuul/df.txt || true # If 'df' times out, the mount output helps debug which NFS share # is unresponsive. mount > /home/zuul/mount.txt for py_ver in 2 3; do if [[ `which python${py_ver}` ]]; then python${py_ver} -m pip freeze > /home/zuul/pip${py_ver}-freeze.txt fi done if [ `command -v dpkg` ]; then dpkg -l> /home/zuul/dpkg-l.txt fi if [ `command -v rpm` ]; then rpm -qa | sort > /home/zuul/rpm-qa.txt fi # Services status sudo systemctl status --all > services.txt 2>/dev/null # NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU # failed to start due to denials from SELinux — useful for CentOS # and Fedora machines. For Ubuntu (which runs AppArmor), DevStack # already captures the contents of /var/log/kern.log (via # `journalctl -t kernel` redirected into syslog.txt.gz), which # contains AppArmor-related messages. if [ -f /var/log/audit/audit.log ] ; then sudo cp /var/log/audit/audit.log /home/zuul/audit.log && chmod +r /home/zuul/audit.log; fi # gzip and save any coredumps in /var/core if [ -d /var/core ]; then sudo gzip -r /var/core sudo cp -r /var/core /home/zuul/ fi sudo ss -lntup | grep ':53' > /home/zuul/listen53.txt # NOTE(andreaf) Service logs are already in logs/ thanks for the # export-devstack-journal log. Apache logs are under apache/ thans to the # apache-logs-conf role. grep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \ sed -r 's/[0-9]{1,2}\:[0-9]{1,2}\:[0-9]{1,2}\.[0-9]{1,3}/ /g' | \ sed -r 's/[0-9]{1,2}\:[0-9]{1,2}\:[0-9]{1,2}/ /g' | \ sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' | sed -r 's/\[.*\]/ /g' | \ sed -r 's/\s[0-9]+\s/ /g' | \ awk '{if ($0 in seen) {seen[$0]++} else {out[++n]=$0;seen[$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]" :: " out[i] }' > /home/zuul/deprecations.log _uses_shell=True zuul_no_log=False zuul_log_id=0242ac17-0011-42b5-8cef-000000000033-1-controller zuul_output_max_bytes=1073741824 zuul_ansible_split_streams=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None creates=None removes=None stdin=None Apr 02 14:51:53 np0000163014 sudo[130114]: root : PWD=/home/zuul ; USER=root ; COMMAND=/usr/sbin/iptables-save Apr 02 14:51:53 np0000163014 sudo[130114]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=0) Apr 02 14:51:53 np0000163014 sudo[130114]: pam_unix(sudo:session): session closed for user root Apr 02 14:51:53 np0000163014 sudo[130125]: root : PWD=/home/zuul ; USER=root ; COMMAND=/usr/bin/systemctl status --all Apr 02 14:51:53 np0000163014 sudo[130125]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=0) ● apache-htcacheclean.service - Disk Cache Cleaning Daemon for Apache HTTP Server Loaded: loaded (/usr/lib/systemd/system/apache-htcacheclean.service; disabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:28:22 UTC; 23min ago Docs: https://httpd.apache.org/docs/2.4/programs/htcacheclean.html Main PID: 13911 (htcacheclean) Tasks: 1 (limit: 77075) Memory: 296.0K (peak: 832.0K) CPU: 67ms CGroup: /system.slice/apache-htcacheclean.service └─13911 /usr/bin/htcacheclean -d 120 -p /var/cache/apache2/mod_cache_disk -l 300M -n Apr 02 14:28:22 np0000163014 systemd[1]: Starting apache-htcacheclean.service - Disk Cache Cleaning Daemon for Apache HTTP Server... Apr 02 14:28:22 np0000163014 systemd[1]: Started apache-htcacheclean.service - Disk Cache Cleaning Daemon for Apache HTTP Server. ● apache2.service - The Apache HTTP Server Loaded: loaded (/usr/lib/systemd/system/apache2.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:49:56 UTC; 1min 57s ago Docs: https://httpd.apache.org/docs/2.4/ Main PID: 121459 (apache2) Tasks: 69 (limit: 77075) Memory: 18.7M (peak: 20.0M) CPU: 1.505s CGroup: /system.slice/apache2.service ├─121459 /usr/sbin/apache2 -k start ├─121463 /usr/sbin/apache2 -k start └─121465 /usr/sbin/apache2 -k start Apr 02 14:49:56 np0000163014 systemd[1]: Starting apache2.service - The Apache HTTP Server... Apr 02 14:49:56 np0000163014 systemd[1]: Started apache2.service - The Apache HTTP Server. ○ apt-daily-upgrade.service - Daily apt upgrade and clean activities Loaded: loaded (/usr/lib/systemd/system/apt-daily-upgrade.service; static) Active: inactive (dead) TriggeredBy: ● apt-daily-upgrade.timer Docs: man:apt(8) ○ apt-daily.service - Daily apt download activities Loaded: loaded (/usr/lib/systemd/system/apt-daily.service; static) Active: inactive (dead) TriggeredBy: ● apt-daily.timer Docs: man:apt(8) ○ auth-rpcgss-module.service - Kernel Module supporting RPCSEC_GSS Loaded: loaded (/usr/lib/systemd/system/auth-rpcgss-module.service; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:36:38 UTC; 15min ago Apr 02 14:36:35 np0000163014 systemd[1]: auth-rpcgss-module.service - Kernel Module supporting RPCSEC_GSS was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). Apr 02 14:36:36 np0000163014 systemd[1]: auth-rpcgss-module.service - Kernel Module supporting RPCSEC_GSS was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). Apr 02 14:36:38 np0000163014 systemd[1]: auth-rpcgss-module.service - Kernel Module supporting RPCSEC_GSS was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). ● blk-availability.service - Availability of block devices Loaded: loaded (/usr/lib/systemd/system/blk-availability.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:28:25 UTC; 23min ago Apr 02 14:28:25 np0000163014 systemd[1]: Finished blk-availability.service - Availability of block devices. ● cloud-config.service - Cloud-init: Config Stage Loaded: loaded (/usr/lib/systemd/system/cloud-config.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:21:11 UTC; 30min ago Main PID: 698 (code=exited, status=0/SUCCESS) Tasks: 0 (limit: 77075) Memory: 9.9M (peak: 47.6M) CPU: 573ms CGroup: /system.slice/cloud-config.service Apr 02 14:21:10 np0000163014 systemd[1]: Starting cloud-config.service - Cloud-init: Config Stage... Apr 02 14:21:11 np0000163014 cloud-init[760]: Cloud-init v. 25.3-0ubuntu1~24.04.1 running 'modules:config' at Thu, 02 Apr 2026 14:21:10 +0000. Up 9.70 seconds. Apr 02 14:21:11 np0000163014 systemd[1]: Finished cloud-config.service - Cloud-init: Config Stage. ● cloud-final.service - Cloud-init: Final Stage Loaded: loaded (/usr/lib/systemd/system/cloud-final.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:21:11 UTC; 30min ago Main PID: 781 (code=exited, status=0/SUCCESS) Tasks: 0 Memory: 424.0K (peak: 31.5M) CPU: 499ms CGroup: /system.slice/cloud-final.service Apr 02 14:21:11 np0000163014 cloud-init[801]: ############################################################# Apr 02 14:21:11 np0000163014 cloud-init[803]: -----BEGIN SSH HOST KEY FINGERPRINTS----- Apr 02 14:21:11 np0000163014 cloud-init[806]: 1024 SHA256:DBMKq7qW9eBI5cKYGatWopptUYK/AQi5hyMqhZwVyA0 root@np0000163014 (DSA) Apr 02 14:21:11 np0000163014 cloud-init[808]: 256 SHA256:hW8DxvuCwVHHC9oobbDFO0NvpZgoqD3YkhItNBLUdeY root@np0000163014 (ECDSA) Apr 02 14:21:11 np0000163014 cloud-init[810]: 256 SHA256:PdtmS5N33GO+h+LL6+BomfnhZ6FzTlmEhz4G0FXXaCk root@np0000163014 (ED25519) Apr 02 14:21:11 np0000163014 cloud-init[812]: 3072 SHA256:Ll8fxbsLl9bUmuHsfESCI7aMjTpHM4JwfruQZ2GwSGc root@np0000163014 (RSA) Apr 02 14:21:11 np0000163014 cloud-init[813]: -----END SSH HOST KEY FINGERPRINTS----- Apr 02 14:21:11 np0000163014 cloud-init[814]: ############################################################# Apr 02 14:21:11 np0000163014 cloud-init[796]: Cloud-init v. 25.3-0ubuntu1~24.04.1 finished at Thu, 02 Apr 2026 14:21:11 +0000. Datasource DataSourceConfigDrive [net,ver=2][source=/dev/sr0]. Up 10.61 seconds Apr 02 14:21:11 np0000163014 systemd[1]: Finished cloud-final.service - Cloud-init: Final Stage. ○ cloud-init-hotplugd.service - Cloud-init: Hotplug Hook Loaded: loaded (/usr/lib/systemd/system/cloud-init-hotplugd.service; static) Active: inactive (dead) TriggeredBy: ● cloud-init-hotplugd.socket ● cloud-init-local.service - Cloud-init: Local Stage (pre-network) Loaded: loaded (/usr/lib/systemd/system/cloud-init-local.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:21:05 UTC; 30min ago Main PID: 431 (code=exited, status=0/SUCCESS) Tasks: 0 (limit: 77075) Memory: 22.5M (peak: 67.5M) CPU: 887ms CGroup: /system.slice/cloud-init-local.service Apr 02 14:21:04 ubuntu systemd[1]: Starting cloud-init-local.service - Cloud-init: Local Stage (pre-network)... Apr 02 14:21:04 ubuntu cloud-init[544]: Cloud-init v. 25.3-0ubuntu1~24.04.1 running 'init-local' at Thu, 02 Apr 2026 14:21:04 +0000. Up 3.70 seconds. Apr 02 14:21:05 np0000163014 systemd[1]: Finished cloud-init-local.service - Cloud-init: Local Stage (pre-network). ● cloud-init.service - Cloud-init: Network Stage Loaded: loaded (/usr/lib/systemd/system/cloud-init.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:21:10 UTC; 30min ago Main PID: 609 (code=exited, status=0/SUCCESS) Tasks: 0 (limit: 77075) Memory: 27.4M (peak: 59.4M) CPU: 1.613s CGroup: /system.slice/cloud-init.service Apr 02 14:21:10 np0000163014 cloud-init[613]: | .. Eo+..| Apr 02 14:21:10 np0000163014 cloud-init[613]: | . ooo | Apr 02 14:21:10 np0000163014 cloud-init[613]: | . o o | Apr 02 14:21:10 np0000163014 cloud-init[613]: | S o + . | Apr 02 14:21:10 np0000163014 cloud-init[613]: | =.= ..| Apr 02 14:21:10 np0000163014 cloud-init[613]: | ++.X..++| Apr 02 14:21:10 np0000163014 cloud-init[613]: | =+oO=++.o| Apr 02 14:21:10 np0000163014 cloud-init[613]: | ...+OO+ oo| Apr 02 14:21:10 np0000163014 cloud-init[613]: +----[SHA256]-----+ Apr 02 14:21:10 np0000163014 systemd[1]: Finished cloud-init.service - Cloud-init: Network Stage. ● containerd.service - containerd container runtime Loaded: loaded (/usr/lib/systemd/system/containerd.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:29:33 UTC; 22min ago Docs: https://containerd.io Main PID: 20255 (containerd) Tasks: 33 Memory: 1.3G (peak: 1.3G) CPU: 14.278s CGroup: /system.slice/containerd.service ├─20255 /usr/bin/containerd └─21112 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8 -address /run/containerd/containerd.sock Apr 02 14:29:33 np0000163014 containerd[20255]: time="2026-04-02T14:29:33.155225958Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Apr 02 14:29:33 np0000163014 containerd[20255]: time="2026-04-02T14:29:33.155235108Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Apr 02 14:29:33 np0000163014 containerd[20255]: time="2026-04-02T14:29:33.155244438Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Apr 02 14:29:33 np0000163014 containerd[20255]: time="2026-04-02T14:29:33.155252628Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Apr 02 14:29:33 np0000163014 containerd[20255]: time="2026-04-02T14:29:33.155262999Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Apr 02 14:29:33 np0000163014 containerd[20255]: time="2026-04-02T14:29:33.155474552Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 02 14:29:33 np0000163014 containerd[20255]: time="2026-04-02T14:29:33.155531693Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 02 14:29:33 np0000163014 containerd[20255]: time="2026-04-02T14:29:33.155595174Z" level=info msg="containerd successfully booted in 0.031957s" Apr 02 14:29:33 np0000163014 systemd[1]: Started containerd.service - containerd container runtime. Apr 02 14:29:56 np0000163014 containerd[20255]: time="2026-04-02T14:29:56.800464310Z" level=info msg="connecting to shim 54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8" address="unix:///run/containerd/s/3bfeba5b95ec38c8dbff2d0d7c6ba274780ed72bb3fbb26d0a587aed259e755a" namespace=moby protocol=ttrpc version=3 ● dbus.service - D-Bus System Message Bus Loaded: loaded (/usr/lib/systemd/system/dbus.service; static) Active: active (running) since Thu 2026-04-02 14:21:10 UTC; 30min ago TriggeredBy: ● dbus.socket Docs: man:dbus-daemon(1) Main PID: 699 (dbus-daemon) Tasks: 1 (limit: 77075) Memory: 2.3M (peak: 3.1M) CPU: 6.374s CGroup: /system.slice/dbus.service └─699 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only Apr 02 14:27:17 np0000163014 dbus-daemon[699]: Unknown username "dnsmasq" in message bus configuration file Apr 02 14:27:17 np0000163014 dbus-daemon[699]: [system] Reloaded configuration Apr 02 14:28:26 np0000163014 dbus-daemon[699]: [system] Reloaded configuration Apr 02 14:34:30 np0000163014 dbus-daemon[699]: [system] Reloaded configuration Apr 02 14:34:30 np0000163014 dbus-daemon[699]: [system] Reloaded configuration Apr 02 14:34:30 np0000163014 dbus-daemon[699]: [system] Reloaded configuration Apr 02 14:34:30 np0000163014 dbus-daemon[699]: [system] Reloaded configuration Apr 02 14:34:30 np0000163014 dbus-daemon[699]: [system] Reloaded configuration Apr 02 14:34:30 np0000163014 dbus-daemon[699]: [system] Reloaded configuration Apr 02 14:34:46 np0000163014 dbus-daemon[699]: [system] Reloaded configuration ● devstack@barbican-keystone-listener.service - Devstack devstack@barbican-keystone-listener.service Loaded: loaded (/etc/systemd/system/devstack@barbican-keystone-listener.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:47:40 UTC; 4min 13s ago Main PID: 117066 (barbican-keysto) Tasks: 28 (limit: 77075) Memory: 87.0M (peak: 88.6M) CPU: 1.666s CGroup: /system.slice/system-devstack.slice/devstack@barbican-keystone-listener.service ├─117066 "barbican-keystone-listener: master process [/opt/stack/data/venv/bin/barbican-keystone-listener --config-file=/etc/barbican/barbican.conf]" └─117314 "barbican-keystone-listener: ServiceWrapper worker(0)" Apr 02 14:51:06 np0000163014 barbican-keystone-listener[117314]: 2026-04-02 14:51:06.526 117314 DEBUG barbican.queue.keystone_listener [-] Input keystone event publisher_id = servergroup.np0000163014 process_event /opt/stack/barbican/barbican/queue/keystone_listener.py:74 Apr 02 14:51:06 np0000163014 barbican-keystone-listener[117314]: 2026-04-02 14:51:06.528 117314 DEBUG barbican.queue.keystone_listener [-] Input keystone event payload = {'name': 'kube-kz2tz', 'policy': 'soft-anti-affinity', 'project_id': 'f3c2a2cf652e4eed8bfe6f00c3630d38', 'user_id': 'bb8ec9c5ec0b4bc6b2ab8e5afb1e53a1', 'server_group_id': '30bdf780-b3ff-42f2-afd3-972684fca189'} process_event /opt/stack/barbican/barbican/queue/keystone_listener.py:75 Apr 02 14:51:06 np0000163014 barbican-keystone-listener[117314]: 2026-04-02 14:51:06.529 117314 DEBUG barbican.queue.keystone_listener [-] Input keystone event type = servergroup.create process_event /opt/stack/barbican/barbican/queue/keystone_listener.py:76 Apr 02 14:51:06 np0000163014 barbican-keystone-listener[117314]: 2026-04-02 14:51:06.529 117314 DEBUG barbican.queue.keystone_listener [-] Input keystone event metadata = {'message_id': '322dc9a1-6926-4a71-9563-649f7a593d7e', 'timestamp': '2026-04-02 14:51:06.506211'} process_event /opt/stack/barbican/barbican/queue/keystone_listener.py:77 Apr 02 14:51:06 np0000163014 barbican-keystone-listener[117314]: 2026-04-02 14:51:06.529 117314 DEBUG barbican.queue.keystone_listener [-] Keystone Event: resource type=None, operation type=None, keystone id=None process_event /opt/stack/barbican/barbican/queue/keystone_listener.py:80 Apr 02 14:51:06 np0000163014 barbican-keystone-listener[117314]: 2026-04-02 14:51:06.691 117314 DEBUG barbican.queue.keystone_listener [-] Input keystone event publisher_id = servergroup.np0000163014 process_event /opt/stack/barbican/barbican/queue/keystone_listener.py:74 Apr 02 14:51:06 np0000163014 barbican-keystone-listener[117314]: 2026-04-02 14:51:06.691 117314 DEBUG barbican.queue.keystone_listener [-] Input keystone event payload = {'name': 'kube-kz2tz-default-worker', 'policy': 'soft-anti-affinity', 'project_id': 'f3c2a2cf652e4eed8bfe6f00c3630d38', 'user_id': 'bb8ec9c5ec0b4bc6b2ab8e5afb1e53a1', 'server_group_id': '2a42ffde-27e4-4dbe-9db3-5ebbfda49633'} process_event /opt/stack/barbican/barbican/queue/keystone_listener.py:75 Apr 02 14:51:06 np0000163014 barbican-keystone-listener[117314]: 2026-04-02 14:51:06.691 117314 DEBUG barbican.queue.keystone_listener [-] Input keystone event type = servergroup.create process_event /opt/stack/barbican/barbican/queue/keystone_listener.py:76 Apr 02 14:51:06 np0000163014 barbican-keystone-listener[117314]: 2026-04-02 14:51:06.692 117314 DEBUG barbican.queue.keystone_listener [-] Input keystone event metadata = {'message_id': '72b989de-61df-4a34-9e49-272d007211b7', 'timestamp': '2026-04-02 14:51:06.674881'} process_event /opt/stack/barbican/barbican/queue/keystone_listener.py:77 Apr 02 14:51:06 np0000163014 barbican-keystone-listener[117314]: 2026-04-02 14:51:06.692 117314 DEBUG barbican.queue.keystone_listener [-] Keystone Event: resource type=None, operation type=None, keystone id=None process_event /opt/stack/barbican/barbican/queue/keystone_listener.py:80 ● devstack@barbican-retry.service - Devstack devstack@barbican-retry.service Loaded: loaded (/etc/systemd/system/devstack@barbican-retry.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:47:38 UTC; 4min 15s ago Main PID: 116544 (barbican-retry:) Tasks: 7 (limit: 77075) Memory: 94.4M (peak: 94.6M) CPU: 1.341s CGroup: /system.slice/system-devstack.slice/devstack@barbican-retry.service ├─116544 "barbican-retry: master process [/opt/stack/data/venv/bin/barbican-retry --config-file=/etc/barbican/barbican.conf]" └─116844 "barbican-retry: ServiceWrapper worker(0)" Apr 02 14:51:38 np0000163014 barbican-retry[116544]: 2026-04-02 14:51:38.002 116544 INFO barbican.queue.retry_scheduler [-] Done processing '0' tasks, will check again in '11.049984533449232' seconds. Apr 02 14:51:38 np0000163014 barbican-retry[116544]: 2026-04-02 14:51:38.002 116544 DEBUG oslo.service.backend._threading.loopingcall [-] Dynamic interval looping call 'barbican.queue.retry_scheduler.PeriodicServer._check_retry_tasks' sleeping for 10.00 seconds _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125 Apr 02 14:51:48 np0000163014 barbican-retry[116544]: 2026-04-02 14:51:48.000 116544 DEBUG dbcounter [-] [116544] Writing DB stats barbican:SELECT=2 stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115 Apr 02 14:51:48 np0000163014 barbican-retry[116544]: 2026-04-02 14:51:48.003 116544 INFO barbican.queue.retry_scheduler [-] Processing scheduled retry tasks: Apr 02 14:51:48 np0000163014 barbican-retry[116544]: 2026-04-02 14:51:48.003 116544 DEBUG barbican.model.repositories [-] Clean paging values limit=10, offset=0 clean_paging_values /opt/stack/barbican/barbican/model/repositories.py:267 Apr 02 14:51:48 np0000163014 barbican-retry[116544]: 2026-04-02 14:51:48.004 116544 DEBUG barbican.model.repositories [-] Getting session... get_session /opt/stack/barbican/barbican/model/repositories.py:309 Apr 02 14:51:48 np0000163014 barbican-retry[116544]: 2026-04-02 14:51:48.004 116544 DEBUG barbican.model.repositories [-] Retrieving from 0 to 10 get_by_create_date /opt/stack/barbican/barbican/model/repositories.py:1263 Apr 02 14:51:48 np0000163014 barbican-retry[116544]: 2026-04-02 14:51:48.009 116544 DEBUG barbican.model.repositories [-] Number entities retrieved: 0 out of 0 get_by_create_date /opt/stack/barbican/barbican/model/repositories.py:1266 Apr 02 14:51:48 np0000163014 barbican-retry[116544]: 2026-04-02 14:51:48.010 116544 INFO barbican.queue.retry_scheduler [-] Done processing '0' tasks, will check again in '8.452985119687538' seconds. Apr 02 14:51:48 np0000163014 barbican-retry[116544]: 2026-04-02 14:51:48.010 116544 DEBUG oslo.service.backend._threading.loopingcall [-] Dynamic interval looping call 'barbican.queue.retry_scheduler.PeriodicServer._check_retry_tasks' sleeping for 8.45 seconds _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125 ● devstack@barbican-svc.service - Devstack devstack@barbican-svc.service Loaded: loaded (/etc/systemd/system/devstack@barbican-svc.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:47:26 UTC; 4min 27s ago Main PID: 115994 (uwsgi) Status: "uWSGI is ready" Tasks: 9 (limit: 77075) Memory: 377.5M (peak: 378.5M) CPU: 8.277s CGroup: /system.slice/system-devstack.slice/devstack@barbican-svc.service ├─115994 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv ├─115995 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv ├─115996 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv ├─115997 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv └─115998 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv Apr 02 14:51:06 np0000163014 devstack@barbican-svc.service[115998]: 2026-04-02 14:51:06.303 115998 DEBUG barbican.model.repositories [None req-f56cd750-ae7e-41de-ae25-e2f5aa7a34fb b797868cf1134547b30cb5bd93fbc66f 1405e5e4df0c46848af022b1c5632d76 - - default default] Getting session... get_session /opt/stack/barbican/barbican/model/repositories.py:309 Apr 02 14:51:06 np0000163014 devstack@barbican-svc.service[115998]: 2026-04-02 14:51:06.311 115998 DEBUG barbican.api.controllers.secrets [None req-f56cd750-ae7e-41de-ae25-e2f5aa7a34fb b797868cf1134547b30cb5bd93fbc66f 1405e5e4df0c46848af022b1c5632d76 - - default default] === Creating SecretController === __init__ /opt/stack/barbican/barbican/api/controllers/secrets.py:79 Apr 02 14:51:06 np0000163014 devstack@barbican-svc.service[115998]: 2026-04-02 14:51:06.318 115998 DEBUG barbican.model.repositories [None req-f56cd750-ae7e-41de-ae25-e2f5aa7a34fb b797868cf1134547b30cb5bd93fbc66f 1405e5e4df0c46848af022b1c5632d76 - - default default] Getting session... get_session /opt/stack/barbican/barbican/model/repositories.py:309 Apr 02 14:51:06 np0000163014 devstack@barbican-svc.service[115998]: 2026-04-02 14:51:06.321 115998 INFO barbican.api.controllers.secrets [None req-f56cd750-ae7e-41de-ae25-e2f5aa7a34fb b797868cf1134547b30cb5bd93fbc66f 1405e5e4df0c46848af022b1c5632d76 - - default default] Retrieved secret payload for project: 1405e5e4df0c46848af022b1c5632d76 Apr 02 14:51:06 np0000163014 devstack@barbican-svc.service[115998]: 2026-04-02 14:51:06.322 115998 INFO barbican.api.middleware.context [None req-f56cd750-ae7e-41de-ae25-e2f5aa7a34fb b797868cf1134547b30cb5bd93fbc66f 1405e5e4df0c46848af022b1c5632d76 - - default default] Processed request: 200 OK - GET http://199.204.45.109/key-manager/v1/secrets/78ba91d4-7cb0-4817-926a-47c9748a9af1/payload Apr 02 14:51:06 np0000163014 devstack@barbican-svc.service[115998]: [pid: 115998|app: 0|req: 12/45] 199.204.45.109 () {58 vars in 1325 bytes} [Thu Apr 2 14:51:06 2026] GET /key-manager/v1/secrets/78ba91d4-7cb0-4817-926a-47c9748a9af1/payload => generated 12 bytes in 21 msecs (HTTP/1.1 200) 6 headers in 234 bytes (1 switches on core 0) Apr 02 14:51:16 np0000163014 devstack@barbican-svc.service[115996]: 2026-04-02 14:51:16.242 115996 DEBUG dbcounter [-] [115996] Writing DB stats barbican:SELECT=42,barbican:INSERT=19,barbican:UPDATE=6 stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115 Apr 02 14:51:16 np0000163014 devstack@barbican-svc.service[115995]: 2026-04-02 14:51:16.268 115995 DEBUG dbcounter [-] [115995] Writing DB stats barbican:SELECT=41,barbican:INSERT=17,barbican:UPDATE=5 stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115 Apr 02 14:51:16 np0000163014 devstack@barbican-svc.service[115997]: 2026-04-02 14:51:16.296 115997 DEBUG dbcounter [-] [115997] Writing DB stats barbican:SELECT=41,barbican:INSERT=17,barbican:UPDATE=5 stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115 Apr 02 14:51:16 np0000163014 devstack@barbican-svc.service[115998]: 2026-04-02 14:51:16.320 115998 DEBUG dbcounter [-] [115998] Writing DB stats barbican:SELECT=49,barbican:INSERT=17,barbican:UPDATE=1 stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115 ● devstack@c-api.service - Devstack devstack@c-api.service Loaded: loaded (/etc/systemd/system/devstack@c-api.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:46:17 UTC; 5min ago Main PID: 111318 (uwsgi) Status: "uWSGI is ready" Tasks: 13 (limit: 77075) Memory: 447.9M (peak: 448.4M) CPU: 10.160s CGroup: /system.slice/system-devstack.slice/devstack@c-api.service ├─111318 "cinder-apiuWSGI master" ├─111320 "cinder-apiuWSGI worker 1" ├─111321 "cinder-apiuWSGI worker 2" ├─111322 "cinder-apiuWSGI worker 3" └─111323 "cinder-apiuWSGI worker 4" Apr 02 14:50:29 np0000163014 devstack@c-api.service[111321]: /opt/stack/data/venv/lib/python3.12/site-packages/oslo_policy/policy.py:806: UserWarning: Policy "volume_extension:default_get_all":"role:admin and system_scope:all" was deprecated in X in favor of "volume_extension:default_get_all":"rule:admin_api". Reason: Default policies now support the three Keystone default roles, namely 'admin', 'member', and 'reader' to implement three Cinder "personas". See "Policy Personas and Permissions" in the "Cinder Service Configuration" documentation (Xena release) for details.. Either ensure your deployment is ready for the new default or copy/paste the deprecated policy into your policy file and maintain it manually. Apr 02 14:50:29 np0000163014 devstack@c-api.service[111321]: warnings.warn(deprecated_msg) Apr 02 14:50:29 np0000163014 devstack@c-api.service[111321]: /opt/stack/data/venv/lib/python3.12/site-packages/oslo_policy/policy.py:806: UserWarning: Policy "volume_extension:default_unset":"rule:system_or_domain_or_project_admin" was deprecated in X in favor of "volume_extension:default_unset":"rule:admin_api". Reason: Default policies now support the three Keystone default roles, namely 'admin', 'member', and 'reader' to implement three Cinder "personas". See "Policy Personas and Permissions" in the "Cinder Service Configuration" documentation (Xena release) for details.. Either ensure your deployment is ready for the new default or copy/paste the deprecated policy into your policy file and maintain it manually. Apr 02 14:50:29 np0000163014 devstack@c-api.service[111321]: warnings.warn(deprecated_msg) Apr 02 14:50:29 np0000163014 devstack@c-api.service[111321]: DEBUG cinder.api.middleware.request_id [None req-8f6db66a-a5b6-4264-882a-dc1d13da22c7 None None] RequestId filter calling following filter/app {{(pid=111321) _context_setter /opt/stack/cinder/cinder/api/middleware/request_id.py:62}} Apr 02 14:50:29 np0000163014 devstack@c-api.service[111321]: INFO cinder.api.openstack.wsgi [None req-8f6db66a-a5b6-4264-882a-dc1d13da22c7 None None] GET https://199.204.45.109/volume// Apr 02 14:50:29 np0000163014 devstack@c-api.service[111321]: DEBUG cinder.api.openstack.wsgi [None req-8f6db66a-a5b6-4264-882a-dc1d13da22c7 None None] Empty body provided in request {{(pid=111321) get_body /opt/stack/cinder/cinder/api/openstack/wsgi.py:725}} Apr 02 14:50:29 np0000163014 devstack@c-api.service[111321]: DEBUG cinder.api.openstack.wsgi [None req-8f6db66a-a5b6-4264-882a-dc1d13da22c7 None None] Calling method 'all' {{(pid=111321) _process_stack /opt/stack/cinder/cinder/api/openstack/wsgi.py:878}} Apr 02 14:50:29 np0000163014 devstack@c-api.service[111321]: INFO cinder.api.openstack.wsgi [None req-8f6db66a-a5b6-4264-882a-dc1d13da22c7 None None] https://199.204.45.109/volume// returned with HTTP 300 Apr 02 14:50:29 np0000163014 devstack@c-api.service[111321]: [pid: 111321|app: 0|req: 1/4] 199.204.45.109 () {66 vars in 1456 bytes} [Thu Apr 2 14:50:29 2026] GET /volume/ => generated 390 bytes in 18 msecs (HTTP/1.1 300) 7 headers in 299 bytes (1 switches on core 0) ● devstack@c-bak.service - Devstack devstack@c-bak.service Loaded: loaded (/etc/systemd/system/devstack@c-bak.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:46:23 UTC; 5min ago Main PID: 112602 (cinder-backup) Tasks: 1 (limit: 77075) Memory: 91.6M (peak: 92.1M) CPU: 1.780s CGroup: /system.slice/system-devstack.slice/devstack@c-bak.service └─112602 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-backup --config-file /etc/cinder/cinder.conf Apr 02 14:48:10 np0000163014 cinder-backup[112602]: DEBUG oslo_service.periodic_task [None req-fe38924b-f290-476e-b785-96e93f15dcb8 None None] Running periodic task BackupManager.publish_service_capabilities {{(pid=112602) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:48:10 np0000163014 cinder-backup[112602]: DEBUG cinder.manager [None req-fe38924b-f290-476e-b785-96e93f15dcb8 None None] Notifying Schedulers of capabilities ... {{(pid=112602) _publish_service_capabilities /opt/stack/cinder/cinder/manager.py:202}} Apr 02 14:48:25 np0000163014 cinder-backup[112602]: ERROR cinder.service [-] Manager for service cinder-backup np0000163014 is reporting problems, not sending heartbeat. Service will appear "down". Apr 02 14:49:10 np0000163014 cinder-backup[112602]: DEBUG oslo_service.periodic_task [None req-fe38924b-f290-476e-b785-96e93f15dcb8 None None] Running periodic task BackupManager.publish_service_capabilities {{(pid=112602) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:49:10 np0000163014 cinder-backup[112602]: DEBUG cinder.manager [None req-fe38924b-f290-476e-b785-96e93f15dcb8 None None] Notifying Schedulers of capabilities ... {{(pid=112602) _publish_service_capabilities /opt/stack/cinder/cinder/manager.py:202}} Apr 02 14:50:10 np0000163014 cinder-backup[112602]: DEBUG oslo_service.periodic_task [None req-fe38924b-f290-476e-b785-96e93f15dcb8 None None] Running periodic task BackupManager.publish_service_capabilities {{(pid=112602) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:50:10 np0000163014 cinder-backup[112602]: DEBUG cinder.manager [None req-fe38924b-f290-476e-b785-96e93f15dcb8 None None] Notifying Schedulers of capabilities ... {{(pid=112602) _publish_service_capabilities /opt/stack/cinder/cinder/manager.py:202}} Apr 02 14:50:25 np0000163014 cinder-backup[112602]: ERROR cinder.service [-] Manager for service cinder-backup np0000163014 is reporting problems, not sending heartbeat. Service will appear "down". Apr 02 14:51:10 np0000163014 cinder-backup[112602]: DEBUG oslo_service.periodic_task [None req-fe38924b-f290-476e-b785-96e93f15dcb8 None None] Running periodic task BackupManager.publish_service_capabilities {{(pid=112602) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:10 np0000163014 cinder-backup[112602]: DEBUG cinder.manager [None req-fe38924b-f290-476e-b785-96e93f15dcb8 None None] Notifying Schedulers of capabilities ... {{(pid=112602) _publish_service_capabilities /opt/stack/cinder/cinder/manager.py:202}} ● devstack@c-sch.service - Devstack devstack@c-sch.service Loaded: loaded (/etc/systemd/system/devstack@c-sch.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:46:21 UTC; 5min ago Main PID: 112032 (cinder-schedule) Tasks: 1 (limit: 77075) Memory: 105.7M (peak: 106.2M) CPU: 2.047s CGroup: /system.slice/system-devstack.slice/devstack@c-sch.service └─112032 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-scheduler --config-file /etc/cinder/cinder.conf Apr 02 14:48:10 np0000163014 cinder-scheduler[112032]: DEBUG cinder.scheduler.host_manager [None req-f25f4e68-e63e-4bdd-ac10-3632276f9c55 None None] Received backup service update from np0000163014: {'backend_state': False, 'driver_name': 'cinder.backup.drivers.swift.SwiftBackupDriver', 'availability_zone': 'nova'} {{(pid=112032) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:598}} Apr 02 14:48:44 np0000163014 cinder-scheduler[112032]: DEBUG cinder.scheduler.host_manager [None req-d211bbb5-9b26-487b-ae07-79a94aafd638 None None] Received volume service update from np0000163014@lvmdriver-1: {'volume_backend_name': 'lvmdriver-1', 'vendor_name': 'Open Source', 'driver_version': '3.0.0', 'storage_protocol': 'iSCSI', 'pools': [{'pool_name': 'lvmdriver-1', 'total_capacity_gb': 47.5, 'free_capacity_gb': 47.5, 'reserved_percentage': 0, 'location_info': 'LVMVolumeDriver:np0000163014:stack-volumes-lvmdriver-1:thin:0', 'QoS_support': False, 'provisioned_capacity_gb': 0.0, 'max_over_subscription_ratio': '20.0', 'thin_provisioning_support': True, 'thick_provisioning_support': False, 'total_volumes': 1, 'filter_function': None, 'goodness_function': None, 'multiattach': True, 'backend_state': 'up', 'allocated_capacity_gb': 0, 'cacheable': True}], 'shared_targets': False, 'sparse_copy_volume': True, 'filter_function': None, 'goodness_function': None} {{(pid=112032) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:629}} Apr 02 14:48:46 np0000163014 cinder-scheduler[112032]: DEBUG dbcounter [-] [112032] Writing DB stats cinder:SELECT=1,cinder:UPDATE=1 {{(pid=112032) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:49:10 np0000163014 cinder-scheduler[112032]: DEBUG cinder.scheduler.host_manager [None req-0f93c5ed-0139-4d2f-905e-1223ec1b6fc2 None None] Received backup service update from np0000163014: {'backend_state': False, 'driver_name': 'cinder.backup.drivers.swift.SwiftBackupDriver', 'availability_zone': 'nova'} {{(pid=112032) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:598}} Apr 02 14:49:44 np0000163014 cinder-scheduler[112032]: DEBUG cinder.scheduler.host_manager [None req-27bce39b-a789-4cc3-969d-eab009378c76 None None] Received volume service update from np0000163014@lvmdriver-1: {'volume_backend_name': 'lvmdriver-1', 'vendor_name': 'Open Source', 'driver_version': '3.0.0', 'storage_protocol': 'iSCSI', 'pools': [{'pool_name': 'lvmdriver-1', 'total_capacity_gb': 47.5, 'free_capacity_gb': 47.5, 'reserved_percentage': 0, 'location_info': 'LVMVolumeDriver:np0000163014:stack-volumes-lvmdriver-1:thin:0', 'QoS_support': False, 'provisioned_capacity_gb': 0.0, 'max_over_subscription_ratio': '20.0', 'thin_provisioning_support': True, 'thick_provisioning_support': False, 'total_volumes': 1, 'filter_function': None, 'goodness_function': None, 'multiattach': True, 'backend_state': 'up', 'allocated_capacity_gb': 0, 'cacheable': True}], 'shared_targets': False, 'sparse_copy_volume': True, 'filter_function': None, 'goodness_function': None} {{(pid=112032) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:629}} Apr 02 14:50:10 np0000163014 cinder-scheduler[112032]: DEBUG cinder.scheduler.host_manager [None req-7a1e385f-5dc4-4180-a8e7-35881e5e2b31 None None] Received backup service update from np0000163014: {'backend_state': False, 'driver_name': 'cinder.backup.drivers.swift.SwiftBackupDriver', 'availability_zone': 'nova'} {{(pid=112032) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:598}} Apr 02 14:50:44 np0000163014 cinder-scheduler[112032]: DEBUG cinder.scheduler.host_manager [None req-273c2643-1ece-4016-b1c4-44639332b06c None None] Received volume service update from np0000163014@lvmdriver-1: {'volume_backend_name': 'lvmdriver-1', 'vendor_name': 'Open Source', 'driver_version': '3.0.0', 'storage_protocol': 'iSCSI', 'pools': [{'pool_name': 'lvmdriver-1', 'total_capacity_gb': 47.5, 'free_capacity_gb': 47.5, 'reserved_percentage': 0, 'location_info': 'LVMVolumeDriver:np0000163014:stack-volumes-lvmdriver-1:thin:0', 'QoS_support': False, 'provisioned_capacity_gb': 0.0, 'max_over_subscription_ratio': '20.0', 'thin_provisioning_support': True, 'thick_provisioning_support': False, 'total_volumes': 1, 'filter_function': None, 'goodness_function': None, 'multiattach': True, 'backend_state': 'up', 'allocated_capacity_gb': 0, 'cacheable': True}], 'shared_targets': False, 'sparse_copy_volume': True, 'filter_function': None, 'goodness_function': None} {{(pid=112032) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:629}} Apr 02 14:50:46 np0000163014 cinder-scheduler[112032]: DEBUG dbcounter [-] [112032] Writing DB stats cinder:SELECT=1,cinder:UPDATE=1 {{(pid=112032) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:10 np0000163014 cinder-scheduler[112032]: DEBUG cinder.scheduler.host_manager [None req-a4992da9-d44c-4b86-923f-b3247e73ec4c None None] Received backup service update from np0000163014: {'backend_state': False, 'driver_name': 'cinder.backup.drivers.swift.SwiftBackupDriver', 'availability_zone': 'nova'} {{(pid=112032) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:598}} Apr 02 14:51:44 np0000163014 cinder-scheduler[112032]: DEBUG cinder.scheduler.host_manager [None req-ee1e27b9-67ea-4664-89b5-277a6e155f44 None None] Received volume service update from np0000163014@lvmdriver-1: {'volume_backend_name': 'lvmdriver-1', 'vendor_name': 'Open Source', 'driver_version': '3.0.0', 'storage_protocol': 'iSCSI', 'pools': [{'pool_name': 'lvmdriver-1', 'total_capacity_gb': 47.5, 'free_capacity_gb': 47.5, 'reserved_percentage': 0, 'location_info': 'LVMVolumeDriver:np0000163014:stack-volumes-lvmdriver-1:thin:0', 'QoS_support': False, 'provisioned_capacity_gb': 0.0, 'max_over_subscription_ratio': '20.0', 'thin_provisioning_support': True, 'thick_provisioning_support': False, 'total_volumes': 1, 'filter_function': None, 'goodness_function': None, 'multiattach': True, 'backend_state': 'up', 'allocated_capacity_gb': 0, 'cacheable': True}], 'shared_targets': False, 'sparse_copy_volume': True, 'filter_function': None, 'goodness_function': None} {{(pid=112032) update_service_capabilities /opt/stack/cinder/cinder/scheduler/host_manager.py:629}} ● devstack@c-vol.service - Devstack devstack@c-vol.service Loaded: loaded (/etc/systemd/system/devstack@c-vol.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:46:26 UTC; 5min ago Main PID: 113189 (cinder-volume) Tasks: 2 (limit: 77075) Memory: 135.8M (peak: 165.7M) CPU: 20.377s CGroup: /system.slice/system-devstack.slice/devstack@c-vol.service ├─113189 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf └─113469 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf Apr 02 14:51:43 np0000163014 sudo[129558]: stack : PWD=/ ; USER=root ; COMMAND=/opt/stack/data/venv/bin/cinder-rootwrap /etc/cinder/rootwrap.conf env LC_ALL=C lvs --noheadings --unit=g -o size,data_percent --separator : --nosuffix /dev/stack-volumes-lvmdriver-1/stack-volumes-lvmdriver-1-pool Apr 02 14:51:43 np0000163014 sudo[129558]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=1002) Apr 02 14:51:43 np0000163014 sudo[129558]: pam_unix(sudo:session): session closed for user root Apr 02 14:51:43 np0000163014 cinder-volume[113469]: DEBUG oslo_concurrency.processutils [None req-3a330c88-9322-4259-a90c-796e7fba09fa None None] CMD "sudo cinder-rootwrap /etc/cinder/rootwrap.conf env LC_ALL=C lvs --noheadings --unit=g -o size,data_percent --separator : --nosuffix /dev/stack-volumes-lvmdriver-1/stack-volumes-lvmdriver-1-pool" returned: 0 in 0.383s {{(pid=113469) execute /opt/stack/data/venv/lib/python3.12/site-packages/oslo_concurrency/processutils.py:468}} Apr 02 14:51:43 np0000163014 cinder-volume[113469]: DEBUG oslo_concurrency.processutils [None req-3a330c88-9322-4259-a90c-796e7fba09fa None None] Running cmd (subprocess): sudo cinder-rootwrap /etc/cinder/rootwrap.conf env LC_ALL=C lvs --noheadings --unit=g -o vg_name,name,size --nosuffix --readonly stack-volumes-lvmdriver-1 {{(pid=113469) execute /opt/stack/data/venv/lib/python3.12/site-packages/oslo_concurrency/processutils.py:440}} Apr 02 14:51:43 np0000163014 sudo[129739]: stack : PWD=/ ; USER=root ; COMMAND=/opt/stack/data/venv/bin/cinder-rootwrap /etc/cinder/rootwrap.conf env LC_ALL=C lvs --noheadings --unit=g -o vg_name,name,size --nosuffix --readonly stack-volumes-lvmdriver-1 Apr 02 14:51:43 np0000163014 sudo[129739]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=1002) Apr 02 14:51:44 np0000163014 sudo[129739]: pam_unix(sudo:session): session closed for user root Apr 02 14:51:44 np0000163014 cinder-volume[113469]: DEBUG oslo_concurrency.processutils [None req-3a330c88-9322-4259-a90c-796e7fba09fa None None] CMD "sudo cinder-rootwrap /etc/cinder/rootwrap.conf env LC_ALL=C lvs --noheadings --unit=g -o vg_name,name,size --nosuffix --readonly stack-volumes-lvmdriver-1" returned: 0 in 0.363s {{(pid=113469) execute /opt/stack/data/venv/lib/python3.12/site-packages/oslo_concurrency/processutils.py:468}} Apr 02 14:51:44 np0000163014 cinder-volume[113469]: DEBUG cinder.manager [None req-3a330c88-9322-4259-a90c-796e7fba09fa None None] Notifying Schedulers of capabilities ... {{(pid=113469) _publish_service_capabilities /opt/stack/cinder/cinder/manager.py:202}} ● devstack@etcd.service - Devstack devstack@etcd.service Loaded: loaded (/etc/systemd/system/devstack@etcd.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:39:12 UTC; 12min ago Main PID: 63970 (etcd) Tasks: 17 (limit: 77075) Memory: 133.7M (peak: 134.8M) CPU: 5.233s CGroup: /system.slice/system-devstack.slice/devstack@etcd.service └─63970 /opt/stack/bin/etcd --name np0000163014 --data-dir /opt/stack/data/etcd --initial-cluster-state new --initial-cluster-token etcd-cluster-01 --initial-cluster np0000163014=http://199.204.45.109:2380 --initial-advertise-peer-urls http://199.204.45.109:2380 --advertise-client-urls http://199.204.45.109:2379 --listen-peer-urls http://0.0.0.0:2380 --listen-client-urls http://199.204.45.109:2379 --log-level=debug Apr 02 14:51:33 np0000163014 etcd[63970]: {"level":"debug","ts":"2026-04-02T14:51:33.380764Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-02T14:51:33.300815Z","now":"2026-04-02T14:51:33.380761Z"} Apr 02 14:51:33 np0000163014 etcd[63970]: {"level":"debug","ts":"2026-04-02T14:51:33.400762Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-02T14:51:33.300815Z","now":"2026-04-02T14:51:33.400759Z"} Apr 02 14:51:33 np0000163014 etcd[63970]: {"level":"debug","ts":"2026-04-02T14:51:33.405102Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-02T14:51:33.400789Z","now":"2026-04-02T14:51:33.405100Z"} Apr 02 14:51:33 np0000163014 etcd[63970]: {"level":"debug","ts":"2026-04-02T14:51:33.469135Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-02T14:51:33.400789Z","now":"2026-04-02T14:51:33.469133Z"} Apr 02 14:51:40 np0000163014 etcd[63970]: {"level":"debug","ts":"2026-04-02T14:51:40.472649Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-02T14:51:40.400769Z","now":"2026-04-02T14:51:40.472645Z"} Apr 02 14:51:43 np0000163014 etcd[63970]: {"level":"debug","ts":"2026-04-02T14:51:43.107305Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-02T14:51:43.101340Z","now":"2026-04-02T14:51:43.107302Z"} Apr 02 14:51:48 np0000163014 etcd[63970]: {"level":"debug","ts":"2026-04-02T14:51:48.383086Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-02T14:51:48.300759Z","now":"2026-04-02T14:51:48.383083Z"} Apr 02 14:51:48 np0000163014 etcd[63970]: {"level":"debug","ts":"2026-04-02T14:51:48.403274Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-02T14:51:48.400775Z","now":"2026-04-02T14:51:48.403272Z"} Apr 02 14:51:48 np0000163014 etcd[63970]: {"level":"debug","ts":"2026-04-02T14:51:48.407058Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-02T14:51:48.400775Z","now":"2026-04-02T14:51:48.407056Z"} Apr 02 14:51:48 np0000163014 etcd[63970]: {"level":"debug","ts":"2026-04-02T14:51:48.471183Z","caller":"etcdserver/server.go:1230","msg":"The member is active, skip checking leadership","latestTickTs":"2026-04-02T14:51:48.400775Z","now":"2026-04-02T14:51:48.471181Z"} ● devstack@file_tracker.service - Devstack devstack@file_tracker.service Loaded: loaded (/etc/systemd/system/devstack@file_tracker.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:39:09 UTC; 12min ago Main PID: 63325 (file_tracker.sh) Tasks: 2 (limit: 77075) Memory: 596.0K (peak: 1.1M) CPU: 249ms CGroup: /system.slice/system-devstack.slice/devstack@file_tracker.service ├─ 63325 /bin/bash /opt/stack/devstack/tools/file_tracker.sh └─129995 sleep 20 Apr 02 14:48:49 np0000163014 file_tracker.sh[120117]: 9504 0 9223372036854775807 Apr 02 14:49:09 np0000163014 file_tracker.sh[120131]: 9472 0 9223372036854775807 Apr 02 14:49:29 np0000163014 file_tracker.sh[120155]: 9504 0 9223372036854775807 Apr 02 14:49:49 np0000163014 file_tracker.sh[120389]: 9472 0 9223372036854775807 Apr 02 14:50:09 np0000163014 file_tracker.sh[122116]: 9824 0 9223372036854775807 Apr 02 14:50:29 np0000163014 file_tracker.sh[124927]: 10368 0 9223372036854775807 Apr 02 14:50:49 np0000163014 file_tracker.sh[127845]: 10688 0 9223372036854775807 Apr 02 14:51:09 np0000163014 file_tracker.sh[128008]: 10944 0 9223372036854775807 Apr 02 14:51:29 np0000163014 file_tracker.sh[128368]: 11264 0 9223372036854775807 Apr 02 14:51:49 np0000163014 file_tracker.sh[129994]: 11392 0 9223372036854775807 ● devstack@g-api.service - Devstack devstack@g-api.service Loaded: loaded (/etc/systemd/system/devstack@g-api.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:46:30 UTC; 5min ago Main PID: 114027 (uwsgi) Status: "uWSGI is ready" Tasks: 19 (limit: 77075) Memory: 2.0G (peak: 2.0G) CPU: 20.958s CGroup: /system.slice/system-devstack.slice/devstack@g-api.service ├─114027 "glance-apiuWSGI master" ├─114029 "glance-apiuWSGI worker 1" ├─114030 "glance-apiuWSGI worker 2" ├─114031 "glance-apiuWSGI worker 3" └─114032 "glance-apiuWSGI worker 4" Apr 02 14:51:01 np0000163014 devstack@g-api.service[114031]: DEBUG glance.api.middleware.version_negotiation [None req-003fc029-3e67-42e8-9146-7c0931622f3d demo demo] new path /v2/images/68002f80-9bd4-45bb-8ae2-cf108b868087 {{(pid=114031) process_request /opt/stack/glance/glance/api/middleware/version_negotiation.py:70}} Apr 02 14:51:01 np0000163014 devstack@g-api.service[114031]: [pid: 114031|app: 0|req: 6/25] 127.0.0.1 () {40 vars in 879 bytes} [Thu Apr 2 14:51:01 2026] GET /v2/images/68002f80-9bd4-45bb-8ae2-cf108b868087 => generated 993 bytes in 13 msecs (HTTP/1.1 200) 4 headers in 157 bytes (1 switches on core 0) Apr 02 14:51:01 np0000163014 devstack@g-api.service[114030]: DEBUG glance.api.middleware.version_negotiation [None req-dc4c60f2-7f42-4cfa-8ee7-9a7bdfbec57e demo demo] Determining version of request: GET /v2/schemas/image Accept: */* {{(pid=114030) process_request /opt/stack/glance/glance/api/middleware/version_negotiation.py:44}} Apr 02 14:51:01 np0000163014 devstack@g-api.service[114030]: DEBUG glance.api.middleware.version_negotiation [None req-dc4c60f2-7f42-4cfa-8ee7-9a7bdfbec57e demo demo] Using url versioning {{(pid=114030) process_request /opt/stack/glance/glance/api/middleware/version_negotiation.py:57}} Apr 02 14:51:01 np0000163014 devstack@g-api.service[114030]: DEBUG glance.api.middleware.version_negotiation [None req-dc4c60f2-7f42-4cfa-8ee7-9a7bdfbec57e demo demo] Matched version: v2 {{(pid=114030) process_request /opt/stack/glance/glance/api/middleware/version_negotiation.py:69}} Apr 02 14:51:01 np0000163014 devstack@g-api.service[114030]: DEBUG glance.api.middleware.version_negotiation [None req-dc4c60f2-7f42-4cfa-8ee7-9a7bdfbec57e demo demo] new path /v2/schemas/image {{(pid=114030) process_request /opt/stack/glance/glance/api/middleware/version_negotiation.py:70}} Apr 02 14:51:01 np0000163014 devstack@g-api.service[114030]: [pid: 114030|app: 0|req: 8/26] 127.0.0.1 () {40 vars in 819 bytes} [Thu Apr 2 14:51:01 2026] GET /v2/schemas/image => generated 6083 bytes in 5 msecs (HTTP/1.1 200) 4 headers in 158 bytes (1 switches on core 0) Apr 02 14:51:02 np0000163014 devstack@g-api.service[114030]: DEBUG dbcounter [-] [114030] Writing DB stats glance:SELECT=1 {{(pid=114030) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:06 np0000163014 devstack@g-api.service[114032]: DEBUG dbcounter [-] [114032] Writing DB stats glance:SELECT=4 {{(pid=114032) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:11 np0000163014 devstack@g-api.service[114031]: DEBUG dbcounter [-] [114031] Writing DB stats glance:SELECT=5 {{(pid=114031) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@keystone.service - Devstack devstack@keystone.service Loaded: loaded (/etc/systemd/system/devstack@keystone.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:39:26 UTC; 12min ago Main PID: 65167 (uwsgi) Status: "uWSGI is ready" Tasks: 9 (limit: 77075) Memory: 453.7M (peak: 454.3M) CPU: 1min 24.616s CGroup: /system.slice/system-devstack.slice/devstack@keystone.service ├─65167 "keystoneuWSGI master" ├─65168 "keystoneuWSGI worker 1" ├─65169 "keystoneuWSGI worker 2" ├─65170 "keystoneuWSGI worker 3" └─65171 "keystoneuWSGI worker 4" Apr 02 14:51:12 np0000163014 devstack@keystone.service[65169]: DEBUG keystone.server.flask.request_processing.req_logging [None req-45841de7-4f74-4ab9-892b-2545fc99dbd7 None None] SCRIPT_NAME: `/identity` {{(pid=65169) log_request_info /opt/stack/keystone/keystone/server/flask/request_processing/req_logging.py:27}} Apr 02 14:51:12 np0000163014 devstack@keystone.service[65169]: DEBUG keystone.server.flask.request_processing.req_logging [None req-45841de7-4f74-4ab9-892b-2545fc99dbd7 None None] PATH_INFO: `/v3/auth/tokens` {{(pid=65169) log_request_info /opt/stack/keystone/keystone/server/flask/request_processing/req_logging.py:28}} Apr 02 14:51:12 np0000163014 devstack@keystone.service[65169]: WARNING keystone.common.password_hashing [None req-45841de7-4f74-4ab9-892b-2545fc99dbd7 None None] Truncating password to algorithm specific maximum length 72 characters. Apr 02 14:51:12 np0000163014 devstack@keystone.service[65169]: DEBUG keystone.auth.core [None req-45841de7-4f74-4ab9-892b-2545fc99dbd7 None None] MFA Rules not processed for user `013b00565eff483799e86099bedaee7a`. Rule list: `[]` (Enabled: `True`). {{(pid=65169) check_auth_methods_against_rules /opt/stack/keystone/keystone/auth/core.py:476}} Apr 02 14:51:12 np0000163014 devstack@keystone.service[65169]: DEBUG keystone.common.fernet_utils [None req-45841de7-4f74-4ab9-892b-2545fc99dbd7 None None] Loaded 2 Fernet keys from /etc/keystone/fernet-keys/, but `[fernet_tokens] max_active_keys = 3`; perhaps there have not been enough key rotations to reach `max_active_keys` yet? {{(pid=65169) load_keys /opt/stack/keystone/keystone/common/fernet_utils.py:297}} Apr 02 14:51:12 np0000163014 devstack@keystone.service[65169]: [pid: 65169|app: 0|req: 424/1695] 199.204.45.109 () {66 vars in 1153 bytes} [Thu Apr 2 14:51:12 2026] POST /identity/v3/auth/tokens => generated 4527 bytes in 23 msecs (HTTP/1.1 201) 6 headers in 385 bytes (1 switches on core 0) Apr 02 14:51:21 np0000163014 devstack@keystone.service[65168]: DEBUG dbcounter [-] [65168] Writing DB stats keystone:SELECT=63,keystone:UPDATE=1,keystone:INSERT=2 {{(pid=65168) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:22 np0000163014 devstack@keystone.service[65171]: DEBUG dbcounter [-] [65171] Writing DB stats keystone:SELECT=90,keystone:INSERT=2,keystone:UPDATE=1 {{(pid=65171) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:22 np0000163014 devstack@keystone.service[65170]: DEBUG dbcounter [-] [65170] Writing DB stats keystone:SELECT=42 {{(pid=65170) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:22 np0000163014 devstack@keystone.service[65169]: DEBUG dbcounter [-] [65169] Writing DB stats keystone:SELECT=103,keystone:INSERT=4,keystone:UPDATE=1 {{(pid=65169) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@m-api.service - Devstack devstack@m-api.service Loaded: loaded (/etc/systemd/system/devstack@m-api.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:49:53 UTC; 2min 0s ago Main PID: 121134 (uwsgi) Status: "uWSGI is ready" Tasks: 13 (limit: 77075) Memory: 428.8M (peak: 429.3M) CPU: 10.736s CGroup: /system.slice/system-devstack.slice/devstack@m-api.service ├─121134 "manila-apiuWSGI master" ├─121135 "manila-apiuWSGI worker 1" ├─121136 "manila-apiuWSGI worker 2" ├─121137 "manila-apiuWSGI worker 3" └─121138 "manila-apiuWSGI worker 4" Apr 02 14:50:03 np0000163014 devstack@m-api.service[121136]: INFO manila.api.openstack.wsgi [None req-35c3603a-1dc7-4e2e-ba91-786b09f0847a admin admin] https://199.204.45.109/share/v2/types returned with HTTP 200 Apr 02 14:50:03 np0000163014 devstack@m-api.service[121136]: [pid: 121136|app: 0|req: 2/9] 199.204.45.109 () {70 vars in 1329 bytes} [Thu Apr 2 14:50:03 2026] POST /share/v2/types => generated 715 bytes in 33 msecs (HTTP/1.1 200) 7 headers in 297 bytes (1 switches on core 0) Apr 02 14:50:03 np0000163014 devstack@m-api.service[121137]: INFO manila.api.openstack.wsgi [None req-b7e40c2f-60b3-41ad-bf65-df246506ec67 admin admin] POST https://199.204.45.109/share/v2/types Apr 02 14:50:03 np0000163014 devstack@m-api.service[121137]: DEBUG manila.api.openstack.wsgi [None req-b7e40c2f-60b3-41ad-bf65-df246506ec67 admin admin] Action: 'create', calling method: Controller.__getattribute__..version_select, body: {"share_type": {"name": "dhss_false", "share_type_access:is_public": true, "extra_specs": {"snapshot_support": "True", "create_share_from_snapshot_support": "True", "driver_handles_share_servers": false}}} {{(pid=121137) _process_stack /opt/stack/manila/manila/api/openstack/wsgi.py:792}} Apr 02 14:50:03 np0000163014 devstack@m-api.service[121137]: INFO manila.api.openstack.wsgi [None req-b7e40c2f-60b3-41ad-bf65-df246506ec67 admin admin] https://199.204.45.109/share/v2/types returned with HTTP 200 Apr 02 14:50:03 np0000163014 devstack@m-api.service[121137]: [pid: 121137|app: 0|req: 3/10] 199.204.45.109 () {70 vars in 1329 bytes} [Thu Apr 2 14:50:03 2026] POST /share/v2/types => generated 721 bytes in 29 msecs (HTTP/1.1 200) 7 headers in 297 bytes (1 switches on core 0) Apr 02 14:50:12 np0000163014 devstack@m-api.service[121138]: DEBUG dbcounter [-] [121138] Writing DB stats manila:SELECT=2 {{(pid=121138) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:12 np0000163014 devstack@m-api.service[121135]: DEBUG dbcounter [-] [121135] Writing DB stats manila:INSERT=6,manila:SELECT=4 {{(pid=121135) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:13 np0000163014 devstack@m-api.service[121136]: DEBUG dbcounter [-] [121136] Writing DB stats manila:SELECT=2,manila:INSERT=4 {{(pid=121136) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:13 np0000163014 devstack@m-api.service[121137]: DEBUG dbcounter [-] [121137] Writing DB stats manila:SELECT=2,manila:INSERT=4 {{(pid=121137) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@m-dat.service - Devstack devstack@m-dat.service Loaded: loaded (/etc/systemd/system/devstack@m-dat.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:50:42 UTC; 1min 12s ago Main PID: 127317 (manila-data) Tasks: 1 (limit: 77075) Memory: 90.9M (peak: 91.2M) CPU: 1.560s CGroup: /system.slice/system-devstack.slice/devstack@m-dat.service └─127317 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-data --config-file /etc/manila/manila.conf Apr 02 14:50:43 np0000163014 manila-data[127317]: INFO manila.service [-] Starting manila-data node (version 22.1.0) Apr 02 14:50:43 np0000163014 manila-data[127317]: DEBUG oslo_db.api [None req-17e10dba-7ed3-4c26-9feb-4cf3bcb7d472 None None] Loading backend 'sqlalchemy' from 'manila.db.sqlalchemy.api' {{(pid=127317) _load_backend /opt/stack/data/venv/lib/python3.12/site-packages/oslo_db/api.py:259}} Apr 02 14:50:43 np0000163014 manila-data[127317]: INFO dbcounter [None req-17e10dba-7ed3-4c26-9feb-4cf3bcb7d472 None None] Registered counter for database manila Apr 02 14:50:43 np0000163014 manila-data[127317]: DEBUG oslo_db.sqlalchemy.engines [None req-17e10dba-7ed3-4c26-9feb-4cf3bcb7d472 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_ENGINE_SUBSTITUTION {{(pid=127317) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.12/site-packages/oslo_db/sqlalchemy/engines.py:325}} Apr 02 14:50:43 np0000163014 manila-data[127317]: DEBUG dbcounter [-] [127317] Writer thread running {{(pid=127317) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:102}} Apr 02 14:50:43 np0000163014 manila-data[127317]: DEBUG manila.service [None req-17e10dba-7ed3-4c26-9feb-4cf3bcb7d472 None None] Creating RPC server for service manila-data. {{(pid=127317) start /opt/stack/manila/manila/service.py:159}} Apr 02 14:50:53 np0000163014 manila-data[127317]: DEBUG dbcounter [-] [127317] Writing DB stats manila:SELECT=3,manila:INSERT=1 {{(pid=127317) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:31 np0000163014 manila-data[127317]: DEBUG oslo_service.periodic_task [None req-f364e2ab-e54a-4080-b286-34b10f78609f None None] Running periodic task DataManager.create_backup_continue {{(pid=127317) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:31 np0000163014 manila-data[127317]: DEBUG oslo_service.periodic_task [None req-f364e2ab-e54a-4080-b286-34b10f78609f None None] Running periodic task DataManager.restore_backup_continue {{(pid=127317) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:53 np0000163014 manila-data[127317]: DEBUG dbcounter [-] [127317] Writing DB stats manila:SELECT=27,manila:UPDATE=6 {{(pid=127317) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@m-sch.service - Devstack devstack@m-sch.service Loaded: loaded (/etc/systemd/system/devstack@m-sch.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:50:39 UTC; 1min 14s ago Main PID: 126741 (manila-schedule) Tasks: 1 (limit: 77075) Memory: 95.8M (peak: 96.3M) CPU: 1.655s CGroup: /system.slice/system-devstack.slice/devstack@m-sch.service └─126741 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-scheduler --config-file /etc/manila/manila.conf Apr 02 14:50:43 np0000163014 manila-scheduler[126741]: DEBUG manila.scheduler.host_manager [None req-a30599b4-0260-4272-9724-d138fbdbdb31 None None] Received share service update from np0000163014@generic: {'share_backend_name': 'GENERIC', 'driver_handles_share_servers': True, 'vendor_name': 'Open Source', 'driver_version': '1.0', 'storage_protocol': 'NFS_CIFS', 'total_capacity_gb': 'unknown', 'free_capacity_gb': 'unknown', 'reserved_percentage': 0, 'reserved_snapshot_percentage': 0, 'reserved_share_extend_percentage': 0, 'qos': False, 'pools': None, 'snapshot_support': True, 'create_share_from_snapshot_support': True, 'revert_to_snapshot_support': False, 'mount_snapshot_support': False, 'replication_domain': None, 'filter_function': None, 'goodness_function': None, 'security_service_update_support': False, 'network_allocation_update_support': False, 'share_server_multiple_subnet_support': False, 'mount_point_name_support': False, 'share_replicas_migration_support': False, 'encryption_support': None, 'qos_type_support': False, 'max_shares_per_share_server': -1, 'max_share_server_size': -1, 'share_group_stats': {'consistent_snapshot_support': None}, 'ipv4_support': True, 'ipv6_support': False, 'server_pools_mapping': {}} {{(pid=126741) update_service_capabilities /opt/stack/manila/manila/scheduler/host_manager.py:644}} Apr 02 14:50:51 np0000163014 manila-scheduler[126741]: DEBUG dbcounter [-] [126741] Writing DB stats manila:SELECT=2,manila:INSERT=1 {{(pid=126741) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:12 np0000163014 manila-scheduler[126741]: DEBUG oslo_service.periodic_task [None req-ea810d44-e3e5-455a-8f94-7eb265b56f6c None None] Running periodic task SchedulerManager._expire_reservations {{(pid=126741) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:12 np0000163014 manila-scheduler[126741]: DEBUG oslo_service.periodic_task [None req-ea810d44-e3e5-455a-8f94-7eb265b56f6c None None] Running periodic task SchedulerManager._clean_expired_messages {{(pid=126741) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:12 np0000163014 manila-scheduler[126741]: DEBUG manila.coordination [None req-ea810d44-e3e5-455a-8f94-7eb265b56f6c None None] Lock "locked-clean-expired-messages" acquired by "_clean_expired_messages". {{(pid=126741) _synchronized /opt/stack/manila/manila/coordination.py:199}} Apr 02 14:51:12 np0000163014 manila-scheduler[126741]: INFO manila.message.api [None req-ea810d44-e3e5-455a-8f94-7eb265b56f6c None None] Deleted 0 expired messages. Apr 02 14:51:12 np0000163014 manila-scheduler[126741]: DEBUG oslo_service.periodic_task [None req-ea810d44-e3e5-455a-8f94-7eb265b56f6c None None] Running periodic task SchedulerManager._mark_services_as_down {{(pid=126741) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:12 np0000163014 manila-scheduler[126741]: DEBUG manila.coordination [None req-ea810d44-e3e5-455a-8f94-7eb265b56f6c None None] Lock "locked-mark-services-as-down" acquired by "_mark_services_as_down". {{(pid=126741) _synchronized /opt/stack/manila/manila/coordination.py:199}} Apr 02 14:51:40 np0000163014 manila-scheduler[126741]: DEBUG manila.scheduler.host_manager [None req-d3533d41-1217-4721-a5f7-af2405ce2003 None None] Received share service update from np0000163014@generic: {'share_backend_name': 'GENERIC', 'driver_handles_share_servers': True, 'vendor_name': 'Open Source', 'driver_version': '1.0', 'storage_protocol': 'NFS_CIFS', 'total_capacity_gb': 'unknown', 'free_capacity_gb': 'unknown', 'reserved_percentage': 0, 'reserved_snapshot_percentage': 0, 'reserved_share_extend_percentage': 0, 'qos': False, 'pools': None, 'snapshot_support': True, 'create_share_from_snapshot_support': True, 'revert_to_snapshot_support': False, 'mount_snapshot_support': False, 'replication_domain': None, 'filter_function': None, 'goodness_function': None, 'security_service_update_support': False, 'network_allocation_update_support': False, 'share_server_multiple_subnet_support': False, 'mount_point_name_support': False, 'share_replicas_migration_support': False, 'encryption_support': None, 'qos_type_support': False, 'max_shares_per_share_server': -1, 'max_share_server_size': -1, 'share_group_stats': {'consistent_snapshot_support': None}, 'ipv4_support': True, 'ipv6_support': False, 'server_pools_mapping': {}} {{(pid=126741) update_service_capabilities /opt/stack/manila/manila/scheduler/host_manager.py:644}} Apr 02 14:51:51 np0000163014 manila-scheduler[126741]: DEBUG dbcounter [-] [126741] Writing DB stats manila:SELECT=30,manila:UPDATE=7,manila:DELETE=1 {{(pid=126741) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@m-shr.service - Devstack devstack@m-shr.service Loaded: loaded (/etc/systemd/system/devstack@m-shr.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:50:37 UTC; 1min 16s ago Main PID: 126208 (manila-share) Tasks: 2 (limit: 77075) Memory: 158.4M (peak: 180.1M) CPU: 5.732s CGroup: /system.slice/system-devstack.slice/devstack@m-shr.service ├─126208 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-share --config-file /etc/manila/manila.conf └─126626 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-share --config-file /etc/manila/manila.conf Apr 02 14:51:40 np0000163014 manila-share[126626]: DEBUG manila.share.manager [None req-d3533d41-1217-4721-a5f7-af2405ce2003 None None] Checking for shares in 'deferred_deleting' status to process their deletion. {{(pid=126626) do_deferred_share_deletion /opt/stack/manila/manila/share/manager.py:3911}} Apr 02 14:51:40 np0000163014 manila-share[126626]: DEBUG oslo_service.periodic_task [None req-d3533d41-1217-4721-a5f7-af2405ce2003 None None] Running periodic task ShareManager.do_deferred_snapshot_deletion {{(pid=126626) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:40 np0000163014 manila-share[126626]: DEBUG manila.share.manager [None req-d3533d41-1217-4721-a5f7-af2405ce2003 None None] Checking for snapshots in 'deferred_deleting' status to process their deletion. {{(pid=126626) do_deferred_snapshot_deletion /opt/stack/manila/manila/share/manager.py:4254}} Apr 02 14:51:40 np0000163014 manila-share[126626]: DEBUG oslo_service.periodic_task [None req-d3533d41-1217-4721-a5f7-af2405ce2003 None None] Running periodic task ShareManager._report_driver_status {{(pid=126626) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:40 np0000163014 manila-share[126626]: INFO manila.share.manager [None req-d3533d41-1217-4721-a5f7-af2405ce2003 None None] Updating share status Apr 02 14:51:40 np0000163014 manila-share[126626]: DEBUG manila.share.driver [None req-d3533d41-1217-4721-a5f7-af2405ce2003 None None] Updating share stats. {{(pid=126626) _update_share_stats /opt/stack/manila/manila/share/driver.py:1336}} Apr 02 14:51:40 np0000163014 manila-share[126626]: DEBUG oslo_service.periodic_task [None req-d3533d41-1217-4721-a5f7-af2405ce2003 None None] Running periodic task ShareManager.create_backup_continue {{(pid=126626) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:40 np0000163014 manila-share[126626]: DEBUG oslo_service.periodic_task [None req-d3533d41-1217-4721-a5f7-af2405ce2003 None None] Running periodic task ShareManager.restore_backup_continue {{(pid=126626) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:40 np0000163014 manila-share[126626]: DEBUG oslo_service.periodic_task [None req-d3533d41-1217-4721-a5f7-af2405ce2003 None None] Running periodic task ShareManager.periodic_share_status_update {{(pid=126626) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:40 np0000163014 manila-share[126626]: DEBUG manila.share.manager [None req-d3533d41-1217-4721-a5f7-af2405ce2003 None None] Updating status of share instances. {{(pid=126626) periodic_share_status_update /opt/stack/manila/manila/share/manager.py:5832}} ● devstack@magnum-api.service - Devstack devstack@magnum-api.service Loaded: loaded (/etc/systemd/system/devstack@magnum-api.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:47:56 UTC; 3min 57s ago Main PID: 118635 (uwsgi) Status: "uWSGI is ready" Tasks: 42 (limit: 77075) Memory: 461.4M (peak: 461.9M) CPU: 10.919s CGroup: /system.slice/system-devstack.slice/devstack@magnum-api.service ├─118635 "magnum-apiuWSGI master" ├─118636 "magnum-apiuWSGI worker 1" ├─118637 "magnum-apiuWSGI worker 2" ├─118638 "magnum-apiuWSGI worker 3" └─118639 "magnum-apiuWSGI worker 4" Apr 02 14:51:03 np0000163014 devstack@magnum-api.service[118637]: DEBUG oslo_db.sqlalchemy.engines [None req-100e4f98-31ee-47cc-b4d0-c527cb5d770d demo demo] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_ENGINE_SUBSTITUTION {{(pid=118637) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.12/site-packages/oslo_db/sqlalchemy/engines.py:325}} Apr 02 14:51:03 np0000163014 devstack@magnum-api.service[118637]: DEBUG dbcounter [-] [118637] Writer thread running {{(pid=118637) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:102}} Apr 02 14:51:03 np0000163014 devstack@magnum-api.service[118637]: [pid: 118637|app: 0|req: 1/4] 199.204.45.109 () {68 vars in 1382 bytes} [Thu Apr 2 14:51:03 2026] GET /container-infra/v1/clusters/k8s-cluster => generated 1309 bytes in 439 msecs (HTTP/1.1 200) 8 headers in 397 bytes (1 switches on core 0) Apr 02 14:51:05 np0000163014 devstack@magnum-api.service[118639]: [pid: 118639|app: 0|req: 2/5] 199.204.45.109 () {68 vars in 1382 bytes} [Thu Apr 2 14:51:05 2026] GET /container-infra/v1/clusters/k8s-cluster => generated 1340 bytes in 215 msecs (HTTP/1.1 200) 8 headers in 397 bytes (1 switches on core 0) Apr 02 14:51:08 np0000163014 devstack@magnum-api.service[118636]: DEBUG dbcounter [-] [118636] Writing DB stats magnum:SELECT=1 {{(pid=118636) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:11 np0000163014 devstack@magnum-api.service[118638]: DEBUG dbcounter [-] [118638] Writing DB stats magnum:SELECT=6 {{(pid=118638) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:12 np0000163014 devstack@magnum-api.service[118636]: [pid: 118636|app: 0|req: 2/6] 199.204.45.109 () {68 vars in 1382 bytes} [Thu Apr 2 14:51:11 2026] GET /container-infra/v1/clusters/k8s-cluster => generated 1464 bytes in 194 msecs (HTTP/1.1 200) 8 headers in 397 bytes (1 switches on core 0) Apr 02 14:51:13 np0000163014 devstack@magnum-api.service[118637]: DEBUG dbcounter [-] [118637] Writing DB stats magnum:SELECT=10 {{(pid=118637) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:15 np0000163014 devstack@magnum-api.service[118639]: DEBUG dbcounter [-] [118639] Writing DB stats magnum:INSERT=1,magnum:SELECT=10 {{(pid=118639) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:22 np0000163014 devstack@magnum-api.service[118636]: DEBUG dbcounter [-] [118636] Writing DB stats magnum:SELECT=11 {{(pid=118636) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@magnum-cond.service - Devstack devstack@magnum-cond.service Loaded: loaded (/etc/systemd/system/devstack@magnum-cond.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:47:59 UTC; 3min 55s ago Main PID: 119227 (magnum-conducto) Tasks: 69 (limit: 77075) Memory: 393.3M (peak: 394.8M) CPU: 7.030s CGroup: /system.slice/system-devstack.slice/devstack@magnum-cond.service ├─119227 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119467 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119468 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119469 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119470 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119471 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119472 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119473 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119474 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119475 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119476 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119477 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119478 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119479 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119480 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─119481 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor └─119482 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor Apr 02 14:51:29 np0000163014 magnum-conductor[119471]: DEBUG rustls.common_state [-] Sending warning alert CloseNotify {{(pid=119471) None /opt/stack/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/rustls-0.23.31/src/common_state.rs:584}} Apr 02 14:51:29 np0000163014 magnum-conductor[119471]: DEBUG rustls.common_state [-] Sending warning alert CloseNotify {{(pid=119471) None /opt/stack/.cargo/registry/src/index.crates.io-1949cf8c6b5b557f/rustls-0.23.31/src/common_state.rs:584}} Apr 02 14:51:30 np0000163014 magnum-conductor[119227]: DEBUG oslo_service.periodic_task [-] Running periodic task MagnumPeriodicTasks.sync_cluster_status {{(pid=119227) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:30 np0000163014 magnum-conductor[119227]: DEBUG magnum.service.periodic [None req-f64fffbb-27eb-4e7a-ae0f-fc42cf6aa666 None None] Starting to sync up cluster status {{(pid=119227) sync_cluster_status /opt/stack/magnum/magnum/service/periodic.py:182}} Apr 02 14:51:40 np0000163014 magnum-conductor[119227]: DEBUG dbcounter [-] [119227] Writing DB stats magnum:SELECT=2 {{(pid=119227) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:40 np0000163014 magnum-conductor[119227]: DEBUG oslo_service.periodic_task [-] Running periodic task MagnumPeriodicTasks.sync_cluster_status {{(pid=119227) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:40 np0000163014 magnum-conductor[119227]: DEBUG magnum.service.periodic [None req-7dda8d35-d39b-4027-9e74-37b27f90fcc2 None None] Starting to sync up cluster status {{(pid=119227) sync_cluster_status /opt/stack/magnum/magnum/service/periodic.py:182}} Apr 02 14:51:50 np0000163014 magnum-conductor[119227]: DEBUG dbcounter [-] [119227] Writing DB stats magnum:SELECT=1 {{(pid=119227) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:50 np0000163014 magnum-conductor[119227]: DEBUG oslo_service.periodic_task [-] Running periodic task MagnumPeriodicTasks.sync_cluster_status {{(pid=119227) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:50 np0000163014 magnum-conductor[119227]: DEBUG magnum.service.periodic [None req-b225c356-cfe9-4caa-bcaf-2642e2f1b5e3 None None] Starting to sync up cluster status {{(pid=119227) sync_cluster_status /opt/stack/magnum/magnum/service/periodic.py:182}} ● devstack@memory_tracker.service - Devstack devstack@memory_tracker.service Loaded: loaded (/etc/systemd/system/devstack@memory_tracker.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:39:07 UTC; 12min ago Main PID: 62827 (memory_tracker.) Tasks: 2 (limit: 77075) Memory: 1.7M (peak: 9.4M) CPU: 2.510s CGroup: /system.slice/system-devstack.slice/devstack@memory_tracker.service ├─ 62827 /bin/bash /opt/stack/devstack/tools/memory_tracker.sh └─130003 sleep 20 Apr 02 14:51:49 np0000163014 memory_tracker.sh[130002]: 23560 0.0 692 23506 00:00:00 1 do_sys_pause /pause Apr 02 14:51:49 np0000163014 memory_tracker.sh[130002]: 23568 0.0 692 23523 00:00:00 1 do_sys_pause /pause Apr 02 14:51:49 np0000163014 memory_tracker.sh[130002]: 24432 0.0 692 24412 00:00:00 1 do_sys_pause /pause Apr 02 14:51:49 np0000163014 memory_tracker.sh[130002]: 24759 0.0 692 24739 00:00:00 1 do_sys_pause /pause Apr 02 14:51:49 np0000163014 memory_tracker.sh[130002]: 21844 0.0 688 21818 00:00:00 1 do_sys_pause /pause Apr 02 14:51:49 np0000163014 memory_tracker.sh[130002]: 21907 0.0 688 21863 00:00:00 1 do_sys_pause /pause Apr 02 14:51:49 np0000163014 memory_tracker.sh[130002]: 23224 0.0 688 23188 00:00:00 1 do_sys_pause /pause Apr 02 14:51:49 np0000163014 memory_tracker.sh[130002]: 24363 0.0 688 24342 00:00:00 1 do_sys_pause /pause Apr 02 14:51:49 np0000163014 memory_tracker.sh[62827]: --- Apr 02 14:51:49 np0000163014 memory_tracker.sh[62827]: ]]] ● devstack@n-api-meta.service - Devstack devstack@n-api-meta.service Loaded: loaded (/etc/systemd/system/devstack@n-api-meta.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:46:00 UTC; 5min ago Main PID: 107210 (uwsgi) Status: "uWSGI is ready" Tasks: 30 (limit: 77075) Memory: 424.8M (peak: 426.1M) CPU: 9.231s CGroup: /system.slice/system-devstack.slice/devstack@n-api-meta.service ├─107210 "nova-api-metauWSGI master" ├─107211 "nova-api-metauWSGI worker 1" ├─107212 "nova-api-metauWSGI worker 2" ├─107213 "nova-api-metauWSGI worker 3" ├─107214 "nova-api-metauWSGI worker 4" └─107215 "nova-api-metauWSGI http 1" Apr 02 14:46:12 np0000163014 devstack@n-api-meta.service[107212]: DEBUG dbcounter [-] [107212] Writing DB stats nova_cell0:SELECT=2,nova_cell0:INSERT=1 {{(pid=107212) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:46:12 np0000163014 devstack@n-api-meta.service[107213]: DEBUG dbcounter [-] [107213] Writing DB stats nova_cell0:SELECT=1 {{(pid=107213) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:46:12 np0000163014 devstack@n-api-meta.service[107213]: DEBUG dbcounter [-] [107213] Writing DB stats nova_cell1:SELECT=1 {{(pid=107213) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:46:12 np0000163014 devstack@n-api-meta.service[107214]: DEBUG dbcounter [-] [107214] Writing DB stats nova_cell0:SELECT=1 {{(pid=107214) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:46:12 np0000163014 devstack@n-api-meta.service[107214]: DEBUG dbcounter [-] [107214] Writing DB stats nova_cell1:SELECT=1 {{(pid=107214) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:46:12 np0000163014 devstack@n-api-meta.service[107211]: DEBUG dbcounter [-] [107211] Writing DB stats nova_cell1:SELECT=1 {{(pid=107211) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:46:12 np0000163014 devstack@n-api-meta.service[107211]: DEBUG dbcounter [-] [107211] Writing DB stats nova_cell0:SELECT=1 {{(pid=107211) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:46:12 np0000163014 devstack@n-api-meta.service[107213]: DEBUG dbcounter [-] [107213] Writing DB stats nova_cell0:SELECT=1 {{(pid=107213) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:46:12 np0000163014 devstack@n-api-meta.service[107214]: DEBUG dbcounter [-] [107214] Writing DB stats nova_cell0:SELECT=1 {{(pid=107214) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:46:12 np0000163014 devstack@n-api-meta.service[107211]: DEBUG dbcounter [-] [107211] Writing DB stats nova_cell0:SELECT=1 {{(pid=107211) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@n-api.service - Devstack devstack@n-api.service Loaded: loaded (/etc/systemd/system/devstack@n-api.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:45:01 UTC; 6min ago Main PID: 98732 (uwsgi) Status: "uWSGI is ready" Tasks: 34 (limit: 77075) Memory: 528.6M (peak: 529.0M) CPU: 18.103s CGroup: /system.slice/system-devstack.slice/devstack@n-api.service ├─98732 "nova-apiuWSGI master" ├─98733 "nova-apiuWSGI worker 1" ├─98734 "nova-apiuWSGI worker 2" ├─98735 "nova-apiuWSGI worker 3" └─98736 "nova-apiuWSGI worker 4" Apr 02 14:51:06 np0000163014 devstack@n-api.service[98734]: DEBUG nova.quota [None req-75035176-3589-4a96-8f58-c9b32860fb69 demo demo] Getting quotas for project f3c2a2cf652e4eed8bfe6f00c3630d38. Resources: {'server_groups'} {{(pid=98734) _get_quotas /opt/stack/nova/nova/quota.py:392}} Apr 02 14:51:06 np0000163014 devstack@n-api.service[98734]: DEBUG nova.quota [None req-75035176-3589-4a96-8f58-c9b32860fb69 demo demo] Getting quotas for user bb8ec9c5ec0b4bc6b2ab8e5afb1e53a1 and project f3c2a2cf652e4eed8bfe6f00c3630d38. Resources: {'server_groups'} {{(pid=98734) _get_quotas /opt/stack/nova/nova/quota.py:382}} Apr 02 14:51:06 np0000163014 devstack@n-api.service[98734]: INFO nova.api.openstack.requestlog [None req-75035176-3589-4a96-8f58-c9b32860fb69 demo demo] 199.204.45.109 "POST /compute/v2.1/os-server-groups" status: 200 len: 265 microversion: 2.15 time: 0.106362 Apr 02 14:51:06 np0000163014 devstack@n-api.service[98734]: [pid: 98734|app: 0|req: 18/72] 199.204.45.109 () {70 vars in 1365 bytes} [Thu Apr 2 14:51:06 2026] POST /compute/v2.1/os-server-groups => generated 265 bytes in 108 msecs (HTTP/1.1 200) 9 headers in 359 bytes (1 switches on core 0) Apr 02 14:51:16 np0000163014 devstack@n-api.service[98735]: DEBUG dbcounter [-] [98735] Writing DB stats nova_api:SELECT=4 {{(pid=98735) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:16 np0000163014 devstack@n-api.service[98736]: DEBUG dbcounter [-] [98736] Writing DB stats nova_cell0:SELECT=4 {{(pid=98736) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:16 np0000163014 devstack@n-api.service[98736]: DEBUG dbcounter [-] [98736] Writing DB stats nova_api:SELECT=16,nova_api:INSERT=2 {{(pid=98736) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:16 np0000163014 devstack@n-api.service[98733]: DEBUG dbcounter [-] [98733] Writing DB stats nova_api:SELECT=4 {{(pid=98733) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:16 np0000163014 devstack@n-api.service[98734]: DEBUG dbcounter [-] [98734] Writing DB stats nova_cell0:SELECT=4 {{(pid=98734) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:16 np0000163014 devstack@n-api.service[98734]: DEBUG dbcounter [-] [98734] Writing DB stats nova_api:SELECT=16,nova_api:INSERT=2 {{(pid=98734) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@n-cond-cell1.service - Devstack devstack@n-cond-cell1.service Loaded: loaded (/etc/systemd/system/devstack@n-cond-cell1.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:46:07 UTC; 5min ago Main PID: 109228 (nova-conductor:) Tasks: 84 (limit: 77075) Memory: 224.6M (peak: 225.3M) CPU: 5.343s CGroup: /system.slice/system-devstack.slice/devstack@n-cond-cell1.service ├─109228 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova_cell1.conf]" ├─109927 "nova-conductor: ServiceWrapper worker(0)" ├─109935 "nova-conductor: ServiceWrapper worker(1)" ├─109944 "nova-conductor: ServiceWrapper worker(2)" └─109954 "nova-conductor: ServiceWrapper worker(3)" Apr 02 14:51:34 np0000163014 nova-conductor[109954]: DEBUG dbcounter [-] [109954] Writing DB stats nova_cell1:SELECT=4,nova_cell1:DELETE=1,nova_cell1:UPDATE=1 {{(pid=109954) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:35 np0000163014 nova-conductor[109944]: DEBUG dbcounter [-] [109944] Writing DB stats nova_cell1:SELECT=5,nova_cell1:UPDATE=1 {{(pid=109944) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:35 np0000163014 nova-conductor[109954]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.99 seconds {{(pid=109954) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:35 np0000163014 nova-conductor[109944]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.99 seconds {{(pid=109944) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:45 np0000163014 nova-conductor[109927]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.89 seconds {{(pid=109927) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:45 np0000163014 nova-conductor[109954]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.92 seconds {{(pid=109954) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:45 np0000163014 nova-conductor[109944]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.93 seconds {{(pid=109944) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:45 np0000163014 nova-conductor[109935]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.88 seconds {{(pid=109935) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:49 np0000163014 nova-conductor[109935]: DEBUG oslo.service.backend._threading.loopingcall [None req-2d0e1c3a-8ded-440f-b5dd-5d9853959d4f None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=109935) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:52 np0000163014 nova-conductor[109927]: DEBUG oslo.service.backend._threading.loopingcall [None req-d09c5126-d665-43ce-8a4a-f61444a0d106 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=109927) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} ● devstack@n-cpu.service - Devstack devstack@n-cpu.service Loaded: loaded (/etc/systemd/system/devstack@n-cpu.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:46:11 UTC; 5min ago Main PID: 110381 (nova-compute) Tasks: 10 (limit: 77075) Memory: 140.4M (peak: 143.9M) CPU: 3.748s CGroup: /system.slice/system-devstack.slice/devstack@n-cpu.service └─110381 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/nova-compute --config-file /etc/nova/nova-cpu.conf Apr 02 14:51:19 np0000163014 nova-compute[110381]: DEBUG oslo_concurrency.lockutils [None req-da7ba3cc-3fa6-434c-ae4b-9cfd97d259fb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.150s {{(pid=110381) inner /opt/stack/data/venv/lib/python3.12/site-packages/oslo_concurrency/lockutils.py:538}} Apr 02 14:51:19 np0000163014 nova-compute[110381]: DEBUG oslo.service.backend._threading.loopingcall [None req-da7ba3cc-3fa6-434c-ae4b-9cfd97d259fb None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 1.48 seconds {{(pid=110381) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:21 np0000163014 nova-compute[110381]: DEBUG oslo_service.periodic_task [None req-da7ba3cc-3fa6-434c-ae4b-9cfd97d259fb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=110381) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:21 np0000163014 nova-compute[110381]: DEBUG oslo_service.periodic_task [None req-da7ba3cc-3fa6-434c-ae4b-9cfd97d259fb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=110381) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:21 np0000163014 nova-compute[110381]: DEBUG oslo_service.periodic_task [None req-da7ba3cc-3fa6-434c-ae4b-9cfd97d259fb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=110381) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:21 np0000163014 nova-compute[110381]: DEBUG oslo_service.periodic_task [None req-da7ba3cc-3fa6-434c-ae4b-9cfd97d259fb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=110381) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:21 np0000163014 nova-compute[110381]: DEBUG oslo_service.periodic_task [None req-da7ba3cc-3fa6-434c-ae4b-9cfd97d259fb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=110381) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:21 np0000163014 nova-compute[110381]: DEBUG oslo_service.periodic_task [None req-da7ba3cc-3fa6-434c-ae4b-9cfd97d259fb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=110381) run_periodic_tasks /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/periodic_task.py:210}} Apr 02 14:51:21 np0000163014 nova-compute[110381]: DEBUG nova.compute.manager [None req-da7ba3cc-3fa6-434c-ae4b-9cfd97d259fb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=110381) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11402}} Apr 02 14:51:21 np0000163014 nova-compute[110381]: DEBUG oslo.service.backend._threading.loopingcall [None req-da7ba3cc-3fa6-434c-ae4b-9cfd97d259fb None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 54.25 seconds {{(pid=110381) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} ● devstack@n-novnc-cell1.service - Devstack devstack@n-novnc-cell1.service Loaded: loaded (/etc/systemd/system/devstack@n-novnc-cell1.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:46:02 UTC; 5min ago Main PID: 107911 (nova-novncproxy) Tasks: 16 (limit: 77075) Memory: 106.2M (peak: 106.4M) CPU: 3.008s CGroup: /system.slice/system-devstack.slice/devstack@n-novnc-cell1.service └─107911 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/nova-novncproxy --config-file /etc/nova/nova_cell1.conf --web /opt/stack/novnc Apr 02 14:46:02 np0000163014 systemd[1]: Started devstack@n-novnc-cell1.service - Devstack devstack@n-novnc-cell1.service. Apr 02 14:46:04 np0000163014 nova-novncproxy[107911]: INFO nova.console.websocketproxy [-] WebSocket server settings: Apr 02 14:46:04 np0000163014 nova-novncproxy[107911]: INFO nova.console.websocketproxy [-]  - Listen on 0.0.0.0:6080 Apr 02 14:46:04 np0000163014 nova-novncproxy[107911]: INFO nova.console.websocketproxy [-]  - Web server (no directory listings). Web root: /opt/stack/novnc Apr 02 14:46:04 np0000163014 nova-novncproxy[107911]: INFO nova.console.websocketproxy [-]  - No SSL/TLS support (no cert file) Apr 02 14:46:04 np0000163014 nova-novncproxy[107911]: INFO nova.console.websocketproxy [-]  - proxying from 0.0.0.0:6080 to None:None ● devstack@n-sch.service - Devstack devstack@n-sch.service Loaded: loaded (/etc/systemd/system/devstack@n-sch.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:45:57 UTC; 5min ago Main PID: 106605 (nova-scheduler:) Tasks: 33 (limit: 77075) Memory: 197.1M (peak: 197.9M) CPU: 2.626s CGroup: /system.slice/system-devstack.slice/devstack@n-sch.service ├─106605 "nova-scheduler: master process [/opt/stack/data/venv/bin/nova-scheduler --config-file /etc/nova/nova.conf]" ├─107339 "nova-scheduler: ServiceWrapper worker(0)" ├─107345 "nova-scheduler: ServiceWrapper worker(1)" ├─107352 "nova-scheduler: ServiceWrapper worker(2)" └─107362 "nova-scheduler: ServiceWrapper worker(3)" Apr 02 14:50:15 np0000163014 nova-scheduler[107345]: DEBUG dbcounter [-] [107345] Writing DB stats nova_cell0:SELECT=1,nova_cell0:UPDATE=1 {{(pid=107345) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:15 np0000163014 nova-scheduler[107339]: DEBUG dbcounter [-] [107339] Writing DB stats nova_cell0:SELECT=1,nova_cell0:UPDATE=1 {{(pid=107339) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:15 np0000163014 nova-scheduler[107362]: DEBUG dbcounter [-] [107362] Writing DB stats nova_cell0:SELECT=1 {{(pid=107362) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:15 np0000163014 nova-scheduler[107352]: DEBUG dbcounter [-] [107352] Writing DB stats nova_cell0:SELECT=1 {{(pid=107352) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:32 np0000163014 nova-scheduler[107339]: DEBUG oslo.service.backend._threading.loopingcall [None req-ef1a0d54-2593-40bb-81e1-fa34bdb5d6da None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=107339) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:50:41 np0000163014 nova-scheduler[107362]: DEBUG oslo.service.backend._threading.loopingcall [None req-6451383f-17df-44d9-97c7-bb86fc301bd0 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=107362) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:00 np0000163014 nova-scheduler[107345]: DEBUG oslo.service.backend._threading.loopingcall [None req-8d60ac53-7f90-4221-832e-d09e65a5ef50 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=107345) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:11 np0000163014 nova-scheduler[107352]: DEBUG oslo.service.backend._threading.loopingcall [None req-59673e56-15a5-4d3a-b070-32c589abf2db None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=107352) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:32 np0000163014 nova-scheduler[107339]: DEBUG oslo.service.backend._threading.loopingcall [None req-ef1a0d54-2593-40bb-81e1-fa34bdb5d6da None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=107339) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:41 np0000163014 nova-scheduler[107362]: DEBUG oslo.service.backend._threading.loopingcall [None req-6451383f-17df-44d9-97c7-bb86fc301bd0 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=107362) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} ● devstack@n-super-cond.service - Devstack devstack@n-super-cond.service Loaded: loaded (/etc/systemd/system/devstack@n-super-cond.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:46:05 UTC; 5min ago Main PID: 108690 (nova-conductor:) Tasks: 33 (limit: 77075) Memory: 197.2M (peak: 198.3M) CPU: 2.773s CGroup: /system.slice/system-devstack.slice/devstack@n-super-cond.service ├─108690 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova.conf]" ├─109512 "nova-conductor: ServiceWrapper worker(0)" ├─109517 "nova-conductor: ServiceWrapper worker(1)" ├─109523 "nova-conductor: ServiceWrapper worker(2)" └─109534 "nova-conductor: ServiceWrapper worker(3)" Apr 02 14:50:23 np0000163014 nova-conductor[109517]: DEBUG dbcounter [-] [109517] Writing DB stats nova_cell0:SELECT=1,nova_cell0:UPDATE=1 {{(pid=109517) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:23 np0000163014 nova-conductor[109523]: DEBUG dbcounter [-] [109523] Writing DB stats nova_cell0:SELECT=1 {{(pid=109523) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:23 np0000163014 nova-conductor[109534]: DEBUG dbcounter [-] [109534] Writing DB stats nova_cell0:SELECT=1 {{(pid=109534) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:36 np0000163014 nova-conductor[109534]: DEBUG oslo.service.backend._threading.loopingcall [None req-2d5b4dc6-7f0e-4644-a50d-e0546866d67c None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=109534) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:50:48 np0000163014 nova-conductor[109523]: DEBUG oslo.service.backend._threading.loopingcall [None req-9526be07-aa3c-4599-9d5a-560651ef2ae3 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=109523) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:50:52 np0000163014 nova-conductor[109512]: DEBUG oslo.service.backend._threading.loopingcall [None req-22d1cb6a-d38f-45ce-abe0-5fb7d2956fc8 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=109512) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:50:57 np0000163014 nova-conductor[109517]: DEBUG oslo.service.backend._threading.loopingcall [None req-4501417d-2666-459f-890d-50a79fbcae3e None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=109517) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:36 np0000163014 nova-conductor[109534]: DEBUG oslo.service.backend._threading.loopingcall [None req-2d5b4dc6-7f0e-4644-a50d-e0546866d67c None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=109534) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:48 np0000163014 nova-conductor[109523]: DEBUG oslo.service.backend._threading.loopingcall [None req-9526be07-aa3c-4599-9d5a-560651ef2ae3 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=109523) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:52 np0000163014 nova-conductor[109512]: DEBUG oslo.service.backend._threading.loopingcall [None req-22d1cb6a-d38f-45ce-abe0-5fb7d2956fc8 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=109512) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} ● devstack@neutron-api.service - Devstack devstack@neutron-api.service Loaded: loaded (/etc/systemd/system/devstack@neutron-api.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:45:23 UTC; 6min ago Main PID: 102068 (uwsgi) Status: "uWSGI is ready" Tasks: 141 (limit: 77075) Memory: 722.3M (peak: 723.7M) CPU: 45.989s CGroup: /system.slice/system-devstack.slice/devstack@neutron-api.service ├─102068 "neutron-apiuWSGI master" ├─102069 "neutron-apiuWSGI worker 1" ├─102070 "neutron-apiuWSGI worker 2" ├─102071 "neutron-apiuWSGI worker 3" └─102072 "neutron-apiuWSGI worker 4" Apr 02 14:51:40 np0000163014 devstack@neutron-api.service[102069]: DEBUG dbcounter [-] [102069] Writing DB stats neutron:UPDATE=1,neutron:SELECT=1 {{(pid=102069) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:40 np0000163014 devstack@neutron-api.service[102070]: DEBUG dbcounter [-] [102070] Writing DB stats neutron:UPDATE=1,neutron:SELECT=1 {{(pid=102070) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:45 np0000163014 devstack@neutron-api.service[102072]: DEBUG futurist.periodics [-] Submitting periodic callback 'neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance.HashRingHealthCheckPeriodics.touch_hash_ring_node' {{(pid=102072) _process_scheduled /opt/stack/data/venv/lib/python3.12/site-packages/futurist/periodics.py:784}} Apr 02 14:51:45 np0000163014 devstack@neutron-api.service[102072]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [-] Touching Hash Ring node "290039c87a875ad097878f929f8bc7e7" from periodic health check thread {{(pid=102072) touch_hash_ring_node /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:1232}} Apr 02 14:51:45 np0000163014 devstack@neutron-api.service[102071]: DEBUG futurist.periodics [-] Submitting periodic callback 'neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance.HashRingHealthCheckPeriodics.touch_hash_ring_node' {{(pid=102071) _process_scheduled /opt/stack/data/venv/lib/python3.12/site-packages/futurist/periodics.py:784}} Apr 02 14:51:45 np0000163014 devstack@neutron-api.service[102071]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [-] Touching Hash Ring node "5b67e3a394625cb1b63d1c4d43b5f3d5" from periodic health check thread {{(pid=102071) touch_hash_ring_node /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:1232}} Apr 02 14:51:45 np0000163014 devstack@neutron-api.service[102069]: DEBUG futurist.periodics [-] Submitting periodic callback 'neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance.HashRingHealthCheckPeriodics.touch_hash_ring_node' {{(pid=102069) _process_scheduled /opt/stack/data/venv/lib/python3.12/site-packages/futurist/periodics.py:784}} Apr 02 14:51:45 np0000163014 devstack@neutron-api.service[102069]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [-] Touching Hash Ring node "eea2c844571557648f57620d655938d0" from periodic health check thread {{(pid=102069) touch_hash_ring_node /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:1232}} Apr 02 14:51:45 np0000163014 devstack@neutron-api.service[102070]: DEBUG futurist.periodics [-] Submitting periodic callback 'neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance.HashRingHealthCheckPeriodics.touch_hash_ring_node' {{(pid=102070) _process_scheduled /opt/stack/data/venv/lib/python3.12/site-packages/futurist/periodics.py:784}} Apr 02 14:51:45 np0000163014 devstack@neutron-api.service[102070]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [-] Touching Hash Ring node "061b6d8e2cf55230b4c38406daf7842a" from periodic health check thread {{(pid=102070) touch_hash_ring_node /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:1232}} ● devstack@neutron-ovn-maintenance-worker.service - Devstack devstack@neutron-ovn-maintenance-worker.service Loaded: loaded (/etc/systemd/system/devstack@neutron-ovn-maintenance-worker.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:45:30 UTC; 6min ago Main PID: 103560 (neutron-ovn-mai) Tasks: 35 (limit: 77075) Memory: 258.6M (peak: 259.0M) CPU: 5.864s CGroup: /system.slice/system-devstack.slice/devstack@neutron-ovn-maintenance-worker.service ├─103560 "neutron-ovn-maintenance-worker: master process [/opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" └─104364 "neutron-server: maintenance worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" Apr 02 14:50:53 np0000163014 neutron-ovn-maintenance-worker[104364]: INFO neutron.db.ovn_revision_numbers_db [None req-96b814e0-c298-45b1-b8fe-1a16f2aaf5d2 None None] Successfully bumped revision number for resource 64163d45-23b4-43df-af66-0873bea56e2f (type: networks) to 3 Apr 02 14:50:53 np0000163014 neutron-ovn-maintenance-worker[104364]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [None req-96b814e0-c298-45b1-b8fe-1a16f2aaf5d2 None None] Maintenance task: Fixing resource 9673545f-69af-49b8-baa9-83fd5797a4d3 (type: networks) at create/update {{(pid=104364) check_for_inconsistencies /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:396}} Apr 02 14:50:53 np0000163014 neutron-ovn-maintenance-worker[104364]: DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=0): CheckRevisionNumberCommand(_result=None, name=neutron-9673545f-69af-49b8-baa9-83fd5797a4d3, resource={'id': '9673545f-69af-49b8-baa9-83fd5797a4d3', 'name': 'public', 'tenant_id': '5d51f7b3c36d494ba6c66f17a427497e', 'project_id': '5d51f7b3c36d494ba6c66f17a427497e', 'admin_state_up': True, 'mtu': 1430, 'status': 'ACTIVE', 'subnets': ['40c224d9-0a3a-4e4a-9b31-93bb15494baf', '7369b3d0-529c-4126-aa3e-56cdceeb69ae'], 'standard_attr_id': 18, 'shared': False, 'availability_zone_hints': [], 'availability_zones': [], 'ipv4_address_scope': None, 'ipv6_address_scope': None, 'router:external': True, 'qinq': False, 'vlan_transparent': False, 'description': '', 'port_security_enabled': True, 'is_default': True, 'tags': [], 'created_at': '2026-04-02T14:45:44Z', 'updated_at': '2026-04-02T14:45:53Z', 'revision_number': 3, 'provider:network_type': 'flat', 'provider:physical_network': 'public', 'provider:segmentation_id': None}, resource_type=networks, if_exists=True) {{(pid=104364) do_commit /opt/stack/data/venv/lib/python3.12/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89}} Apr 02 14:50:53 np0000163014 neutron-ovn-maintenance-worker[104364]: DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=1): DbSetCommand(_result=None, table=Logical_Switch, record=neutron-9673545f-69af-49b8-baa9-83fd5797a4d3, col_values=(('external_ids', {'neutron:network_name': 'public', 'neutron:mtu': '1430', 'neutron:revision_number': '3', 'neutron:availability_zone_hints': '', 'neutron:provnet-network-type': 'flat', 'neutron:provnet-physical-network': 'public'}), ('other_config', {'mcast_snoop': 'false', 'mcast_flood_unregistered': 'false', 'vlan-passthru': 'false', 'fdb_age_threshold': '300', 'broadcast-arps-to-all-routers': 'true'})), if_exists=True) {{(pid=104364) do_commit /opt/stack/data/venv/lib/python3.12/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89}} Apr 02 14:50:53 np0000163014 neutron-ovn-maintenance-worker[104364]: DEBUG ovsdbapp.backend.ovs_idl.transaction [-] Running txn n=1 command(idx=2): SetLSwitchPortCommand(_result=None, lport=provnet-8556b1dc-5fdc-45a9-a252-4d811ec5cb42, external_ids_update=None, columns={'tag': []}, if_exists=True) {{(pid=104364) do_commit /opt/stack/data/venv/lib/python3.12/site-packages/ovsdbapp/backend/ovs_idl/transaction.py:89}} Apr 02 14:50:53 np0000163014 neutron-ovn-maintenance-worker[104364]: INFO neutron.db.ovn_revision_numbers_db [None req-96b814e0-c298-45b1-b8fe-1a16f2aaf5d2 None None] Successfully bumped revision number for resource 9673545f-69af-49b8-baa9-83fd5797a4d3 (type: networks) to 3 Apr 02 14:50:53 np0000163014 neutron-ovn-maintenance-worker[104364]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [None req-96b814e0-c298-45b1-b8fe-1a16f2aaf5d2 None None] Maintenance task: Fixing resource dccd65ab-21c1-493e-8fcd-d5b6cdd84a20 (type: security_groups) at create/update {{(pid=104364) check_for_inconsistencies /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/maintenance.py:396}} Apr 02 14:50:53 np0000163014 neutron-ovn-maintenance-worker[104364]: INFO neutron.db.ovn_revision_numbers_db [None req-96b814e0-c298-45b1-b8fe-1a16f2aaf5d2 None None] Successfully bumped revision number for resource dccd65ab-21c1-493e-8fcd-d5b6cdd84a20 (type: security_groups) to 9 Apr 02 14:50:53 np0000163014 neutron-ovn-maintenance-worker[104364]: INFO neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.maintenance [None req-96b814e0-c298-45b1-b8fe-1a16f2aaf5d2 None None] Maintenance task: Synchronization completed (took 0.18 seconds) Apr 02 14:51:03 np0000163014 neutron-ovn-maintenance-worker[104364]: DEBUG dbcounter [-] [104364] Writing DB stats neutron:SELECT=99,neutron:UPDATE=3 {{(pid=104364) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@neutron-periodic-workers.service - Devstack devstack@neutron-periodic-workers.service Loaded: loaded (/etc/systemd/system/devstack@neutron-periodic-workers.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:45:27 UTC; 6min ago Main PID: 103064 (neutron-periodi) Tasks: 23 (limit: 77075) Memory: 209.3M (peak: 209.6M) CPU: 3.165s CGroup: /system.slice/system-devstack.slice/devstack@neutron-periodic-workers.service ├─103064 "neutron-periodic-workers: master process [/opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" ├─103784 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" ├─103796 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" ├─103807 "neutron-server: Periodic worker for \"DbQuotaNoLockDriver\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" └─103819 "neutron-server: Periodic worker for \"L3_NAT_dbonly_mixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" Apr 02 14:51:03 np0000163014 neutron-periodic-workers[103796]: DEBUG neutron.db.agents_db [None req-57e5a4f8-ee4e-491e-8d78-54e1a95d5fc3 None None] Agent healthcheck: found 0 active agents {{(pid=103796) agent_health_check /opt/stack/neutron/neutron/db/agents_db.py:317}} Apr 02 14:51:03 np0000163014 neutron-periodic-workers[103796]: DEBUG oslo.service.backend._threading.loopingcall [None req-57e5a4f8-ee4e-491e-8d78-54e1a95d5fc3 None None] Fixed interval looping call 'neutron.plugins.ml2.plugin.AgentDbMixin.agent_health_check' sleeping for 36.99 seconds {{(pid=103796) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:13 np0000163014 neutron-periodic-workers[103796]: DEBUG dbcounter [-] [103796] Writing DB stats neutron:SELECT=1 {{(pid=103796) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:28 np0000163014 neutron-periodic-workers[103784]: DEBUG oslo.service.backend._threading.loopingcall [None req-9c3751da-6e46-4b48-8f20-6d974a8aa7d5 None None] Fixed interval looping call 'neutron.plugins.ml2.plugin.DhcpAgentSchedulerDbMixin.remove_networks_from_down_agents' sleeping for 36.99 seconds {{(pid=103784) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:30 np0000163014 neutron-periodic-workers[103807]: DEBUG oslo.service.backend._threading.loopingcall [None req-16bc2f08-204f-443a-b1ef-aa61f75ba2fd None None] Fixed interval looping call 'neutron.db.quota.driver_nolock.DbQuotaNoLockDriver._remove_expired_reservations' sleeping for 120.00 seconds {{(pid=103807) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:38 np0000163014 neutron-periodic-workers[103784]: DEBUG dbcounter [-] [103784] Writing DB stats neutron:SELECT=3 {{(pid=103784) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:40 np0000163014 neutron-periodic-workers[103796]: DEBUG neutron.db.agents_db [None req-57e5a4f8-ee4e-491e-8d78-54e1a95d5fc3 None None] Agent healthcheck: found 0 active agents {{(pid=103796) agent_health_check /opt/stack/neutron/neutron/db/agents_db.py:317}} Apr 02 14:51:40 np0000163014 neutron-periodic-workers[103796]: DEBUG oslo.service.backend._threading.loopingcall [None req-57e5a4f8-ee4e-491e-8d78-54e1a95d5fc3 None None] Fixed interval looping call 'neutron.plugins.ml2.plugin.AgentDbMixin.agent_health_check' sleeping for 36.99 seconds {{(pid=103796) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:40 np0000163014 neutron-periodic-workers[103807]: DEBUG dbcounter [-] [103807] Writing DB stats neutron:DELETE=1 {{(pid=103807) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:50 np0000163014 neutron-periodic-workers[103796]: DEBUG dbcounter [-] [103796] Writing DB stats neutron:SELECT=1 {{(pid=103796) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@neutron-rpc-server.service - Devstack devstack@neutron-rpc-server.service Loaded: loaded (/etc/systemd/system/devstack@neutron-rpc-server.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:45:25 UTC; 6min ago Main PID: 102556 (neutron-rpc-ser) Tasks: 29 (limit: 77075) Memory: 218.8M (peak: 219.5M) CPU: 4.183s CGroup: /system.slice/system-devstack.slice/devstack@neutron-rpc-server.service ├─102556 "neutron-rpc-server: master process [/opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" ├─103752 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" └─103760 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" Apr 02 14:50:31 np0000163014 neutron-rpc-server[103752]: DEBUG dbcounter [-] [103752] Writing DB stats neutron:SELECT=2 {{(pid=103752) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:31 np0000163014 neutron-rpc-server[103760]: DEBUG dbcounter [-] [103760] Writing DB stats neutron:SELECT=2 {{(pid=103760) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:33 np0000163014 neutron-rpc-server[103760]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.ovsdb_monitor [-] ChassisAgentWriteEvent : Matched Chassis_Private, update, None None {{(pid=103760) matches /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/ovsdb_monitor.py:64}} Apr 02 14:50:33 np0000163014 neutron-rpc-server[103752]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.ovsdb_monitor [-] ChassisAgentWriteEvent : Matched Chassis_Private, update, None None {{(pid=103752) matches /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/ovsdb_monitor.py:64}} Apr 02 14:50:33 np0000163014 neutron-rpc-server[103760]: DEBUG neutron.common.ovn.hash_ring_manager [-] Hash Ring loaded. 4 active nodes. 0 offline nodes {{(pid=103760) _load_hash_ring /opt/stack/neutron/neutron/common/ovn/hash_ring_manager.py:102}} Apr 02 14:50:33 np0000163014 neutron-rpc-server[103752]: DEBUG neutron.common.ovn.hash_ring_manager [-] Hash Ring loaded. 4 active nodes. 0 offline nodes {{(pid=103752) _load_hash_ring /opt/stack/neutron/neutron/common/ovn/hash_ring_manager.py:102}} Apr 02 14:50:38 np0000163014 neutron-rpc-server[103760]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.ovsdb_monitor [-] ChassisOVNAgentWriteEvent : Matched Chassis_Private, update, None None {{(pid=103760) matches /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/ovsdb_monitor.py:64}} Apr 02 14:50:38 np0000163014 neutron-rpc-server[103752]: DEBUG neutron.plugins.ml2.drivers.ovn.mech_driver.ovsdb.ovsdb_monitor [-] ChassisOVNAgentWriteEvent : Matched Chassis_Private, update, None None {{(pid=103752) matches /opt/stack/neutron/neutron/plugins/ml2/drivers/ovn/mech_driver/ovsdb/ovsdb_monitor.py:64}} Apr 02 14:50:43 np0000163014 neutron-rpc-server[103760]: DEBUG dbcounter [-] [103760] Writing DB stats neutron:SELECT=2 {{(pid=103760) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:43 np0000163014 neutron-rpc-server[103752]: DEBUG dbcounter [-] [103752] Writing DB stats neutron:SELECT=2 {{(pid=103752) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@o-api.service - Devstack devstack@o-api.service Loaded: loaded (/etc/systemd/system/devstack@o-api.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:50:15 UTC; 1min 39s ago Main PID: 122913 (uwsgi) Status: "uWSGI is ready" Tasks: 9 (limit: 77075) Memory: 462.3M (peak: 463.3M) CPU: 17.912s CGroup: /system.slice/system-devstack.slice/devstack@o-api.service ├─122913 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv ├─122915 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv ├─122916 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv ├─122917 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv └─122918 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv Apr 02 14:50:22 np0000163014 devstack@o-api.service[122917]: DEBUG ovsdbapp.backend.ovs_idl [-] Created schema index Logical_Switch_Port.name {{(pid=122917) autocreate_indices /opt/stack/data/venv/lib/python3.12/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106}} Apr 02 14:50:22 np0000163014 devstack@o-api.service[122917]: DEBUG ovsdbapp.backend.ovs_idl [-] Created schema index Logical_Router_Port.name {{(pid=122917) autocreate_indices /opt/stack/data/venv/lib/python3.12/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106}} Apr 02 14:50:22 np0000163014 devstack@o-api.service[122917]: DEBUG ovsdbapp.backend.ovs_idl [-] Created schema index Gateway_Chassis.name {{(pid=122917) autocreate_indices /opt/stack/data/venv/lib/python3.12/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106}} Apr 02 14:50:22 np0000163014 devstack@o-api.service[122917]: DEBUG ovsdbapp.backend.ovs_idl [-] Created schema index HA_Chassis_Group.name {{(pid=122917) autocreate_indices /opt/stack/data/venv/lib/python3.12/site-packages/ovsdbapp/backend/ovs_idl/__init__.py:106}} Apr 02 14:50:22 np0000163014 devstack@o-api.service[122916]: WARNING keystonemiddleware.auth_token [-] AuthToken middleware is set with keystone_authtoken.service_token_roles_required set to False. This is backwards compatible but deprecated behaviour. Please set this to True. Apr 02 14:50:22 np0000163014 devstack@o-api.service[122916]: WSGI app 0 (mountpoint='') ready in 7 seconds on interpreter 0x785ad58a3668 pid: 122916 (default app) Apr 02 14:50:22 np0000163014 devstack@o-api.service[122915]: WARNING keystonemiddleware.auth_token [-] AuthToken middleware is set with keystone_authtoken.service_token_roles_required set to False. This is backwards compatible but deprecated behaviour. Please set this to True. Apr 02 14:50:22 np0000163014 devstack@o-api.service[122915]: WSGI app 0 (mountpoint='') ready in 7 seconds on interpreter 0x785ad58a3668 pid: 122915 (default app) Apr 02 14:50:22 np0000163014 devstack@o-api.service[122917]: WARNING keystonemiddleware.auth_token [-] AuthToken middleware is set with keystone_authtoken.service_token_roles_required set to False. This is backwards compatible but deprecated behaviour. Please set this to True. Apr 02 14:50:22 np0000163014 devstack@o-api.service[122917]: WSGI app 0 (mountpoint='') ready in 7 seconds on interpreter 0x785ad58a3668 pid: 122917 (default app) ● devstack@o-da.service - Devstack devstack@o-da.service Loaded: loaded (/etc/systemd/system/devstack@o-da.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:50:17 UTC; 1min 37s ago Main PID: 123448 (octavia-driver-) Tasks: 17 (limit: 77075) Memory: 189.8M (peak: 190.2M) CPU: 3.841s CGroup: /system.slice/system-devstack.slice/devstack@o-da.service ├─123448 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/octavia-driver-agent --config-file /etc/octavia/octavia.conf ├─124243 "octavia-driver-agent - status_listener" ├─124245 "octavia-driver-agent - stats_listener" ├─124248 "octavia-driver-agent - get_listener" └─124363 "octavia-driver-agent - provider_agent -- ovn" Apr 02 14:50:20 np0000163014 octavia-driver-agent[124363]: INFO ovn_octavia_provider.maintenance [-] Periodic task found: DBInconsistenciesPeriodics.format_ip_port_mappings_ipv6 Apr 02 14:50:20 np0000163014 octavia-driver-agent[124363]: DEBUG futurist.periodics [-] Submitting immediate callback 'ovn_octavia_provider.maintenance.DBInconsistenciesPeriodics.change_device_owner_lb_hm_ports' {{(pid=124363) _process_immediates /opt/stack/data/venv/lib/python3.12/site-packages/futurist/periodics.py:826}} Apr 02 14:50:20 np0000163014 octavia-driver-agent[124363]: INFO ovn_octavia_provider.agent [-] OVN provider agent has started. Apr 02 14:50:20 np0000163014 octavia-driver-agent[124363]: DEBUG ovn_octavia_provider.maintenance [-] Maintenance task: checking device_owner for OVN LB HM ports. {{(pid=124363) change_device_owner_lb_hm_ports /opt/stack/ovn-octavia-provider/ovn_octavia_provider/maintenance.py:81}} Apr 02 14:50:21 np0000163014 octavia-driver-agent[124363]: DEBUG ovn_octavia_provider.maintenance [-] Maintenance task: no more ports left, stopping the periodic task. {{(pid=124363) change_device_owner_lb_hm_ports /opt/stack/ovn-octavia-provider/ovn_octavia_provider/maintenance.py:121}} Apr 02 14:50:21 np0000163014 octavia-driver-agent[124363]: DEBUG futurist.periodics [-] Periodic callback 'ovn_octavia_provider.maintenance.DBInconsistenciesPeriodics.change_device_owner_lb_hm_ports' raised 'NeverAgain' exception, stopping any further execution of it. {{(pid=124363) _on_done /opt/stack/data/venv/lib/python3.12/site-packages/futurist/periodics.py:881}} Apr 02 14:50:21 np0000163014 octavia-driver-agent[124363]: DEBUG futurist.periodics [-] Submitting immediate callback 'ovn_octavia_provider.maintenance.DBInconsistenciesPeriodics.format_ip_port_mappings_ipv6' {{(pid=124363) _process_immediates /opt/stack/data/venv/lib/python3.12/site-packages/futurist/periodics.py:826}} Apr 02 14:50:21 np0000163014 octavia-driver-agent[124363]: DEBUG ovn_octavia_provider.maintenance [-] Maintenance task: Ensure correct formatting of ip_port_mappings for IPv6 backend members. {{(pid=124363) format_ip_port_mappings_ipv6 /opt/stack/ovn-octavia-provider/ovn_octavia_provider/maintenance.py:138}} Apr 02 14:50:21 np0000163014 octavia-driver-agent[124363]: DEBUG ovn_octavia_provider.maintenance [-] Maintenance task: no more ip_port_mappings to format, stopping the periodic task. {{(pid=124363) format_ip_port_mappings_ipv6 /opt/stack/ovn-octavia-provider/ovn_octavia_provider/maintenance.py:163}} Apr 02 14:50:21 np0000163014 octavia-driver-agent[124363]: DEBUG futurist.periodics [-] Periodic callback 'ovn_octavia_provider.maintenance.DBInconsistenciesPeriodics.format_ip_port_mappings_ipv6' raised 'NeverAgain' exception, stopping any further execution of it. {{(pid=124363) _on_done /opt/stack/data/venv/lib/python3.12/site-packages/futurist/periodics.py:881}} ● devstack@o-hk.service - Devstack devstack@o-hk.service Loaded: loaded (/etc/systemd/system/devstack@o-hk.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:50:19 UTC; 1min 34s ago Main PID: 124055 (octavia-houseke) Tasks: 3 (limit: 77075) Memory: 105.1M (peak: 105.6M) CPU: 3.302s CGroup: /system.slice/system-devstack.slice/devstack@o-hk.service └─124055 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/octavia-housekeeping --config-file /etc/octavia/octavia.conf Apr 02 14:50:22 np0000163014 octavia-housekeeping[124055]: INFO octavia.cmd.house_keeping [-] DB cleanup interval is set to 30 sec Apr 02 14:50:22 np0000163014 octavia-housekeeping[124055]: INFO octavia.cmd.house_keeping [-] Amphora expiry age is 3600 seconds Apr 02 14:50:22 np0000163014 octavia-housekeeping[124055]: INFO octavia.cmd.house_keeping [-] Load balancer expiry age is 3600 seconds Apr 02 14:50:22 np0000163014 octavia-housekeeping[124055]: DEBUG octavia.cmd.house_keeping [-] Initiating the cleanup of old resources... {{(pid=124055) db_cleanup /opt/stack/octavia/octavia/cmd/house_keeping.py:49}} Apr 02 14:50:22 np0000163014 octavia-housekeeping[124055]: INFO octavia.cmd.house_keeping [-] Expiring certificate check interval is set to 3600 sec Apr 02 14:50:22 np0000163014 octavia-housekeeping[124055]: DEBUG octavia.cmd.house_keeping [-] Initiating certification rotation ... {{(pid=124055) cert_rotation /opt/stack/octavia/octavia/cmd/house_keeping.py:66}} Apr 02 14:50:22 np0000163014 octavia-housekeeping[124055]: DEBUG oslo_db.sqlalchemy.engines [-] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_ENGINE_SUBSTITUTION {{(pid=124055) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.12/site-packages/oslo_db/sqlalchemy/engines.py:325}} Apr 02 14:50:52 np0000163014 octavia-housekeeping[124055]: DEBUG octavia.cmd.house_keeping [-] Initiating the cleanup of old resources... {{(pid=124055) db_cleanup /opt/stack/octavia/octavia/cmd/house_keeping.py:49}} Apr 02 14:51:22 np0000163014 octavia-housekeeping[124055]: DEBUG octavia.cmd.house_keeping [-] Initiating the cleanup of old resources... {{(pid=124055) db_cleanup /opt/stack/octavia/octavia/cmd/house_keeping.py:49}} Apr 02 14:51:52 np0000163014 octavia-housekeeping[124055]: DEBUG octavia.cmd.house_keeping [-] Initiating the cleanup of old resources... {{(pid=124055) db_cleanup /opt/stack/octavia/octavia/cmd/house_keeping.py:49}} ● devstack@openstack-cli-server.service - Devstack devstack@openstack-cli-server.service Loaded: loaded (/etc/systemd/system/devstack@openstack-cli-server.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:38:52 UTC; 13min ago Main PID: 61420 (python3) Tasks: 1 (limit: 77075) Memory: 94.5M (peak: 98.1M) CPU: 28.978s CGroup: /system.slice/system-devstack.slice/devstack@openstack-cli-server.service └─61420 /opt/stack/data/venv/bin/python3 /opt/stack/devstack/files/openstack-cli-server/openstack-cli-server Apr 02 14:50:11 np0000163014 python3[61420]: openstack ['--os-cloud', 'devstack-system-admin', 'role', 'assignment', 'list', '--role', 'load-balancer_member', '--user', 'demo', '--project', 'demo', '-c', 'Role', '-f', 'value'] Apr 02 14:50:21 np0000163014 python3[61420]: openstack ['--os-cloud', 'devstack-admin', '--os-region', 'RegionOne', 'compute', 'service', 'list', '--host', 'np0000163014', '--service', 'nova-compute', '-c', 'ID', '-f', 'value'] Apr 02 14:50:30 np0000163014 python3[61420]: openstack ['complete'] Apr 02 14:50:32 np0000163014 python3[61420]: openstack ['--os-cloud', 'devstack-admin', 'project', 'show', 'service', '-c', 'id', '-f', 'value'] Apr 02 14:50:32 np0000163014 python3[61420]: openstack ['--os-cloud', 'devstack-admin', 'network', 'show', 'admin_net', '-f', 'value', '-c', 'id'] Apr 02 14:50:32 np0000163014 python3[61420]: openstack ['--os-cloud', 'devstack-admin', 'network', 'create', 'admin_net', '--project', '1405e5e4df0c46848af022b1c5632d76'] Apr 02 14:50:33 np0000163014 python3[61420]: openstack ['--os-cloud', 'devstack-admin', 'network', 'show', 'admin_net', '-f', 'value', '-c', 'id'] Apr 02 14:50:33 np0000163014 python3[61420]: openstack ['--os-cloud', 'devstack-admin', 'subnet', 'show', 'admin_subnet', '-f', 'value', '-c', 'id'] Apr 02 14:50:34 np0000163014 python3[61420]: openstack ['--os-cloud', 'devstack-admin', 'subnet', 'create', 'admin_subnet', '--project', '1405e5e4df0c46848af022b1c5632d76', '--ip-version', '4', '--network', '6fccf94a-88ef-49fd-a0c3-4d8e39aeea0c', '--gateway', 'None', '--subnet-range', '10.2.5.0/24'] Apr 02 14:50:35 np0000163014 python3[61420]: openstack ['--os-cloud', 'devstack-admin', 'subnet', 'show', 'admin_subnet', '-f', 'value', '-c', 'id'] ● devstack@placement-api.service - Devstack devstack@placement-api.service Loaded: loaded (/etc/systemd/system/devstack@placement-api.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:45:32 UTC; 6min ago Main PID: 104337 (uwsgi) Status: "uWSGI is ready" Tasks: 9 (limit: 77075) Memory: 331.8M (peak: 332.3M) CPU: 5.977s CGroup: /system.slice/system-devstack.slice/devstack@placement-api.service ├─104337 "placementuWSGI master" ├─104338 "placementuWSGI worker 1" ├─104339 "placementuWSGI worker 2" ├─104340 "placementuWSGI worker 3" └─104341 "placementuWSGI worker 4" Apr 02 14:50:29 np0000163014 devstack@placement-api.service[104339]: DEBUG dbcounter [-] [104339] Writing DB stats placement:SELECT=2 {{(pid=104339) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:50:29 np0000163014 devstack@placement-api.service[104341]: DEBUG dbcounter [-] [104341] Writing DB stats placement:SELECT=2 {{(pid=104341) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:18 np0000163014 devstack@placement-api.service[104338]: DEBUG placement.requestlog [req-5ed727a8-c702-4ea1-9b1f-05d55be6a6e7 req-7ab64734-9b9b-48c0-b0e5-50e1c3a75ae4 None None] Starting request: 199.204.45.109 "GET /placement/resource_providers/bf3a6875-73c0-49fe-9350-8dc3315eae6b/allocations" {{(pid=104338) __call__ /opt/stack/placement/placement/requestlog.py:55}} Apr 02 14:51:18 np0000163014 devstack@placement-api.service[104338]: INFO placement.requestlog [req-5ed727a8-c702-4ea1-9b1f-05d55be6a6e7 req-7ab64734-9b9b-48c0-b0e5-50e1c3a75ae4 service nova] 199.204.45.109 "GET /placement/resource_providers/bf3a6875-73c0-49fe-9350-8dc3315eae6b/allocations" status: 200 len: 54 microversion: 1.0 Apr 02 14:51:18 np0000163014 devstack@placement-api.service[104338]: [pid: 104338|app: 0|req: 6/21] 199.204.45.109 () {66 vars in 1543 bytes} [Thu Apr 2 14:51:18 2026] GET /placement/resource_providers/bf3a6875-73c0-49fe-9350-8dc3315eae6b/allocations => generated 54 bytes in 14 msecs (HTTP/1.1 200) 6 headers in 223 bytes (1 switches on core 0) Apr 02 14:51:18 np0000163014 devstack@placement-api.service[104339]: DEBUG placement.requestlog [req-5ed727a8-c702-4ea1-9b1f-05d55be6a6e7 req-46a32252-7214-40f4-a63e-0b661e035cd4 None None] Starting request: 199.204.45.109 "GET /placement/resource_providers/bf3a6875-73c0-49fe-9350-8dc3315eae6b/allocations" {{(pid=104339) __call__ /opt/stack/placement/placement/requestlog.py:55}} Apr 02 14:51:18 np0000163014 devstack@placement-api.service[104339]: INFO placement.requestlog [req-5ed727a8-c702-4ea1-9b1f-05d55be6a6e7 req-46a32252-7214-40f4-a63e-0b661e035cd4 service nova] 199.204.45.109 "GET /placement/resource_providers/bf3a6875-73c0-49fe-9350-8dc3315eae6b/allocations" status: 200 len: 54 microversion: 1.0 Apr 02 14:51:18 np0000163014 devstack@placement-api.service[104339]: [pid: 104339|app: 0|req: 6/22] 199.204.45.109 () {66 vars in 1543 bytes} [Thu Apr 2 14:51:18 2026] GET /placement/resource_providers/bf3a6875-73c0-49fe-9350-8dc3315eae6b/allocations => generated 54 bytes in 11 msecs (HTTP/1.1 200) 6 headers in 223 bytes (1 switches on core 0) Apr 02 14:51:28 np0000163014 devstack@placement-api.service[104338]: DEBUG dbcounter [-] [104338] Writing DB stats placement:SELECT=2 {{(pid=104338) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:28 np0000163014 devstack@placement-api.service[104339]: DEBUG dbcounter [-] [104339] Writing DB stats placement:SELECT=2 {{(pid=104339) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● devstack@q-ovn-agent.service - Devstack devstack@q-ovn-agent.service Loaded: loaded (/etc/systemd/system/devstack@q-ovn-agent.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:45:19 UTC; 6min ago Main PID: 100946 (neutron-ovn-age) Tasks: 32 (limit: 77075) Memory: 473.4M (peak: 474.7M) CPU: 8.482s CGroup: /system.slice/system-devstack.slice/devstack@q-ovn-agent.service ├─100946 "neutron-ovn-agent: master process [/opt/stack/data/venv/bin/neutron-ovn-agent --config-file /etc/neutron/plugins/ml2/ovn_agent.ini]" ├─101518 "neutron-ovn-agent: ServiceWrapper worker(0)" ├─101743 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmpzexbsa_i/privsep.sock ├─105256 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmp2clb52x1/privsep.sock ├─127213 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.link_cmd --privsep_sock_path /tmp/tmpypayx3b_/privsep.sock ├─127750 sudo /opt/stack/data/venv/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf ├─127751 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf └─127791 haproxy -f /opt/stack/data/neutron/ovn-metadata-proxy/6fccf94a-88ef-49fd-a0c3-4d8e39aeea0c.conf Apr 02 14:50:42 np0000163014 neutron-ovn-agent[101518]: DEBUG oslo_rootwrap.client [-] Popen for ['sudo', '/opt/stack/data/venv/bin/neutron-rootwrap-daemon', '/etc/neutron/rootwrap.conf'] command has been instantiated {{(pid=101518) _initialize /opt/stack/data/venv/lib/python3.12/site-packages/oslo_rootwrap/client.py:85}} Apr 02 14:50:42 np0000163014 sudo[127750]: stack : PWD=/ ; USER=root ; COMMAND=/opt/stack/data/venv/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf Apr 02 14:50:42 np0000163014 sudo[127750]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=1002) Apr 02 14:50:43 np0000163014 neutron-ovn-agent[101518]: INFO oslo_rootwrap.client [-] Spawned new rootwrap daemon process with pid=127750 Apr 02 14:51:21 np0000163014 neutron-ovn-agent[101518]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" {{(pid=101518) inner /opt/stack/data/venv/lib/python3.12/site-packages/oslo_concurrency/lockutils.py:506}} Apr 02 14:51:21 np0000163014 neutron-ovn-agent[101518]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "_check_child_processes" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" {{(pid=101518) inner /opt/stack/data/venv/lib/python3.12/site-packages/oslo_concurrency/lockutils.py:506}} Apr 02 14:51:21 np0000163014 neutron-ovn-agent[101518]: DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s {{(pid=101518) inner /opt/stack/data/venv/lib/python3.12/site-packages/oslo_concurrency/lockutils.py:519}} Apr 02 14:51:21 np0000163014 neutron-ovn-agent[101518]: DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s {{(pid=101518) inner /opt/stack/data/venv/lib/python3.12/site-packages/oslo_concurrency/lockutils.py:538}} Apr 02 14:51:21 np0000163014 neutron-ovn-agent[101518]: DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" acquired by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: waited 0.001s {{(pid=101518) inner /opt/stack/data/venv/lib/python3.12/site-packages/oslo_concurrency/lockutils.py:519}} Apr 02 14:51:21 np0000163014 neutron-ovn-agent[101518]: DEBUG oslo_concurrency.lockutils [-] Lock "_check_child_processes" "released" by "neutron.agent.linux.external_process.ProcessMonitor._check_child_processes" :: held 0.001s {{(pid=101518) inner /opt/stack/data/venv/lib/python3.12/site-packages/oslo_concurrency/lockutils.py:538}} ● dm-event.service - Device-mapper event daemon Loaded: loaded (/usr/lib/systemd/system/dm-event.service; static) Active: active (running) since Thu 2026-04-02 14:46:29 UTC; 5min ago TriggeredBy: ● dm-event.socket Docs: man:dmeventd(8) Main PID: 113941 (dmeventd) Tasks: 3 (limit: 77075) Memory: 14.4M (peak: 15.0M) CPU: 64ms CGroup: /system.slice/dm-event.service └─113941 /usr/sbin/dmeventd -f Apr 02 14:46:29 np0000163014 systemd[1]: Started dm-event.service - Device-mapper event daemon. Apr 02 14:46:29 np0000163014 dmeventd[113941]: dmeventd ready for processing. Apr 02 14:46:29 np0000163014 dmeventd[113941]: Monitoring thin pool stack--volumes--lvmdriver--1-stack--volumes--lvmdriver--1--pool. ○ dmesg.service - Save initial kernel messages after boot Loaded: loaded (/usr/lib/systemd/system/dmesg.service; enabled; preset: enabled) Active: inactive (dead) since Thu 2026-04-02 14:27:55 UTC; 23min ago Duration: 112ms Main PID: 11711 (code=exited, status=0/SUCCESS) CPU: 59ms Apr 02 14:27:55 np0000163014 systemd[1]: Started dmesg.service - Save initial kernel messages after boot. Apr 02 14:27:55 np0000163014 systemd[1]: dmesg.service: Deactivated successfully. ● docker.service - Docker Application Container Engine Loaded: loaded (/usr/lib/systemd/system/docker.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:29:34 UTC; 22min ago TriggeredBy: ● docker.socket Docs: https://docs.docker.com Main PID: 20379 (dockerd) Tasks: 28 Memory: 442.4M (peak: 452.3M) CPU: 4.812s CGroup: /system.slice/docker.service ├─20379 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock └─21206 /usr/bin/docker-proxy -proto tcp -host-ip 127.0.0.1 -host-port 42841 -container-ip 172.18.0.2 -container-port 6443 -use-listen-fd Apr 02 14:29:34 np0000163014 dockerd[20379]: time="2026-04-02T14:29:34.901158705Z" level=info msg="Docker daemon" commit=f78c987 containerd-snapshotter=true storage-driver=overlayfs version=29.3.1 Apr 02 14:29:34 np0000163014 dockerd[20379]: time="2026-04-02T14:29:34.901314467Z" level=info msg="Initializing buildkit" Apr 02 14:29:34 np0000163014 dockerd[20379]: time="2026-04-02T14:29:34.923042716Z" level=info msg="Completed buildkit initialization" Apr 02 14:29:34 np0000163014 dockerd[20379]: time="2026-04-02T14:29:34.936948739Z" level=info msg="Daemon has completed initialization" Apr 02 14:29:34 np0000163014 dockerd[20379]: time="2026-04-02T14:29:34.937129052Z" level=info msg="API listen on /run/docker.sock" Apr 02 14:29:34 np0000163014 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 02 14:29:47 np0000163014 dockerd[20379]: time="2026-04-02T14:29:47.079364820Z" level=info msg="image pulled" digest="sha256:9be91e9e9cdf116809841fc77ebdb8845443c4c72fe5218f3ae9eb57fdb4bace" remote="docker.io/kindest/node@sha256:9be91e9e9cdf116809841fc77ebdb8845443c4c72fe5218f3ae9eb57fdb4bace" Apr 02 14:29:47 np0000163014 dockerd[20379]: time="2026-04-02T14:29:47.113608372Z" level=error msg="failed to validate image signature" error="resolving signature chain for image sha256:9be91e9e9cdf116809841fc77ebdb8845443c4c72fe5218f3ae9eb57fdb4bace: expected image index descriptor, got application/vnd.docker.distribution.manifest.list.v2+json" Apr 02 14:29:56 np0000163014 dockerd[20379]: time="2026-04-02T14:29:56.990558921Z" level=info msg="Skipping check for route to send NA, EMSGSIZE" eid=3a65e1d2bc40 ep=kind-control-plane net=kind nid=94c6ce52c320 Apr 02 14:29:56 np0000163014 dockerd[20379]: time="2026-04-02T14:29:56.993580213Z" level=info msg="sbJoin: gwep4 ''->'3a65e1d2bc40', gwep6 ''->'3a65e1d2bc40'" eid=3a65e1d2bc40 ep=kind-control-plane net=kind nid=94c6ce52c320 ○ dpkg-db-backup.service - Daily dpkg database backup service Loaded: loaded (/usr/lib/systemd/system/dpkg-db-backup.service; static) Active: inactive (dead) TriggeredBy: ● dpkg-db-backup.timer Docs: man:dpkg(1) ○ e2scrub_all.service - Online ext4 Metadata Check for All Filesystems Loaded: loaded (/usr/lib/systemd/system/e2scrub_all.service; static) Active: inactive (dead) TriggeredBy: ● e2scrub_all.timer Docs: man:e2scrub_all(8) ○ e2scrub_reap.service - Remove Stale Online ext4 Metadata Check Snapshots Loaded: loaded (/usr/lib/systemd/system/e2scrub_reap.service; enabled; preset: enabled) Active: inactive (dead) since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:e2scrub_all(8) Main PID: 700 (code=exited, status=0/SUCCESS) CPU: 33ms Apr 02 14:21:10 np0000163014 systemd[1]: Starting e2scrub_reap.service - Remove Stale Online ext4 Metadata Check Snapshots... Apr 02 14:21:10 np0000163014 systemd[1]: e2scrub_reap.service: Deactivated successfully. Apr 02 14:21:10 np0000163014 systemd[1]: Finished e2scrub_reap.service - Remove Stale Online ext4 Metadata Check Snapshots. ○ emergency.service - Emergency Shell Loaded: loaded (/usr/lib/systemd/system/emergency.service; static) Active: inactive (dead) Docs: man:sulogin(8) ● epmd.service - Erlang Port Mapper Daemon Loaded: loaded (/usr/lib/systemd/system/epmd.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:31:29 UTC; 20min ago TriggeredBy: ● epmd.socket Main PID: 25717 (epmd) Tasks: 1 (limit: 77075) Memory: 452.0K (peak: 1.4M) CPU: 39ms CGroup: /system.slice/epmd.service └─25717 /usr/bin/epmd -systemd Apr 02 14:31:29 np0000163014 systemd[1]: Started epmd.service - Erlang Port Mapper Daemon. ● fsidd.service - NFS FSID Daemon Loaded: loaded (/usr/lib/systemd/system/fsidd.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:36:36 UTC; 15min ago Main PID: 53941 (fsidd) Tasks: 1 (limit: 77075) Memory: 436.0K (peak: 676.0K) CPU: 5ms CGroup: /system.slice/fsidd.service └─53941 /usr/sbin/fsidd Apr 02 14:36:36 np0000163014 systemd[1]: Started fsidd.service - NFS FSID Daemon. ○ fstrim.service - Discard unused blocks on filesystems from /etc/fstab Loaded: loaded (/usr/lib/systemd/system/fstrim.service; static) Active: inactive (dead) TriggeredBy: ● fstrim.timer Docs: man:fstrim(8) ○ getty-static.service - getty on tty2-tty6 if dbus and logind are not available Loaded: loaded (/usr/lib/systemd/system/getty-static.service; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:10 UTC; 30min ago Apr 02 14:21:10 np0000163014 systemd[1]: getty-static.service - getty on tty2-tty6 if dbus and logind are not available was skipped because of an unmet condition check (ConditionPathExists=!/usr/bin/dbus-daemon). ● getty@tty1.service - Getty on tty1 Loaded: loaded (/usr/lib/systemd/system/getty@.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:agetty(8) man:systemd-getty-generator(8) https://0pointer.de/blog/projects/serial-console.html Main PID: 714 (agetty) Tasks: 1 (limit: 77075) Memory: 292.0K (peak: 1.8M) CPU: 17ms CGroup: /system.slice/system-getty.slice/getty@tty1.service └─714 /sbin/agetty -o "-p -- \\u" --noclear - linux Apr 02 14:21:10 np0000163014 systemd[1]: Started getty@tty1.service - Getty on tty1. ○ grub-common.service - Record successful boot for GRUB Loaded: loaded (/usr/lib/systemd/system/grub-common.service; enabled; preset: enabled) Active: inactive (dead) since Thu 2026-04-02 14:21:10 UTC; 30min ago Main PID: 707 (code=exited, status=0/SUCCESS) CPU: 49ms Apr 02 14:21:10 np0000163014 systemd[1]: Starting grub-common.service - Record successful boot for GRUB... Apr 02 14:21:10 np0000163014 systemd[1]: grub-common.service: Deactivated successfully. Apr 02 14:21:10 np0000163014 systemd[1]: Finished grub-common.service - Record successful boot for GRUB. ○ grub-initrd-fallback.service - GRUB failed boot detection Loaded: loaded (/usr/lib/systemd/system/grub-initrd-fallback.service; enabled; preset: enabled) Active: inactive (dead) since Thu 2026-04-02 14:21:10 UTC; 30min ago Main PID: 738 (code=exited, status=0/SUCCESS) CPU: 21ms Apr 02 14:21:10 np0000163014 systemd[1]: Starting grub-initrd-fallback.service - GRUB failed boot detection... Apr 02 14:21:10 np0000163014 systemd[1]: grub-initrd-fallback.service: Deactivated successfully. Apr 02 14:21:10 np0000163014 systemd[1]: Finished grub-initrd-fallback.service - GRUB failed boot detection. ● haproxy.service - HAProxy Load Balancer Loaded: loaded (/usr/lib/systemd/system/haproxy.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:28:11 UTC; 23min ago Docs: man:haproxy(1) file:/usr/share/doc/haproxy/configuration.txt.gz Main PID: 12955 (haproxy) Status: "Ready." Tasks: 17 (limit: 77075) Memory: 42.9M (peak: 45.6M) CPU: 201ms CGroup: /system.slice/haproxy.service ├─12955 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock └─12957 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock Apr 02 14:28:11 np0000163014 systemd[1]: Starting haproxy.service - HAProxy Load Balancer... Apr 02 14:28:11 np0000163014 haproxy[12955]: [NOTICE] (12955) : New worker (12957) forked Apr 02 14:28:11 np0000163014 haproxy[12955]: [NOTICE] (12955) : Loading success. Apr 02 14:28:11 np0000163014 systemd[1]: Started haproxy.service - HAProxy Load Balancer. ○ initrd-cleanup.service - Cleaning Up and Shutting Down Daemons Loaded: loaded (/usr/lib/systemd/system/initrd-cleanup.service; static) Active: inactive (dead) ○ initrd-parse-etc.service - Mountpoints Configured in the Real Root Loaded: loaded (/usr/lib/systemd/system/initrd-parse-etc.service; static) Active: inactive (dead) ○ initrd-switch-root.service - Switch Root Loaded: loaded (/usr/lib/systemd/system/initrd-switch-root.service; static) Active: inactive (dead) ○ initrd-udevadm-cleanup-db.service - Cleanup udev Database Loaded: loaded (/usr/lib/systemd/system/initrd-udevadm-cleanup-db.service; static) Active: inactive (dead) ● iscsid.service - iSCSI initiator daemon (iscsid) Loaded: loaded (/usr/lib/systemd/system/iscsid.service; disabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:35:01 UTC; 16min ago TriggeredBy: ● iscsid.socket Docs: man:iscsid(8) Main PID: 43376 (iscsid) Tasks: 2 (limit: 77075) Memory: 2.7M (peak: 3.1M) CPU: 70ms CGroup: /system.slice/iscsid.service ├─43375 /usr/sbin/iscsid └─43376 /usr/sbin/iscsid Apr 02 14:35:01 np0000163014 systemd[1]: Starting iscsid.service - iSCSI initiator daemon (iscsid)... Apr 02 14:35:01 np0000163014 iscsid[43373]: iSCSI logger with pid=43375 started! Apr 02 14:35:01 np0000163014 systemd[1]: Started iscsid.service - iSCSI initiator daemon (iscsid). Apr 02 14:35:02 np0000163014 iscsid[43375]: iSCSI daemon with pid=43376 started! ● kmod-static-nodes.service - Create List of Static Device Nodes Loaded: loaded (/usr/lib/systemd/system/kmod-static-nodes.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Main PID: 387 (code=exited, status=0/SUCCESS) CPU: 11ms Apr 02 14:21:04 ubuntu systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Notice: journal has been rotated since unit was started, output may be incomplete. ● ksm.service - Kernel Samepage Merging Loaded: loaded (/usr/lib/systemd/system/ksm.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:26:34 UTC; 25min ago Main PID: 4936 (code=exited, status=0/SUCCESS) CPU: 1ms Apr 02 14:26:34 np0000163014 systemd[1]: Starting ksm.service - Kernel Samepage Merging... Apr 02 14:26:34 np0000163014 systemd[1]: Finished ksm.service - Kernel Samepage Merging. ● ksmtuned.service - Kernel Samepage Merging (KSM) Tuning Daemon Loaded: loaded (/usr/lib/systemd/system/ksmtuned.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:26:34 UTC; 25min ago Main PID: 4941 (ksmtuned) Tasks: 2 (limit: 77075) Memory: 2.5M (peak: 4.2M) CPU: 1.510s CGroup: /system.slice/ksmtuned.service ├─ 4941 /bin/bash /usr/sbin/ksmtuned └─128547 sleep 60 Apr 02 14:26:34 np0000163014 systemd[1]: Starting ksmtuned.service - Kernel Samepage Merging (KSM) Tuning Daemon... Apr 02 14:26:34 np0000163014 systemd[1]: Started ksmtuned.service - Kernel Samepage Merging (KSM) Tuning Daemon. ● ldconfig.service - Rebuild Dynamic Linker Cache Loaded: loaded (/usr/lib/systemd/system/ldconfig.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:ldconfig(8) Main PID: 448 (code=exited, status=0/SUCCESS) CPU: 44ms Apr 02 14:21:04 ubuntu systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 02 14:21:04 ubuntu systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. ● libvirt-guests.service - libvirt guests suspend/resume service Loaded: loaded (/usr/lib/systemd/system/libvirt-guests.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:34:43 UTC; 17min ago Docs: man:libvirt-guests(8) https://libvirt.org/ Main PID: 42433 (code=exited, status=0/SUCCESS) CPU: 20ms Apr 02 14:34:43 np0000163014 systemd[1]: Starting libvirt-guests.service - libvirt guests suspend/resume service... Apr 02 14:34:43 np0000163014 systemd[1]: Finished libvirt-guests.service - libvirt guests suspend/resume service. ● libvirtd.service - libvirt legacy monolithic daemon Loaded: loaded (/usr/lib/systemd/system/libvirtd.service; enabled; preset: enabled) Drop-In: /etc/systemd/system/libvirtd.service.d └─coredump.conf Active: active (running) since Thu 2026-04-02 14:46:15 UTC; 5min ago TriggeredBy: ● libvirtd-admin.socket ● libvirtd-ro.socket ● libvirtd.socket Docs: man:libvirtd(8) https://libvirt.org/ Main PID: 110870 (libvirtd) Tasks: 22 (limit: 32768) Memory: 33.2M (peak: 63.0M) CPU: 6.428s CGroup: /system.slice/libvirtd.service ├─ 42323 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/lib/libvirt/libvirt_leaseshelper ├─ 42324 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/lib/libvirt/libvirt_leaseshelper └─110870 /usr/sbin/libvirtd --timeout 120 Apr 02 14:46:15 np0000163014 systemd[1]: Starting libvirtd.service - libvirt legacy monolithic daemon... Apr 02 14:46:15 np0000163014 libvirtd[110870]: 2026-04-02 14:46:15.448+0000: 110870: info : libvirt version: 10.0.0, package: 10.0.0-2ubuntu8.12 (Ubuntu) Apr 02 14:46:15 np0000163014 libvirtd[110870]: 2026-04-02 14:46:15.448+0000: 110870: info : hostname: np0000163014 Apr 02 14:46:15 np0000163014 libvirtd[110870]: 2026-04-02 14:46:15.448+0000: 110870: debug : virLogParseOutputs:1638 : outputs=1:file:/var/log/libvirt/libvirtd.log Apr 02 14:46:15 np0000163014 libvirtd[110870]: 2026-04-02 14:46:15.448+0000: 110870: debug : virLogParseOutput:1485 : output=1:file:/var/log/libvirt/libvirtd.log Apr 02 14:46:15 np0000163014 systemd[1]: Started libvirtd.service - libvirt legacy monolithic daemon. Apr 02 14:46:16 np0000163014 dnsmasq[42323]: read /etc/hosts - 8 names Apr 02 14:46:16 np0000163014 dnsmasq[42323]: read /var/lib/libvirt/dnsmasq/default.addnhosts - 0 names Apr 02 14:46:16 np0000163014 dnsmasq-dhcp[42323]: read /var/lib/libvirt/dnsmasq/default.hostsfile ○ logrotate.service - Rotate log files Loaded: loaded (/usr/lib/systemd/system/logrotate.service; static) Active: inactive (dead) TriggeredBy: ● logrotate.timer Docs: man:logrotate(8) man:logrotate.conf(5) ○ lvm2-lvmpolld.service - LVM2 poll daemon Loaded: loaded (/usr/lib/systemd/system/lvm2-lvmpolld.service; static) Active: inactive (dead) TriggeredBy: ● lvm2-lvmpolld.socket Docs: man:lvmpolld(8) ● lvm2-monitor.service - Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling Loaded: loaded (/usr/lib/systemd/system/lvm2-monitor.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:28:25 UTC; 23min ago Docs: man:dmeventd(8) man:lvcreate(8) man:lvchange(8) man:vgchange(8) Main PID: 14237 (code=exited, status=0/SUCCESS) CPU: 22ms Apr 02 14:28:25 np0000163014 systemd[1]: Starting lvm2-monitor.service - Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling... Apr 02 14:28:25 np0000163014 systemd[1]: Finished lvm2-monitor.service - Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling. ○ man-db.service - Daily man-db regeneration Loaded: loaded (/usr/lib/systemd/system/man-db.service; static) Active: inactive (dead) TriggeredBy: ● man-db.timer Docs: man:mandb(8) ● memcached.service - memcached daemon Loaded: loaded (/usr/lib/systemd/system/memcached.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:39:31 UTC; 12min ago Docs: man:memcached(1) Main PID: 65584 (memcached) Tasks: 10 (limit: 77075) Memory: 16.6M (peak: 17.3M) CPU: 2.720s CGroup: /system.slice/memcached.service └─65584 /usr/bin/memcached -m 64 -p 11211 -u memcache -l 127.0.0.1 -l ::1 -P /var/run/memcached/memcached.pid Apr 02 14:39:31 np0000163014 systemd[1]: Started memcached.service - memcached daemon. ○ modprobe@configfs.service - Load Kernel Module configfs Loaded: loaded (/usr/lib/systemd/system/modprobe@.service; static) Active: inactive (dead) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:modprobe(8) Main PID: 388 (code=exited, status=0/SUCCESS) CPU: 14ms Apr 02 14:21:04 ubuntu systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 02 14:21:04 ubuntu systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Notice: journal has been rotated since unit was started, output may be incomplete. ○ modprobe@dm_mod.service - Load Kernel Module dm_mod Loaded: loaded (/usr/lib/systemd/system/modprobe@.service; static) Active: inactive (dead) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:modprobe(8) Main PID: 389 (code=exited, status=0/SUCCESS) CPU: 12ms Apr 02 14:21:04 ubuntu systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 02 14:21:04 ubuntu systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Notice: journal has been rotated since unit was started, output may be incomplete. ○ modprobe@drm.service - Load Kernel Module drm Loaded: loaded (/usr/lib/systemd/system/modprobe@.service; static) Active: inactive (dead) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:modprobe(8) Main PID: 390 (code=exited, status=0/SUCCESS) CPU: 9ms Apr 02 14:21:04 ubuntu systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 02 14:21:04 ubuntu systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Notice: journal has been rotated since unit was started, output may be incomplete. ○ modprobe@efi_pstore.service - Load Kernel Module efi_pstore Loaded: loaded (/usr/lib/systemd/system/modprobe@.service; static) Active: inactive (dead) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:modprobe(8) Main PID: 391 (code=exited, status=0/SUCCESS) CPU: 10ms Apr 02 14:21:04 ubuntu systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 02 14:21:04 ubuntu systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Notice: journal has been rotated since unit was started, output may be incomplete. ○ modprobe@fuse.service - Load Kernel Module fuse Loaded: loaded (/usr/lib/systemd/system/modprobe@.service; static) Active: inactive (dead) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:modprobe(8) Main PID: 392 (code=exited, status=0/SUCCESS) CPU: 13ms Apr 02 14:21:04 ubuntu systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 02 14:21:04 ubuntu systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Notice: journal has been rotated since unit was started, output may be incomplete. ○ modprobe@loop.service - Load Kernel Module loop Loaded: loaded (/usr/lib/systemd/system/modprobe@.service; static) Active: inactive (dead) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:modprobe(8) Main PID: 393 (code=exited, status=0/SUCCESS) CPU: 13ms Apr 02 14:21:04 ubuntu systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 02 14:21:04 ubuntu systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Notice: journal has been rotated since unit was started, output may be incomplete. ○ motd-news.service - Message of the Day Loaded: loaded (/usr/lib/systemd/system/motd-news.service; static) Active: inactive (dead) since Thu 2026-04-02 14:45:14 UTC; 6min ago TriggeredBy: ● motd-news.timer Docs: man:update-motd(8) Main PID: 99972 (code=exited, status=0/SUCCESS) CPU: 2ms Apr 02 14:45:14 np0000163014 systemd[1]: Starting motd-news.service - Message of the Day... Apr 02 14:45:14 np0000163014 systemd[1]: motd-news.service: Deactivated successfully. Apr 02 14:45:14 np0000163014 systemd[1]: Finished motd-news.service - Message of the Day. ● mysql.service - MySQL Community Server Loaded: loaded (/usr/lib/systemd/system/mysql.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:39:04 UTC; 12min ago Main PID: 62164 (mysqld) Status: "Server is operational" Tasks: 259 (limit: 77075) Memory: 1.2G (peak: 1.2G) CPU: 2min 2.679s CGroup: /system.slice/mysql.service └─62164 /usr/sbin/mysqld Apr 02 14:39:03 np0000163014 systemd[1]: Starting mysql.service - MySQL Community Server... Apr 02 14:39:04 np0000163014 systemd[1]: Started mysql.service - MySQL Community Server. ○ netplan-ovs-cleanup.service - OpenVSwitch configuration for cleanup Loaded: loaded (/run/systemd/system/netplan-ovs-cleanup.service; enabled-runtime; preset: enabled) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:05 UTC; 30min ago Apr 02 14:21:05 np0000163014 systemd[1]: netplan-ovs-cleanup.service - OpenVSwitch configuration for cleanup was skipped because of an unmet condition check (ConditionFileIsExecutable=/usr/bin/ovs-vsctl). Apr 02 14:21:05 np0000163014 systemd[1]: netplan-ovs-cleanup.service - OpenVSwitch configuration for cleanup was skipped because of an unmet condition check (ConditionFileIsExecutable=/usr/bin/ovs-vsctl). Apr 02 14:21:05 np0000163014 systemd[1]: netplan-ovs-cleanup.service - OpenVSwitch configuration for cleanup was skipped because of an unmet condition check (ConditionFileIsExecutable=/usr/bin/ovs-vsctl). ○ networkd-dispatcher.service - Dispatcher daemon for systemd-networkd Loaded: loaded (/usr/lib/systemd/system/networkd-dispatcher.service; enabled; preset: enabled) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:10 UTC; 30min ago Apr 02 14:21:10 np0000163014 systemd[1]: networkd-dispatcher.service - Dispatcher daemon for systemd-networkd was skipped because no trigger condition checks were met. ● nfs-blkmap.service - pNFS block layout mapping daemon Loaded: loaded (/usr/lib/systemd/system/nfs-blkmap.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:36:36 UTC; 15min ago Main PID: 53946 (blkmapd) Tasks: 1 (limit: 77075) Memory: 316.0K (peak: 1.5M) CPU: 13ms CGroup: /system.slice/nfs-blkmap.service └─53946 /usr/sbin/blkmapd Apr 02 14:36:36 np0000163014 systemd[1]: Starting nfs-blkmap.service - pNFS block layout mapping daemon... Apr 02 14:36:36 np0000163014 blkmapd[53946]: open pipe file /run/rpc_pipefs/nfs/blocklayout failed: No such file or directory Apr 02 14:36:36 np0000163014 systemd[1]: Started nfs-blkmap.service - pNFS block layout mapping daemon. ● nfs-idmapd.service - NFSv4 ID-name mapping service Loaded: loaded (/usr/lib/systemd/system/nfs-idmapd.service; static) Active: active (running) since Thu 2026-04-02 14:36:36 UTC; 15min ago Main PID: 53949 (rpc.idmapd) Tasks: 1 (limit: 77075) Memory: 408.0K (peak: 1.5M) CPU: 11ms CGroup: /system.slice/nfs-idmapd.service └─53949 /usr/sbin/rpc.idmapd Apr 02 14:36:36 np0000163014 systemd[1]: Starting nfs-idmapd.service - NFSv4 ID-name mapping service... Apr 02 14:36:36 np0000163014 rpc.idmapd[53949]: Setting log level to 0 Apr 02 14:36:36 np0000163014 systemd[1]: Started nfs-idmapd.service - NFSv4 ID-name mapping service. ● nfs-mountd.service - NFS Mount Daemon Loaded: loaded (/usr/lib/systemd/system/nfs-mountd.service; static) Active: active (running) since Thu 2026-04-02 14:36:36 UTC; 15min ago Main PID: 53960 (rpc.mountd) Tasks: 1 (limit: 77075) Memory: 904.0K (peak: 1.5M) CPU: 11ms CGroup: /system.slice/nfs-mountd.service └─53960 /usr/sbin/rpc.mountd Apr 02 14:36:36 np0000163014 systemd[1]: Starting nfs-mountd.service - NFS Mount Daemon... Apr 02 14:36:36 np0000163014 rpc.mountd[53960]: Version 2.6.4 starting Apr 02 14:36:36 np0000163014 systemd[1]: Started nfs-mountd.service - NFS Mount Daemon. ● nfs-server.service - NFS server and services Loaded: loaded (/usr/lib/systemd/system/nfs-server.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:36:37 UTC; 15min ago Main PID: 53967 (code=exited, status=0/SUCCESS) CPU: 22ms Apr 02 14:36:36 np0000163014 systemd[1]: Starting nfs-server.service - NFS server and services... Apr 02 14:36:36 np0000163014 exportfs[53964]: exportfs: can't open /etc/exports for reading Apr 02 14:36:37 np0000163014 systemd[1]: Finished nfs-server.service - NFS server and services. ○ nfs-utils.service - NFS server and client services Loaded: loaded (/usr/lib/systemd/system/nfs-utils.service; static) Active: inactive (dead) ● nfsdcld.service - NFSv4 Client Tracking Daemon Loaded: loaded (/usr/lib/systemd/system/nfsdcld.service; static) Active: active (running) since Thu 2026-04-02 14:36:36 UTC; 15min ago Main PID: 53961 (nfsdcld) Tasks: 1 (limit: 77075) Memory: 684.0K (peak: 1.5M) CPU: 16ms CGroup: /system.slice/nfsdcld.service └─53961 /usr/sbin/nfsdcld Apr 02 14:36:36 np0000163014 systemd[1]: Starting nfsdcld.service - NFSv4 Client Tracking Daemon... Apr 02 14:36:36 np0000163014 systemd[1]: Started nfsdcld.service - NFSv4 Client Tracking Daemon. ● nmbd.service - Samba NMB Daemon Loaded: loaded (/usr/lib/systemd/system/nmbd.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:36:41 UTC; 15min ago Docs: man:nmbd(8) man:samba(7) man:smb.conf(5) Main PID: 54587 (nmbd) Status: "nmbd: ready to serve connections..." Tasks: 1 (limit: 77075) Memory: 3.2M (peak: 4.7M) CPU: 226ms CGroup: /system.slice/nmbd.service └─54587 /usr/sbin/nmbd --foreground --no-process-group Apr 02 14:36:41 np0000163014 systemd[1]: Starting nmbd.service - Samba NMB Daemon... Apr 02 14:36:41 np0000163014 (nmbd)[54587]: nmbd.service: Referenced but unset environment variable evaluates to an empty string: NMBDOPTIONS Apr 02 14:36:41 np0000163014 systemd[1]: Started nmbd.service - Samba NMB Daemon. ○ open-iscsi.service - Login to default iSCSI targets Loaded: loaded (/usr/lib/systemd/system/open-iscsi.service; enabled; preset: enabled) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:iscsiadm(8) man:iscsid(8) Apr 02 14:21:10 np0000163014 systemd[1]: open-iscsi.service - Login to default iSCSI targets was skipped because no trigger condition checks were met. ● openvswitch-switch.service - Open vSwitch Loaded: loaded (/usr/lib/systemd/system/openvswitch-switch.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:45:13 UTC; 6min ago Main PID: 99657 (code=exited, status=0/SUCCESS) CPU: 2ms Apr 02 14:45:13 np0000163014 systemd[1]: Starting openvswitch-switch.service - Open vSwitch... Apr 02 14:45:13 np0000163014 systemd[1]: Finished openvswitch-switch.service - Open vSwitch. ● ovn-central.service - Open Virtual Network central components Loaded: loaded (/usr/lib/systemd/system/ovn-central.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:45:14 UTC; 6min ago Main PID: 99983 (code=exited, status=0/SUCCESS) CPU: 10ms Apr 02 14:45:14 np0000163014 systemd[1]: Starting ovn-central.service - Open Virtual Network central components... Apr 02 14:45:14 np0000163014 systemd[1]: Finished ovn-central.service - Open Virtual Network central components. ● ovn-controller-vtep.service - Open Virtual Network VTEP gateway controller daemon Loaded: loaded (/usr/lib/systemd/system/ovn-controller-vtep.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:45:13 UTC; 6min ago Main PID: 99715 (ovn-controller-) Tasks: 1 (limit: 77075) Memory: 864.0K (peak: 2.3M) CPU: 61ms CGroup: /system.slice/ovn-controller-vtep.service └─99715 ovn-controller-vtep -vconsole:emer -vsyslog:err -vfile:info --vtep-db=/var/run/openvswitch/db.sock --ovnsb-db=/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-controller-vtep.log --pidfile=/var/run/ovn/ovn-controller-vtep.pid --detach Apr 02 14:45:13 np0000163014 systemd[1]: Starting ovn-controller-vtep.service - Open Virtual Network VTEP gateway controller daemon... Apr 02 14:45:13 np0000163014 (ovn-ctl)[99665]: ovn-controller-vtep.service: Referenced but unset environment variable evaluates to an empty string: OVN_CTL_OPTS Apr 02 14:45:13 np0000163014 ovn-ctl[99665]: * Starting ovn-controller-vtep Apr 02 14:45:13 np0000163014 systemd[1]: Started ovn-controller-vtep.service - Open Virtual Network VTEP gateway controller daemon. ● ovn-controller.service - Open Virtual Network host control daemon Loaded: loaded (/usr/lib/systemd/system/ovn-controller.service; static) Active: active (running) since Thu 2026-04-02 14:45:17 UTC; 6min ago Main PID: 100402 (ovn-controller) Tasks: 5 (limit: 77075) Memory: 5.5M (peak: 6.0M) CPU: 308ms CGroup: /system.slice/ovn-controller.service └─100402 ovn-controller unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --no-chdir --log-file=/var/log/ovn/ovn-controller.log --pidfile=/var/run/ovn/ovn-controller.pid --detach Apr 02 14:45:17 np0000163014 systemd[1]: Starting ovn-controller.service - Open Virtual Network host control daemon... Apr 02 14:45:17 np0000163014 (ovn-ctl)[100373]: ovn-controller.service: Referenced but unset environment variable evaluates to an empty string: OVN_CTL_OPTS Apr 02 14:45:17 np0000163014 ovn-ctl[100373]: * Starting ovn-controller Apr 02 14:45:17 np0000163014 systemd[1]: Started ovn-controller.service - Open Virtual Network host control daemon. ● ovn-host.service - Open Virtual Network host components Loaded: loaded (/usr/lib/systemd/system/ovn-host.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:34:16 UTC; 17min ago Main PID: 40319 (code=exited, status=0/SUCCESS) CPU: 8ms Apr 02 14:34:16 np0000163014 systemd[1]: Starting ovn-host.service - Open Virtual Network host components... Apr 02 14:34:16 np0000163014 systemd[1]: Finished ovn-host.service - Open Virtual Network host components. ● ovn-northd.service - Open Virtual Network central control daemon Loaded: loaded (/usr/lib/systemd/system/ovn-northd.service; static) Active: active (running) since Thu 2026-04-02 14:45:15 UTC; 6min ago Main PID: 100136 (ovn-northd) Tasks: 3 (limit: 77075) Memory: 3.0M (peak: 3.3M) CPU: 259ms CGroup: /system.slice/ovn-northd.service └─100136 ovn-northd -vconsole:emer -vsyslog:err -vfile:info --ovnnb-db=unix:/var/run/ovn/ovnnb_db.sock --ovnsb-db=unix:/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-northd.log --pidfile=/var/run/ovn/ovn-northd.pid --detach Apr 02 14:45:15 np0000163014 systemd[1]: Starting ovn-northd.service - Open Virtual Network central control daemon... Apr 02 14:45:15 np0000163014 (ovn-ctl)[100061]: ovn-northd.service: Referenced but unset environment variable evaluates to an empty string: OVN_CTL_OPTS Apr 02 14:45:15 np0000163014 ovn-ctl[100061]: * Starting ovn-northd Apr 02 14:45:15 np0000163014 systemd[1]: Started ovn-northd.service - Open Virtual Network central control daemon. ● ovn-ovsdb-server-nb.service - Open vSwitch database server for OVN Northbound database Loaded: loaded (/usr/lib/systemd/system/ovn-ovsdb-server-nb.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:45:15 UTC; 6min ago Main PID: 100056 (ovsdb-server) Tasks: 1 (limit: 77075) Memory: 4.8M (peak: 5.8M) CPU: 3.232s CGroup: /system.slice/ovn-ovsdb-server-nb.service └─100056 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-nb.log --remote=punix:/var/run/ovn/ovnnb_db.sock --pidfile=/var/run/ovn/ovnnb_db.pid --unixctl=/var/run/ovn/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=db:OVN_Northbound,SSL,private_key --certificate=db:OVN_Northbound,SSL,certificate --ca-cert=db:OVN_Northbound,SSL,ca_cert --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers /var/lib/ovn/ovnnb_db.db Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05760|reconnect|DBG|ssl:199.204.45.109:42614: entering ACTIVE Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05761|poll_loop|DBG|wakeup due to 0-ms timeout at ../lib/stream-ssl.c:844 (0% CPU usage) Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05762|poll_loop|DBG|wakeup due to [POLLIN] on fd 31 (199.204.45.109:6641<->199.204.45.109:51512) at ../lib/stream-ssl.c:842 (0% CPU usage) Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05763|stream_ssl|DBG|server27<--ssl:199.204.45.109:51512 type 256 (5 bytes) Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05764|stream_ssl|DBG|server27<--ssl:199.204.45.109:51512 type 257 (1 bytes) Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05765|jsonrpc|DBG|ssl:199.204.45.109:51512: received request, method="echo", params=[], id="echo" Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05766|jsonrpc|DBG|ssl:199.204.45.109:51512: send reply, result=[], id="echo" Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05767|stream_ssl|DBG|server27-->ssl:199.204.45.109:51512 type 256 (5 bytes) Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05768|stream_ssl|DBG|server27-->ssl:199.204.45.109:51512 type 257 (1 bytes) Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05769|poll_loop|DBG|wakeup due to 0-ms timeout at ../lib/stream-ssl.c:844 (0% CPU usage) ● ovn-ovsdb-server-sb.service - Open vSwitch database server for OVN Southbound database Loaded: loaded (/usr/lib/systemd/system/ovn-ovsdb-server-sb.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:45:15 UTC; 6min ago Main PID: 100058 (ovsdb-server) Tasks: 1 (limit: 77075) Memory: 5.3M (peak: 5.8M) CPU: 2.908s CGroup: /system.slice/ovn-ovsdb-server-sb.service └─100058 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-sb.log --remote=punix:/var/run/ovn/ovnsb_db.sock --pidfile=/var/run/ovn/ovnsb_db.pid --unixctl=/var/run/ovn/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=db:OVN_Southbound,SSL,private_key --certificate=db:OVN_Southbound,SSL,certificate --ca-cert=db:OVN_Southbound,SSL,ca_cert --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers /var/lib/ovn/ovnsb_db.db Apr 02 14:51:50 np0000163014 ovsdb-server[100058]: ovs|04166|poll_loop|DBG|wakeup due to 2163-ms timeout at ../ovsdb/ovsdb-server.c:400 (0% CPU usage) Apr 02 14:51:52 np0000163014 ovsdb-server[100058]: ovs|04167|poll_loop|DBG|wakeup due to 2530-ms timeout at ../ovsdb/ovsdb-server.c:400 (0% CPU usage) Apr 02 14:51:52 np0000163014 ovsdb-server[100058]: ovs|04168|poll_loop|DBG|wakeup due to [POLLIN] on fd 22 (199.204.45.109:6642<->199.204.45.109:53410) at ../lib/stream-ssl.c:842 (0% CPU usage) Apr 02 14:51:52 np0000163014 ovsdb-server[100058]: ovs|04169|stream_ssl|DBG|server0<--ssl:199.204.45.109:53410 type 256 (5 bytes) Apr 02 14:51:52 np0000163014 ovsdb-server[100058]: ovs|04170|stream_ssl|DBG|server0<--ssl:199.204.45.109:53410 type 257 (1 bytes) Apr 02 14:51:52 np0000163014 ovsdb-server[100058]: ovs|04171|jsonrpc|DBG|ssl:199.204.45.109:53410: received request, method="echo", params=[], id="echo" Apr 02 14:51:52 np0000163014 ovsdb-server[100058]: ovs|04172|jsonrpc|DBG|ssl:199.204.45.109:53410: send reply, result=[], id="echo" Apr 02 14:51:52 np0000163014 ovsdb-server[100058]: ovs|04173|stream_ssl|DBG|server0-->ssl:199.204.45.109:53410 type 256 (5 bytes) Apr 02 14:51:52 np0000163014 ovsdb-server[100058]: ovs|04174|stream_ssl|DBG|server0-->ssl:199.204.45.109:53410 type 257 (1 bytes) Apr 02 14:51:52 np0000163014 ovsdb-server[100058]: ovs|04175|poll_loop|DBG|wakeup due to 0-ms timeout at ../lib/stream-ssl.c:844 (0% CPU usage) ● ovs-record-hostname.service - Open vSwitch Record Hostname Loaded: loaded (/usr/lib/systemd/system/ovs-record-hostname.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:45:13 UTC; 6min ago Main PID: 99662 (code=exited, status=0/SUCCESS) CPU: 33ms Apr 02 14:45:13 np0000163014 systemd[1]: Starting ovs-record-hostname.service - Open vSwitch Record Hostname... Apr 02 14:45:13 np0000163014 ovs-vsctl[99694]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --no-wait add Open_vSwitch . external-ids hostname=np0000163014.novalocal Apr 02 14:45:13 np0000163014 systemd[1]: Finished ovs-record-hostname.service - Open vSwitch Record Hostname. ● ovs-vswitchd.service - Open vSwitch Forwarding Unit Loaded: loaded (/usr/lib/systemd/system/ovs-vswitchd.service; static) Active: active (running) since Thu 2026-04-02 14:45:13 UTC; 6min ago Main PID: 99627 (ovs-vswitchd) Tasks: 23 (limit: 77075) Memory: 183.7M (peak: 184.7M) CPU: 1.967s CGroup: /system.slice/ovs-vswitchd.service └─99627 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach Apr 02 14:45:13 np0000163014 systemd[1]: Starting ovs-vswitchd.service - Open vSwitch Forwarding Unit... Apr 02 14:45:13 np0000163014 (ovs-ctl)[99589]: ovs-vswitchd.service: Referenced but unset environment variable evaluates to an empty string: OVS_CTL_OPTS Apr 02 14:45:13 np0000163014 ovs-ctl[99589]: * Starting ovs-vswitchd Apr 02 14:45:13 np0000163014 ovs-ctl[99589]: * Enabling remote OVSDB managers Apr 02 14:45:13 np0000163014 systemd[1]: Started ovs-vswitchd.service - Open vSwitch Forwarding Unit. ● ovsdb-server.service - Open vSwitch Database Unit Loaded: loaded (/usr/lib/systemd/system/ovsdb-server.service; static) Active: active (running) since Thu 2026-04-02 14:45:13 UTC; 6min ago Main PID: 99578 (ovsdb-server) Tasks: 1 (limit: 77075) Memory: 2.1M (peak: 4.9M) CPU: 3.765s CGroup: /system.slice/ovsdb-server.service └─99578 ovsdb-server /etc/openvswitch/conf.db -vconsole:emer -vsyslog:err -vfile:info --remote=punix:/var/run/openvswitch/db.sock --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --no-chdir --log-file=/var/log/openvswitch/ovsdb-server.log --pidfile=/var/run/openvswitch/ovsdb-server.pid --detach Apr 02 14:45:12 np0000163014 systemd[1]: Starting ovsdb-server.service - Open vSwitch Database Unit... Apr 02 14:45:12 np0000163014 (ovs-ctl)[99544]: ovsdb-server.service: Referenced but unset environment variable evaluates to an empty string: OVS_CTL_OPTS Apr 02 14:45:13 np0000163014 ovs-ctl[99544]: * Starting ovsdb-server Apr 02 14:45:13 np0000163014 ovs-vsctl[99579]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --no-wait -- init -- set Open_vSwitch . db-version=8.5.1 Apr 02 14:45:13 np0000163014 ovs-vsctl[99584]: ovs|00001|vsctl|INFO|Called as ovs-vsctl --no-wait set Open_vSwitch . ovs-version=3.3.4 "external-ids:system-id=\"42afca65-e149-4810-9fbe-0f474a299117\"" "external-ids:rundir=\"/var/run/openvswitch\"" "system-type=\"ubuntu\"" "system-version=\"24.04\"" Apr 02 14:45:13 np0000163014 ovs-ctl[99544]: * Configuring Open vSwitch system IDs Apr 02 14:45:13 np0000163014 ovs-ctl[99544]: * Enabling remote OVSDB managers Apr 02 14:45:13 np0000163014 systemd[1]: Started ovsdb-server.service - Open vSwitch Database Unit. ● polkit.service - Authorization Manager Loaded: loaded (/usr/lib/systemd/system/polkit.service; static) Active: active (running) since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:polkit(8) Main PID: 742 (polkitd) Tasks: 4 (limit: 77075) Memory: 3.5M (peak: 4.4M) CPU: 254ms CGroup: /system.slice/polkit.service └─742 /usr/lib/polkit-1/polkitd --no-debug Apr 02 14:34:30 np0000163014 polkitd[742]: Reloading rules Apr 02 14:34:30 np0000163014 polkitd[742]: Collecting garbage unconditionally... Apr 02 14:34:30 np0000163014 polkitd[742]: Loading rules from directory /etc/polkit-1/rules.d Apr 02 14:34:30 np0000163014 polkitd[742]: Loading rules from directory /usr/share/polkit-1/rules.d Apr 02 14:34:30 np0000163014 polkitd[742]: Finished loading, compiling and executing 5 rules Apr 02 14:34:30 np0000163014 polkitd[742]: Reloading rules Apr 02 14:34:30 np0000163014 polkitd[742]: Collecting garbage unconditionally... Apr 02 14:34:30 np0000163014 polkitd[742]: Loading rules from directory /etc/polkit-1/rules.d Apr 02 14:34:30 np0000163014 polkitd[742]: Loading rules from directory /usr/share/polkit-1/rules.d Apr 02 14:34:30 np0000163014 polkitd[742]: Finished loading, compiling and executing 5 rules ● postgresql.service - PostgreSQL RDBMS Loaded: loaded (/usr/lib/systemd/system/postgresql.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:28:13 UTC; 23min ago Main PID: 13315 (code=exited, status=0/SUCCESS) CPU: 2ms Apr 02 14:28:13 np0000163014 systemd[1]: Starting postgresql.service - PostgreSQL RDBMS... Apr 02 14:28:13 np0000163014 systemd[1]: Finished postgresql.service - PostgreSQL RDBMS. ● qemu-kvm.service - QEMU KVM preparation - module, ksm, hugepages Loaded: loaded (/usr/lib/systemd/system/qemu-kvm.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:34:39 UTC; 17min ago Main PID: 41894 (code=exited, status=0/SUCCESS) CPU: 31ms Apr 02 14:34:39 np0000163014 systemd[1]: Starting qemu-kvm.service - QEMU KVM preparation - module, ksm, hugepages... Apr 02 14:34:39 np0000163014 systemd[1]: Finished qemu-kvm.service - QEMU KVM preparation - module, ksm, hugepages. ● rabbitmq-server.service - RabbitMQ Messaging Server Loaded: loaded (/usr/lib/systemd/system/rabbitmq-server.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:31:38 UTC; 20min ago Main PID: 25837 (beam.smp) Tasks: 52 (limit: 77075) Memory: 126.3M (peak: 164.1M) CPU: 29.722s CGroup: /system.slice/rabbitmq-server.service ├─25837 /usr/lib/erlang/erts-13.2.2.5/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -- -root /usr/lib/erlang -bindir /usr/lib/erlang/erts-13.2.2.5/bin -progname erl -- -home /var/lib/rabbitmq -- -pa "" -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger "[]" -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -enable-feature maybe_expr ├─25847 erl_child_setup 65536 ├─25955 /usr/lib/erlang/erts-13.2.2.5/bin/inet_gethost 4 ├─25956 /usr/lib/erlang/erts-13.2.2.5/bin/inet_gethost 4 └─25969 /bin/sh -s rabbit_disk_monitor Apr 02 14:31:30 np0000163014 systemd[1]: Starting rabbitmq-server.service - RabbitMQ Messaging Server... Apr 02 14:31:38 np0000163014 systemd[1]: Started rabbitmq-server.service - RabbitMQ Messaging Server. ○ rc-local.service - /etc/rc.local Compatibility Loaded: loaded (/usr/lib/systemd/system/rc-local.service; static) Drop-In: /usr/lib/systemd/system/rc-local.service.d └─debian.conf Active: inactive (dead) Docs: man:systemd-rc-local-generator(8) ○ rescue.service - Rescue Shell Loaded: loaded (/usr/lib/systemd/system/rescue.service; static) Active: inactive (dead) Docs: man:sulogin(8) ○ rpc-gssd.service - RPC security service for NFS client and server Loaded: loaded (/usr/lib/systemd/system/rpc-gssd.service; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:36:38 UTC; 15min ago Apr 02 14:36:35 np0000163014 systemd[1]: rpc-gssd.service - RPC security service for NFS client and server was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). Apr 02 14:36:36 np0000163014 systemd[1]: rpc-gssd.service - RPC security service for NFS client and server was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). Apr 02 14:36:38 np0000163014 systemd[1]: rpc-gssd.service - RPC security service for NFS client and server was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). ● rpc-statd-notify.service - Notify NFS peers of a restart Loaded: loaded (/usr/lib/systemd/system/rpc-statd-notify.service; static) Active: active (exited) since Thu 2026-04-02 14:36:35 UTC; 15min ago CPU: 11ms Apr 02 14:36:35 np0000163014 systemd[1]: Starting rpc-statd-notify.service - Notify NFS peers of a restart... Apr 02 14:36:35 np0000163014 sm-notify[53853]: Version 2.6.4 starting Apr 02 14:36:35 np0000163014 systemd[1]: Started rpc-statd-notify.service - Notify NFS peers of a restart. ● rpc-statd.service - NFS status monitor for NFSv2/3 locking. Loaded: loaded (/usr/lib/systemd/system/rpc-statd.service; static) Active: active (running) since Thu 2026-04-02 14:36:36 UTC; 15min ago Main PID: 53951 (rpc.statd) Tasks: 1 (limit: 77075) Memory: 508.0K (peak: 1.5M) CPU: 15ms CGroup: /system.slice/rpc-statd.service └─53951 /usr/sbin/rpc.statd Apr 02 14:36:36 np0000163014 systemd[1]: Starting rpc-statd.service - NFS status monitor for NFSv2/3 locking.... Apr 02 14:36:36 np0000163014 rpc.statd[53951]: Version 2.6.4 starting Apr 02 14:36:36 np0000163014 rpc.statd[53951]: Flags: TI-RPC Apr 02 14:36:36 np0000163014 rpc.statd[53951]: Failed to read /var/lib/nfs/state: Success Apr 02 14:36:36 np0000163014 rpc.statd[53951]: Initializing NSM state Apr 02 14:36:36 np0000163014 systemd[1]: Started rpc-statd.service - NFS status monitor for NFSv2/3 locking.. ○ rpc-svcgssd.service - RPC security service for NFS server Loaded: loaded (/usr/lib/systemd/system/rpc-svcgssd.service; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:36:38 UTC; 15min ago Apr 02 14:36:36 np0000163014 systemd[1]: rpc-svcgssd.service - RPC security service for NFS server was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). Apr 02 14:36:38 np0000163014 systemd[1]: rpc-svcgssd.service - RPC security service for NFS server was skipped because of an unmet condition check (ConditionPathExists=/etc/krb5.keytab). ● rpcbind.service - RPC bind portmap service Loaded: loaded (/usr/lib/systemd/system/rpcbind.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:36:32 UTC; 15min ago TriggeredBy: ● rpcbind.socket Docs: man:rpcbind(8) Main PID: 53314 (rpcbind) Tasks: 1 (limit: 77075) Memory: 628.0K (peak: 1.5M) CPU: 19ms CGroup: /system.slice/rpcbind.service └─53314 /sbin/rpcbind -f -w Apr 02 14:36:32 np0000163014 systemd[1]: Starting rpcbind.service - RPC bind portmap service... Apr 02 14:36:32 np0000163014 systemd[1]: Started rpcbind.service - RPC bind portmap service. ● rsyslog.service - System Logging Service Loaded: loaded (/usr/lib/systemd/system/rsyslog.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:50:20 UTC; 1min 34s ago TriggeredBy: ● syslog.socket Docs: man:rsyslogd(8) man:rsyslog.conf(5) https://www.rsyslog.com/doc/ Main PID: 124176 (rsyslogd) Tasks: 4 (limit: 77075) Memory: 4.8M (peak: 5.2M) CPU: 689ms CGroup: /system.slice/rsyslog.service └─124176 /usr/sbin/rsyslogd -n -iNONE Apr 02 14:50:20 np0000163014 systemd[1]: Starting rsyslog.service - System Logging Service... Apr 02 14:50:20 np0000163014 rsyslogd[124176]: imuxsock: Acquired UNIX socket '/run/systemd/journal/syslog' (fd 3) from systemd. [v8.2312.0] Apr 02 14:50:20 np0000163014 rsyslogd[124176]: rsyslogd's groupid changed to 104 Apr 02 14:50:20 np0000163014 systemd[1]: Started rsyslog.service - System Logging Service. Apr 02 14:50:20 np0000163014 rsyslogd[124176]: rsyslogd's userid changed to 104 Apr 02 14:50:20 np0000163014 rsyslogd[124176]: [origin software="rsyslogd" swVersion="8.2312.0" x-pid="124176" x-info="https://www.rsyslog.com"] start ● rtslib-fb-targetctl.service - Restore LIO kernel target configuration Loaded: loaded (/usr/lib/systemd/system/rtslib-fb-targetctl.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:33:39 UTC; 18min ago Main PID: 35788 (code=exited, status=0/SUCCESS) CPU: 111ms Apr 02 14:33:39 np0000163014 systemd[1]: Starting rtslib-fb-targetctl.service - Restore LIO kernel target configuration... Apr 02 14:33:39 np0000163014 target[35788]: No saved config file at /etc/rtslib-fb-target/saveconfig.json, ok, exiting Apr 02 14:33:39 np0000163014 systemd[1]: Finished rtslib-fb-targetctl.service - Restore LIO kernel target configuration. ○ samba-ad-dc.service - Samba AD Daemon Loaded: loaded (/usr/lib/systemd/system/samba-ad-dc.service; enabled; preset: enabled) Active: inactive (dead) (Result: exec-condition) since Thu 2026-04-02 14:36:41 UTC; 15min ago Condition: start condition unmet at Thu 2026-04-02 14:36:41 UTC; 15min ago Docs: man:samba(8) man:samba(7) man:smb.conf(5) CPU: 28ms Apr 02 14:36:41 np0000163014 systemd[1]: Starting samba-ad-dc.service - Samba AD Daemon... Apr 02 14:36:41 np0000163014 systemd[1]: samba-ad-dc.service: Skipped due to 'exec-condition'. Apr 02 14:36:41 np0000163014 systemd[1]: Condition check resulted in samba-ad-dc.service - Samba AD Daemon being skipped. ● serial-getty@ttyS0.service - Serial Getty on ttyS0 Loaded: loaded (/usr/lib/systemd/system/serial-getty@.service; enabled-runtime; preset: enabled) Active: active (running) since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:agetty(8) man:systemd-getty-generator(8) https://0pointer.de/blog/projects/serial-console.html Main PID: 715 (agetty) Tasks: 1 (limit: 77075) Memory: 244.0K (peak: 1.8M) CPU: 26ms CGroup: /system.slice/system-serial\x2dgetty.slice/serial-getty@ttyS0.service └─715 /sbin/agetty -o "-p -- \\u" --keep-baud 115200,57600,38400,9600 - vt220 Apr 02 14:21:10 np0000163014 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. ● smbd.service - Samba SMB Daemon Loaded: loaded (/usr/lib/systemd/system/smbd.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:36:40 UTC; 15min ago Docs: man:smbd(8) man:samba(7) man:smb.conf(5) Main PID: 54520 (smbd) Status: "smbd: ready to serve connections..." Tasks: 3 (limit: 77075) Memory: 8.7M (peak: 9.5M) CPU: 107ms CGroup: /system.slice/smbd.service ├─54520 /usr/sbin/smbd --foreground --no-process-group ├─54523 "smbd: notifyd" . └─54524 "smbd: cleanupd " Apr 02 14:36:40 np0000163014 systemd[1]: Starting smbd.service - Samba SMB Daemon... Apr 02 14:36:40 np0000163014 (smbd)[54520]: smbd.service: Referenced but unset environment variable evaluates to an empty string: SMBDOPTIONS Apr 02 14:36:40 np0000163014 systemd[1]: Started smbd.service - Samba SMB Daemon. ● ssh-keygen.service - OpenSSH Server Key Generation Loaded: loaded (/usr/lib/systemd/system/ssh-keygen.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:21:10 UTC; 30min ago Main PID: 702 (code=exited, status=0/SUCCESS) CPU: 90ms Apr 02 14:21:10 np0000163014 runtime-ssh-host-keys.sh[702]: + for key in dsa ecdsa ed25519 rsa Apr 02 14:21:10 np0000163014 runtime-ssh-host-keys.sh[702]: + FILE=/etc/ssh/ssh_host_ecdsa_key Apr 02 14:21:10 np0000163014 runtime-ssh-host-keys.sh[702]: + '[' -f /etc/ssh/ssh_host_ecdsa_key ']' Apr 02 14:21:10 np0000163014 runtime-ssh-host-keys.sh[702]: + for key in dsa ecdsa ed25519 rsa Apr 02 14:21:10 np0000163014 runtime-ssh-host-keys.sh[702]: + FILE=/etc/ssh/ssh_host_ed25519_key Apr 02 14:21:10 np0000163014 runtime-ssh-host-keys.sh[702]: + '[' -f /etc/ssh/ssh_host_ed25519_key ']' Apr 02 14:21:10 np0000163014 runtime-ssh-host-keys.sh[702]: + for key in dsa ecdsa ed25519 rsa Apr 02 14:21:10 np0000163014 runtime-ssh-host-keys.sh[702]: + FILE=/etc/ssh/ssh_host_rsa_key Apr 02 14:21:10 np0000163014 runtime-ssh-host-keys.sh[702]: + '[' -f /etc/ssh/ssh_host_rsa_key ']' Apr 02 14:21:10 np0000163014 systemd[1]: Finished ssh-keygen.service - OpenSSH Server Key Generation. ● ssh.service - OpenBSD Secure Shell server Loaded: loaded (/usr/lib/systemd/system/ssh.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:21:10 UTC; 30min ago TriggeredBy: ● ssh.socket Docs: man:sshd(8) man:sshd_config(5) Main PID: 744 (sshd) Tasks: 42 (limit: 77075) Memory: 43.5M (peak: 43.6M) CPU: 917ms CGroup: /system.slice/ssh.service ├─ 744 "sshd: /usr/sbin/sshd -D [listener] 26 of 10-100 startups" ├─127825 "sshd: root [priv]" ├─127826 "sshd: root [net]" ├─127848 "sshd: [accepted]" ├─127849 "sshd: [net]" ├─127850 "sshd: [accepted]" ├─127851 "sshd: [net]" ├─127852 "sshd: [accepted]" ├─127853 "sshd: [net]" ├─127854 "sshd: [accepted]" ├─127855 "sshd: [net]" ├─127891 "sshd: [accepted]" ├─127893 "sshd: [net]" ├─127897 "sshd: [accepted]" ├─127898 "sshd: [net]" ├─127899 "sshd: [accepted]" ├─127900 "sshd: [net]" ├─127904 "sshd: [accepted]" ├─127905 "sshd: [net]" ├─127928 "sshd: [accepted]" ├─127988 "sshd: [accepted]" ├─127999 "sshd: [accepted]" ├─128090 "sshd: [accepted]" ├─128134 "sshd: [accepted]" ├─128135 "sshd: [net]" ├─128136 "sshd: [accepted]" ├─128137 "sshd: [accepted]" ├─128138 "sshd: [net]" ├─128141 "sshd: [accepted]" ├─128292 "sshd: [accepted]" ├─128364 "sshd: [accepted]" ├─128365 "sshd: [net]" ├─128604 "sshd: [accepted]" ├─129705 "sshd: [accepted]" ├─129861 "sshd: [accepted]" ├─129864 "sshd: [net]" ├─129915 "sshd: [accepted]" ├─130024 "sshd: [accepted]" ├─130095 "sshd: [accepted]" ├─130096 "sshd: [net]" ├─130127 "sshd: [accepted]" └─130128 "sshd: [net]" Apr 02 14:50:35 np0000163014 sshd[125455]: Failed password for root from 172.191.157.64 port 59196 ssh2 Apr 02 14:50:37 np0000163014 sshd[125455]: Received disconnect from 172.191.157.64 port 59196:11: Bye Bye [preauth] Apr 02 14:50:37 np0000163014 sshd[125455]: Disconnected from authenticating user root 172.191.157.64 port 59196 [preauth] Apr 02 14:51:09 np0000163014 sshd[127994]: Invalid user apache from 101.126.135.131 port 42926 Apr 02 14:51:09 np0000163014 sshd[744]: error: beginning MaxStartups throttling Apr 02 14:51:09 np0000163014 sshd[744]: drop connection #13 from [101.126.135.131]:42990 on [199.204.45.109]:22 past MaxStartups Apr 02 14:51:09 np0000163014 sshd[127994]: pam_unix(sshd:auth): check pass; user unknown Apr 02 14:51:09 np0000163014 sshd[127994]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=101.126.135.131 Apr 02 14:51:11 np0000163014 sshd[127994]: Failed password for invalid user apache from 101.126.135.131 port 42926 ssh2 Apr 02 14:51:14 np0000163014 sshd[127994]: Connection closed by invalid user apache 101.126.135.131 port 42926 [preauth] ○ ssl-cert.service - Generate snakeoil SSL keypair Loaded: loaded (/usr/lib/systemd/system/ssl-cert.service; enabled; preset: enabled) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:28:02 UTC; 23min ago Apr 02 14:28:02 np0000163014 systemd[1]: ssl-cert.service - Generate snakeoil SSL keypair was skipped because of an unmet condition check (ConditionPathExists=!/etc/ssl/private/ssl-cert-snakeoil.key). ● stack-volumes-lvmdriver-1-backing-file.service - Activate LVM backing file /opt/stack/data/stack-volumes-lvmdriver-1-backing-file Loaded: loaded (/etc/systemd/system/stack-volumes-lvmdriver-1-backing-file.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:41:37 UTC; 10min ago Main PID: 79733 (code=exited, status=0/SUCCESS) CPU: 13ms Apr 02 14:41:37 np0000163014 systemd[1]: Starting stack-volumes-lvmdriver-1-backing-file.service - Activate LVM backing file /opt/stack/data/stack-volumes-lvmdriver-1-backing-file... Apr 02 14:41:37 np0000163014 losetup[79733]: /dev/loop0 Apr 02 14:41:37 np0000163014 systemd[1]: Finished stack-volumes-lvmdriver-1-backing-file.service - Activate LVM backing file /opt/stack/data/stack-volumes-lvmdriver-1-backing-file. ● sysfsutils.service - Apply sysfs variables Loaded: loaded (/usr/lib/systemd/system/sysfsutils.service; enabled; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:27:53 UTC; 24min ago Docs: man:sysfs.conf(5) man:systool(1) Main PID: 11493 (code=exited, status=0/SUCCESS) CPU: 21ms Apr 02 14:27:53 np0000163014 systemd[1]: Starting sysfsutils.service - Apply sysfs variables... Apr 02 14:27:53 np0000163014 sysfsutils[11493]: * Setting sysfs variables...... Apr 02 14:27:53 np0000163014 sysfsutils[11493]: ...done. Apr 02 14:27:53 np0000163014 systemd[1]: Finished sysfsutils.service - Apply sysfs variables. ○ systemd-ask-password-console.service - Dispatch Password Requests to Console Loaded: loaded (/usr/lib/systemd/system/systemd-ask-password-console.service; static) Active: inactive (dead) TriggeredBy: ● systemd-ask-password-console.path Docs: man:systemd-ask-password-console.service(8) ○ systemd-ask-password-wall.service - Forward Password Requests to Wall Loaded: loaded (/usr/lib/systemd/system/systemd-ask-password-wall.service; static) Active: inactive (dead) TriggeredBy: ● systemd-ask-password-wall.path Docs: man:systemd-ask-password-wall.service(8) ○ systemd-battery-check.service - Check battery level during early boot Loaded: loaded (/usr/lib/systemd/system/systemd-battery-check.service; static) Active: inactive (dead) Docs: man:systemd-battery-check.service(8) ● systemd-binfmt.service - Set Up Additional Binary Formats Loaded: loaded (/usr/lib/systemd/system/systemd-binfmt.service; static) Active: active (exited) since Thu 2026-04-02 14:28:40 UTC; 23min ago Docs: man:systemd-binfmt.service(8) man:binfmt.d(5) https://docs.kernel.org/admin-guide/binfmt-misc.html https://www.freedesktop.org/wiki/Software/systemd/APIFileSystems Main PID: 17034 (code=exited, status=0/SUCCESS) CPU: 13ms Apr 02 14:28:40 np0000163014 systemd[1]: Starting systemd-binfmt.service - Set Up Additional Binary Formats... Apr 02 14:28:40 np0000163014 systemd[1]: Finished systemd-binfmt.service - Set Up Additional Binary Formats. ○ systemd-bsod.service - Displays emergency message in full screen. Loaded: loaded (/usr/lib/systemd/system/systemd-bsod.service; static) Active: inactive (dead) Docs: man:systemd-bsod.service(8) ● systemd-firstboot.service - First Boot Wizard Loaded: loaded (/usr/lib/systemd/system/systemd-firstboot.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-firstboot(1) Main PID: 460 (code=exited, status=0/SUCCESS) CPU: 13ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-firstboot.service - First Boot Wizard... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-firstboot.service - First Boot Wizard. ● systemd-fsck-root.service - File System Check on Root Device Loaded: loaded (/usr/lib/systemd/system/systemd-fsck-root.service; enabled-runtime; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-fsck-root.service(8) Main PID: 395 (code=exited, status=0/SUCCESS) CPU: 26ms Apr 02 14:21:04 ubuntu systemd-fsck[399]: cloudimg-rootfs: clean, 35928/1090720 files, 758546/1090400 blocks Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-fsck-root.service - File System Check on Root Device. Notice: journal has been rotated since unit was started, output may be incomplete. ○ systemd-fsckd.service - File System Check Daemon to report status Loaded: loaded (/usr/lib/systemd/system/systemd-fsckd.service; static) Active: inactive (dead) since Thu 2026-04-02 14:21:34 UTC; 30min ago Duration: 30.038s TriggeredBy: ● systemd-fsckd.socket Docs: man:systemd-fsckd.service(8) Main PID: 416 (code=exited, status=0/SUCCESS) CPU: 17ms Apr 02 14:21:04 ubuntu systemd[1]: Started systemd-fsckd.service - File System Check Daemon to report status. Apr 02 14:21:34 np0000163014 systemd[1]: systemd-fsckd.service: Deactivated successfully. ○ systemd-hibernate-resume.service - Resume from hibernation Loaded: loaded (/usr/lib/systemd/system/systemd-hibernate-resume.service; static) Active: inactive (dead) Docs: man:systemd-hibernate-resume.service(8) ○ systemd-hibernate.service - System Hibernate Loaded: loaded (/usr/lib/systemd/system/systemd-hibernate.service; static) Active: inactive (dead) Docs: man:systemd-hibernate.service(8) ○ systemd-hwdb-update.service - Rebuild Hardware Database Loaded: loaded (/usr/lib/systemd/system/systemd-hwdb-update.service; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:hwdb(7) man:systemd-hwdb(8) Apr 02 14:21:04 ubuntu systemd[1]: systemd-hwdb-update.service - Rebuild Hardware Database was skipped because no trigger condition checks were met. ○ systemd-hybrid-sleep.service - System Hybrid Suspend+Hibernate Loaded: loaded (/usr/lib/systemd/system/systemd-hybrid-sleep.service; static) Active: inactive (dead) Docs: man:systemd-hybrid-sleep.service(8) ○ systemd-initctl.service - initctl Compatibility Daemon Loaded: loaded (/usr/lib/systemd/system/systemd-initctl.service; static) Active: inactive (dead) TriggeredBy: ● systemd-initctl.socket Docs: man:systemd-initctl.service(8) ● systemd-journal-catalog-update.service - Rebuild Journal Catalog Loaded: loaded (/usr/lib/systemd/system/systemd-journal-catalog-update.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-journald.service(8) man:journald.conf(5) Main PID: 462 (code=exited, status=0/SUCCESS) CPU: 26ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. ● systemd-journal-flush.service - Flush Journal to Persistent Storage Loaded: loaded (/usr/lib/systemd/system/systemd-journal-flush.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-journald.service(8) man:journald.conf(5) Main PID: 434 (code=exited, status=0/SUCCESS) CPU: 17ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. ● systemd-journald.service - Journal Service Loaded: loaded (/usr/lib/systemd/system/systemd-journald.service; static) Drop-In: /usr/lib/systemd/system/systemd-journald.service.d └─nice.conf Active: active (running) since Thu 2026-04-02 14:28:56 UTC; 22min ago TriggeredBy: ● systemd-journald.socket ● systemd-journald-dev-log.socket ○ systemd-journald-audit.socket Docs: man:systemd-journald.service(8) man:journald.conf(5) Main PID: 18695 (systemd-journal) Status: "Processing requests..." Tasks: 1 (limit: 77075) FD Store: 59 (limit: 4224) Memory: 72.6M (peak: 72.9M) CPU: 10.275s CGroup: /system.slice/systemd-journald.service └─18695 /usr/lib/systemd/systemd-journald Apr 02 14:28:56 np0000163014 systemd-journald[18695]: Collecting audit messages is disabled. Apr 02 14:28:56 np0000163014 systemd-journald[18695]: Journal started Apr 02 14:28:56 np0000163014 systemd-journald[18695]: System Journal (/var/log/journal/d67238fcb5c64a86ad1cd602ecd11e16) is 24.0M, max 4.0G, 3.9G free. ● systemd-logind.service - User Login Management Loaded: loaded (/usr/lib/systemd/system/systemd-logind.service; static) Drop-In: /usr/lib/systemd/system/systemd-logind.service.d └─dbus.conf Active: active (running) since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:sd-login(3) man:systemd-logind.service(8) man:logind.conf(5) man:org.freedesktop.login1(5) Main PID: 704 (systemd-logind) Status: "Processing requests..." Tasks: 1 (limit: 77075) FD Store: 0 (limit: 512) Memory: 1.8M (peak: 2.0M) CPU: 2.003s CGroup: /system.slice/systemd-logind.service └─704 /usr/lib/systemd/systemd-logind Apr 02 14:21:10 np0000163014 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 02 14:21:10 np0000163014 systemd-logind[704]: New seat seat0. Apr 02 14:21:10 np0000163014 systemd-logind[704]: Watching system buttons on /dev/input/event0 (Power Button) Apr 02 14:21:10 np0000163014 systemd-logind[704]: Watching system buttons on /dev/input/event1 (AT Translated Set 2 keyboard) Apr 02 14:21:10 np0000163014 systemd[1]: Started systemd-logind.service - User Login Management. Apr 02 14:21:59 np0000163014 systemd-logind[704]: New session 1 of user zuul. Apr 02 14:22:13 np0000163014 systemd-logind[704]: New session 3 of user zuul. Apr 02 14:22:39 np0000163014 systemd-logind[704]: Session 3 logged out. Waiting for processes to exit. Apr 02 14:22:39 np0000163014 systemd-logind[704]: Removed session 3. ● systemd-machine-id-commit.service - Commit a transient machine-id on disk Loaded: loaded (/usr/lib/systemd/system/systemd-machine-id-commit.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-machine-id-commit.service(8) Main PID: 494 (code=exited, status=0/SUCCESS) CPU: 18ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. ● systemd-machined.service - Virtual Machine and Container Registration Service Loaded: loaded (/usr/lib/systemd/system/systemd-machined.service; static) Active: active (running) since Thu 2026-04-02 14:34:41 UTC; 17min ago Docs: man:systemd-machined.service(8) man:org.freedesktop.machine1(5) Main PID: 42225 (systemd-machine) Status: "Processing requests..." Tasks: 1 (limit: 77075) Memory: 1.1M (peak: 1.8M) CPU: 1.391s CGroup: /system.slice/systemd-machined.service └─42225 /usr/lib/systemd/systemd-machined Apr 02 14:34:41 np0000163014 systemd[1]: Starting systemd-machined.service - Virtual Machine and Container Registration Service... Apr 02 14:34:41 np0000163014 systemd[1]: Started systemd-machined.service - Virtual Machine and Container Registration Service. ● systemd-modules-load.service - Load Kernel Modules Loaded: loaded (/usr/lib/systemd/system/systemd-modules-load.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-modules-load.service(8) man:modules-load.d(5) Main PID: 396 (code=exited, status=0/SUCCESS) CPU: 13ms Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. ● systemd-networkd-wait-online.service - Wait for Network to be Configured Loaded: loaded (/usr/lib/systemd/system/systemd-networkd-wait-online.service; enabled-runtime; preset: enabled) Drop-In: /run/systemd/system/systemd-networkd-wait-online.service.d └─10-netplan.conf Active: active (exited) since Thu 2026-04-02 14:21:07 UTC; 30min ago Docs: man:systemd-networkd-wait-online.service(8) Main PID: 607 (code=exited, status=0/SUCCESS) CPU: 13ms Apr 02 14:21:05 np0000163014 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 02 14:21:07 np0000163014 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. ● systemd-networkd.service - Network Configuration Loaded: loaded (/usr/lib/systemd/system/systemd-networkd.service; enabled-runtime; preset: enabled) Active: active (running) since Thu 2026-04-02 14:21:05 UTC; 30min ago TriggeredBy: ● systemd-networkd.socket Docs: man:systemd-networkd.service(8) man:org.freedesktop.network1(5) Main PID: 598 (systemd-network) Status: "Processing requests..." Tasks: 1 (limit: 77075) FD Store: 0 (limit: 512) Memory: 3.3M (peak: 3.6M) CPU: 186ms CGroup: /system.slice/systemd-networkd.service └─598 /usr/lib/systemd/systemd-networkd Apr 02 14:34:41 np0000163014 systemd-networkd[598]: virbr0: Link UP Apr 02 14:45:50 np0000163014 systemd-networkd[598]: br-ex: Link UP Apr 02 14:45:50 np0000163014 systemd-networkd[598]: br-ex: Gained carrier Apr 02 14:45:51 np0000163014 systemd-networkd[598]: br-ex: Gained IPv6LL Apr 02 14:50:41 np0000163014 systemd-networkd[598]: tapec2ec1a6-67: Link UP Apr 02 14:50:41 np0000163014 systemd-networkd[598]: tapec2ec1a6-67: Gained carrier Apr 02 14:50:42 np0000163014 systemd-networkd[598]: tap6fccf94a-80: Link UP Apr 02 14:50:42 np0000163014 systemd-networkd[598]: tap6fccf94a-80: Gained carrier Apr 02 14:50:43 np0000163014 systemd-networkd[598]: tapec2ec1a6-67: Gained IPv6LL Apr 02 14:50:44 np0000163014 systemd-networkd[598]: tap6fccf94a-80: Gained IPv6LL ○ systemd-pcrmachine.service - TPM2 PCR Machine ID Measurement Loaded: loaded (/usr/lib/systemd/system/systemd-pcrmachine.service; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-pcrmachine.service(8) ○ systemd-pcrphase-initrd.service - TPM2 PCR Barrier (initrd) Loaded: loaded (/usr/lib/systemd/system/systemd-pcrphase-initrd.service; static) Active: inactive (dead) Docs: man:systemd-pcrphase-initrd.service(8) ○ systemd-pcrphase-sysinit.service - TPM2 PCR Barrier (Initialization) Loaded: loaded (/usr/lib/systemd/system/systemd-pcrphase-sysinit.service; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd-pcrphase-sysinit.service(8) Apr 02 14:21:10 np0000163014 systemd[1]: systemd-pcrphase-sysinit.service - TPM2 PCR Barrier (Initialization) was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ○ systemd-pcrphase.service - TPM2 PCR Barrier (User) Loaded: loaded (/usr/lib/systemd/system/systemd-pcrphase.service; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd-pcrphase.service(8) Apr 02 14:21:10 np0000163014 systemd[1]: systemd-pcrphase.service - TPM2 PCR Barrier (User) was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ○ systemd-pstore.service - Platform Persistent Storage Archival Loaded: loaded (/usr/lib/systemd/system/systemd-pstore.service; enabled; preset: enabled) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-pstore(8) Apr 02 14:21:04 ubuntu systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). ● systemd-random-seed.service - Load/Save OS Random Seed Loaded: loaded (/usr/lib/systemd/system/systemd-random-seed.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-random-seed.service(8) man:random(4) Main PID: 435 (code=exited, status=0/SUCCESS) CPU: 17ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. ● systemd-remount-fs.service - Remount Root and Kernel File Systems Loaded: loaded (/usr/lib/systemd/system/systemd-remount-fs.service; enabled-runtime; preset: enabled) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-remount-fs.service(8) https://www.freedesktop.org/wiki/Software/systemd/APIFileSystems Main PID: 418 (code=exited, status=0/SUCCESS) CPU: 22ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. ○ systemd-repart.service - Repartition Root Disk Loaded: loaded (/usr/lib/systemd/system/systemd-repart.service; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-repart.service(8) Apr 02 14:21:04 ubuntu systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. ● systemd-resolved.service - Network Name Resolution Loaded: loaded (/usr/lib/systemd/system/systemd-resolved.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-resolved.service(8) man:org.freedesktop.resolve1(5) https://www.freedesktop.org/wiki/Software/systemd/writing-network-configuration-managers https://www.freedesktop.org/wiki/Software/systemd/writing-resolver-clients Main PID: 464 (systemd-resolve) Status: "Processing requests..." Tasks: 1 (limit: 77075) Memory: 7.0M (peak: 7.6M) CPU: 486ms CGroup: /system.slice/systemd-resolved.service └─464 /usr/lib/systemd/systemd-resolved Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 02 14:21:04 ubuntu systemd-resolved[464]: Positive Trust Anchors: Apr 02 14:21:04 ubuntu systemd-resolved[464]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 02 14:21:04 ubuntu systemd-resolved[464]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa corp home internal intranet lan local private test Apr 02 14:21:04 ubuntu systemd-resolved[464]: Using system hostname 'ubuntu'. Apr 02 14:21:04 ubuntu systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 02 14:21:05 np0000163014 systemd-resolved[464]: System hostname changed to 'np0000163014'. Apr 02 14:21:39 np0000163014 systemd-resolved[464]: Clock change detected. Flushing caches. ○ systemd-rfkill.service - Load/Save RF Kill Switch Status Loaded: loaded (/usr/lib/systemd/system/systemd-rfkill.service; static) Active: inactive (dead) TriggeredBy: ● systemd-rfkill.socket Docs: man:systemd-rfkill.service(8) ○ systemd-soft-reboot.service - Reboot System Userspace Loaded: loaded (/usr/lib/systemd/system/systemd-soft-reboot.service; static) Active: inactive (dead) Docs: man:systemd-soft-reboot.service(8) ○ systemd-suspend-then-hibernate.service - System Suspend then Hibernate Loaded: loaded (/usr/lib/systemd/system/systemd-suspend-then-hibernate.service; static) Active: inactive (dead) Docs: man:systemd-suspend-then-hibernate.service(8) ○ systemd-suspend.service - System Suspend Loaded: loaded (/usr/lib/systemd/system/systemd-suspend.service; static) Active: inactive (dead) Docs: man:systemd-suspend.service(8) ● systemd-sysctl.service - Apply Kernel Variables Loaded: loaded (/usr/lib/systemd/system/systemd-sysctl.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-sysctl.service(8) man:sysctl.d(5) Main PID: 420 (code=exited, status=0/SUCCESS) CPU: 25ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. ○ systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/ Loaded: loaded (/usr/lib/systemd/system/systemd-sysext.service; disabled; preset: enabled) Active: inactive (dead) Docs: man:systemd-sysext.service(8) ● systemd-sysusers.service - Create System Users Loaded: loaded (/usr/lib/systemd/system/systemd-sysusers.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:sysusers.d(5) man:systemd-sysusers.service(8) Main PID: 438 (code=exited, status=0/SUCCESS) CPU: 26ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-sysusers.service - Create System Users. ● systemd-timesyncd.service - Network Time Synchronization Loaded: loaded (/usr/lib/systemd/system/systemd-timesyncd.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-timesyncd.service(8) Main PID: 465 (systemd-timesyn) Status: "Contacted time server [2620:2d:4000:1::40]:123 (ntp.ubuntu.com)." Tasks: 2 (limit: 77075) Memory: 1.5M (peak: 2.5M) CPU: 82ms CGroup: /system.slice/systemd-timesyncd.service └─465 /usr/lib/systemd/systemd-timesyncd Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Apr 02 14:21:04 ubuntu systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Apr 02 14:21:05 np0000163014 systemd-timesyncd[465]: Network configuration changed, trying to establish connection. Apr 02 14:21:07 np0000163014 systemd-timesyncd[465]: Network configuration changed, trying to establish connection. Apr 02 14:21:07 np0000163014 systemd-timesyncd[465]: Network configuration changed, trying to establish connection. Apr 02 14:21:07 np0000163014 systemd-timesyncd[465]: Network configuration changed, trying to establish connection. Apr 02 14:21:08 np0000163014 systemd-timesyncd[465]: Network configuration changed, trying to establish connection. Apr 02 14:21:39 np0000163014 systemd-timesyncd[465]: Contacted time server [2620:2d:4000:1::40]:123 (ntp.ubuntu.com). Apr 02 14:21:39 np0000163014 systemd-timesyncd[465]: Initial clock synchronization to Thu 2026-04-02 14:21:39.817214 UTC. ○ systemd-tmpfiles-clean.service - Cleanup of Temporary Directories Loaded: loaded (/usr/lib/systemd/system/systemd-tmpfiles-clean.service; static) Active: inactive (dead) since Thu 2026-04-02 14:36:07 UTC; 15min ago TriggeredBy: ● systemd-tmpfiles-clean.timer Docs: man:tmpfiles.d(5) man:systemd-tmpfiles(8) Main PID: 52914 (code=exited, status=0/SUCCESS) CPU: 22ms Apr 02 14:36:07 np0000163014 systemd[1]: Starting systemd-tmpfiles-clean.service - Cleanup of Temporary Directories... Apr 02 14:36:07 np0000163014 systemd-tmpfiles[52914]: /etc/tmpfiles.d/uwsgi.conf:1: Line references path below legacy directory /var/run/, updating /var/run/uwsgi → /run/uwsgi; please update the tmpfiles.d/ drop-in file accordingly. Apr 02 14:36:07 np0000163014 systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully. Apr 02 14:36:07 np0000163014 systemd[1]: Finished systemd-tmpfiles-clean.service - Cleanup of Temporary Directories. ● systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully Loaded: loaded (/usr/lib/systemd/system/systemd-tmpfiles-setup-dev-early.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:tmpfiles.d(5) man:systemd-tmpfiles(8) Main PID: 421 (code=exited, status=0/SUCCESS) CPU: 30ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. ● systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev Loaded: loaded (/usr/lib/systemd/system/systemd-tmpfiles-setup-dev.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:tmpfiles.d(5) man:systemd-tmpfiles(8) Main PID: 442 (code=exited, status=0/SUCCESS) CPU: 23ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. ● systemd-tmpfiles-setup.service - Create Volatile Files and Directories Loaded: loaded (/usr/lib/systemd/system/systemd-tmpfiles-setup.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:tmpfiles.d(5) man:systemd-tmpfiles(8) Main PID: 452 (code=exited, status=0/SUCCESS) CPU: 35ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-tmpfiles-setup.service - Create Volatile Files and Directories... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-tmpfiles-setup.service - Create Volatile Files and Directories. ○ systemd-tpm2-setup-early.service - TPM2 SRK Setup (Early) Loaded: loaded (/usr/lib/systemd/system/systemd-tpm2-setup-early.service; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-tpm2-setup.service(8) ○ systemd-tpm2-setup.service - TPM2 SRK Setup Loaded: loaded (/usr/lib/systemd/system/systemd-tpm2-setup.service; static) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-tpm2-setup.service(8) Apr 02 14:21:04 ubuntu systemd[1]: systemd-tpm2-setup.service - TPM2 SRK Setup was skipped because of an unmet condition check (ConditionSecurity=measured-uki). ● systemd-udev-settle.service - Wait for udev To Complete Device Initialization Loaded: loaded (/usr/lib/systemd/system/systemd-udev-settle.service; static) Active: active (exited) since Thu 2026-04-02 14:41:37 UTC; 10min ago Docs: man:systemd-udev-settle.service(8) Main PID: 79731 (code=exited, status=0/SUCCESS) CPU: 5ms Apr 02 14:41:37 np0000163014 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Apr 02 14:41:37 np0000163014 udevadm[79731]: systemd-udev-settle.service is deprecated. Please fix stack-volumes-lvmdriver-1-backing-file.service not to pull it in. Apr 02 14:41:37 np0000163014 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. ● systemd-udev-trigger.service - Coldplug All udev Devices Loaded: loaded (/usr/lib/systemd/system/systemd-udev-trigger.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:udev(7) man:systemd-udevd.service(8) Main PID: 398 (code=exited, status=0/SUCCESS) CPU: 142ms Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. ● systemd-udevd.service - Rule-based Manager for Device Events and Files Loaded: loaded (/usr/lib/systemd/system/systemd-udevd.service; static) Drop-In: /usr/lib/systemd/system/systemd-udevd.service.d └─syscall-architecture.conf Active: active (running) since Thu 2026-04-02 14:21:04 UTC; 30min ago TriggeredBy: ● systemd-udevd-control.socket ● systemd-udevd-kernel.socket Docs: man:systemd-udevd.service(8) man:udev(7) Main PID: 453 (systemd-udevd) Status: "Processing with 48 children at max" Tasks: 1 Memory: 12.7M (peak: 40.2M) CPU: 2.902s CGroup: /system.slice/systemd-udevd.service └─udev └─453 /usr/lib/systemd/systemd-udevd Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 02 14:21:04 ubuntu systemd-udevd[453]: Using default interface naming scheme 'v255'. Apr 02 14:21:04 ubuntu systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 02 14:41:37 np0000163014 lvm[79784]: PV /dev/loop0 online, VG stack-volumes-lvmdriver-1 is complete. Apr 02 14:41:38 np0000163014 lvm[79824]: PV /dev/loop0 online, VG stack-volumes-lvmdriver-1 is complete. Apr 02 14:41:38 np0000163014 lvm[79824]: VG stack-volumes-lvmdriver-1 finished Apr 02 14:46:29 np0000163014 lvm[113964]: PV /dev/loop0 online, VG stack-volumes-lvmdriver-1 is complete. Apr 02 14:46:29 np0000163014 lvm[113964]: VG stack-volumes-lvmdriver-1 finished ● systemd-update-done.service - Update is Completed Loaded: loaded (/usr/lib/systemd/system/systemd-update-done.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-update-done.service(8) Main PID: 538 (code=exited, status=0/SUCCESS) CPU: 7ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-update-done.service - Update is Completed. ○ systemd-update-utmp-runlevel.service - Record Runlevel Change in UTMP Loaded: loaded (/usr/lib/systemd/system/systemd-update-utmp-runlevel.service; static) Active: inactive (dead) since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd-update-utmp-runlevel.service(8) man:utmp(5) Main PID: 745 (code=exited, status=0/SUCCESS) CPU: 14ms Apr 02 14:21:10 np0000163014 systemd[1]: Starting systemd-update-utmp-runlevel.service - Record Runlevel Change in UTMP... Apr 02 14:21:10 np0000163014 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Apr 02 14:21:10 np0000163014 systemd[1]: Finished systemd-update-utmp-runlevel.service - Record Runlevel Change in UTMP. ● systemd-update-utmp.service - Record System Boot/Shutdown in UTMP Loaded: loaded (/usr/lib/systemd/system/systemd-update-utmp.service; static) Active: active (exited) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-update-utmp.service(8) man:utmp(5) Main PID: 467 (code=exited, status=0/SUCCESS) CPU: 15ms Apr 02 14:21:04 ubuntu systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 02 14:21:04 ubuntu systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. ● systemd-user-sessions.service - Permit User Sessions Loaded: loaded (/usr/lib/systemd/system/systemd-user-sessions.service; static) Active: active (exited) since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd-user-sessions.service(8) Main PID: 706 (code=exited, status=0/SUCCESS) CPU: 17ms Apr 02 14:21:10 np0000163014 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 02 14:21:10 np0000163014 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. ● user-runtime-dir@1000.service - User Runtime Directory /run/user/1000 Loaded: loaded (/usr/lib/systemd/system/user-runtime-dir@.service; static) Active: active (exited) since Thu 2026-04-02 14:21:59 UTC; 29min ago Docs: man:user@.service(5) Main PID: 830 (code=exited, status=0/SUCCESS) CPU: 15ms Apr 02 14:21:59 np0000163014 systemd[1]: Starting user-runtime-dir@1000.service - User Runtime Directory /run/user/1000... Apr 02 14:21:59 np0000163014 systemd[1]: Finished user-runtime-dir@1000.service - User Runtime Directory /run/user/1000. ● user@1000.service - User Manager for UID 1000 Loaded: loaded (/usr/lib/systemd/system/user@.service; static) Drop-In: /usr/lib/systemd/system/user@.service.d └─10-login-barrier.conf, timeout.conf Active: active (running) since Thu 2026-04-02 14:22:00 UTC; 29min ago Docs: man:user@.service(5) Main PID: 833 (systemd) Status: "Ready." Tasks: 2 Memory: 4.5M (peak: 7.6M) CPU: 266ms CGroup: /user.slice/user-1000.slice/user@1000.service └─init.scope ├─833 /usr/lib/systemd/systemd --user └─834 "(sd-pam)" Apr 02 14:22:00 np0000163014 systemd[833]: Listening on keyboxd.socket - GnuPG public key management service. Apr 02 14:22:00 np0000163014 systemd[833]: Listening on pk-debconf-helper.socket - debconf communication socket. Apr 02 14:22:00 np0000163014 systemd[833]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 02 14:22:00 np0000163014 systemd[833]: Listening on gpg-agent-ssh.socket - GnuPG cryptographic agent (ssh-agent emulation). Apr 02 14:22:00 np0000163014 systemd[833]: Reached target sockets.target - Sockets. Apr 02 14:22:00 np0000163014 systemd[833]: Reached target basic.target - Basic System. Apr 02 14:22:00 np0000163014 systemd[833]: Reached target default.target - Main User Target. Apr 02 14:22:00 np0000163014 systemd[833]: Startup finished in 172ms. Apr 02 14:22:00 np0000163014 systemd[1]: Started user@1000.service - User Manager for UID 1000. Apr 02 14:27:44 np0000163014 systemd[833]: launchpadlib-cache-clean.service - Clean up old files in the Launchpadlib cache was skipped because of an unmet condition check (ConditionPathExists=/home/zuul/.launchpadlib/api.launchpad.net/cache). ○ uuidd.service - Daemon for generating UUIDs Loaded: loaded (/usr/lib/systemd/system/uuidd.service; indirect; preset: enabled) Active: inactive (dead) TriggeredBy: ● uuidd.socket Docs: man:uuidd(8) ● uwsgi.service - LSB: Start/stop uWSGI server instance(s) Loaded: loaded (/etc/init.d/uwsgi; generated) Active: active (exited) since Thu 2026-04-02 14:32:23 UTC; 19min ago Docs: man:systemd-sysv-generator(8) CPU: 76ms Apr 02 14:32:22 np0000163014 systemd[1]: Starting uwsgi.service - LSB: Start/stop uWSGI server instance(s)... Apr 02 14:32:22 np0000163014 uwsgi[29808]: * Starting app server(s) uwsgi Apr 02 14:32:23 np0000163014 uwsgi[29808]: ...done. Apr 02 14:32:23 np0000163014 systemd[1]: Started uwsgi.service - LSB: Start/stop uWSGI server instance(s). ● virtlockd.service - libvirt locking daemon Loaded: loaded (/usr/lib/systemd/system/virtlockd.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:34:43 UTC; 17min ago TriggeredBy: ● virtlockd-admin.socket ● virtlockd.socket Docs: man:virtlockd(8) https://libvirt.org/ Main PID: 42437 (virtlockd) Tasks: 1 (limit: 77075) Memory: 2.0M (peak: 2.3M) CPU: 21ms CGroup: /system.slice/virtlockd.service └─42437 /usr/sbin/virtlockd Apr 02 14:34:43 np0000163014 systemd[1]: Starting virtlockd.service - libvirt locking daemon... Apr 02 14:34:43 np0000163014 systemd[1]: Started virtlockd.service - libvirt locking daemon. ● virtlogd.service - libvirt logging daemon Loaded: loaded (/usr/lib/systemd/system/virtlogd.service; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:35:14 UTC; 16min ago TriggeredBy: ● virtlogd.socket ● virtlogd-admin.socket Docs: man:virtlogd(8) https://libvirt.org/ Main PID: 48046 (virtlogd) Tasks: 1 (limit: 77075) Memory: 2.4M (peak: 2.7M) CPU: 59ms CGroup: /system.slice/virtlogd.service └─48046 /usr/sbin/virtlogd Apr 02 14:35:14 np0000163014 systemd[1]: Starting virtlogd.service - libvirt logging daemon... Apr 02 14:35:14 np0000163014 systemd[1]: Started virtlogd.service - libvirt logging daemon. ● -.slice - Root Slice Loaded: loaded Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Docs: man:systemd.special(7) Tasks: 2288 Memory: 12.5G () CPU: 41min 22.540s CGroup: / ├─init.scope │ └─1 /sbin/init nofb ├─system.slice │ ├─apache-htcacheclean.service │ │ └─13911 /usr/bin/htcacheclean -d 120 -p /var/cache/apache2/mod_cache_disk -l 300M -n │ ├─apache2.service │ │ ├─121459 /usr/sbin/apache2 -k start │ │ ├─121463 /usr/sbin/apache2 -k start │ │ └─121465 /usr/sbin/apache2 -k start │ ├─containerd.service │ │ ├─20255 /usr/bin/containerd │ │ └─21112 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8 -address /run/containerd/containerd.sock │ ├─dbus.service │ │ └─699 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only │ ├─dm-event.service │ │ └─113941 /usr/sbin/dmeventd -f │ ├─docker-54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8.scope │ │ ├─init.scope │ │ │ └─21135 /sbin/init │ │ ├─kubelet.slice │ │ │ ├─kubelet-kubepods.slice │ │ │ │ ├─kubelet-kubepods-besteffort.slice │ │ │ │ │ ├─kubelet-kubepods-besteffort-pod0252a4bd_02fc_4e45_b96b_87fd60aae118.slice │ │ │ │ │ │ ├─cri-containerd-38b467f0d9256afa3b38e32675958de195df595af5b33c93ef4ad091fd7ca2ce.scope │ │ │ │ │ │ │ └─24819 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,KubeadmBootstrapFormatIgnition=true,PriorityQueue=false --bootstrap-token-ttl=15m │ │ │ │ │ │ └─cri-containerd-cab203df0fa479277115976d13436c61979f71791765859f21b6ef4d545e0a28.scope │ │ │ │ │ │ └─24432 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-pod0b3e7747_7f3a_4a37_880c_7609beb2bec2.slice │ │ │ │ │ │ ├─cri-containerd-0d1c8a0dc4444ab42de6f1edc1a3d7f9d15e8257c27b683b19c06074509a75fe.scope │ │ │ │ │ │ │ └─25117 /manager --leader-elect --v=2 --diagnostics-address=127.0.0.1:8080 --insecure-diagnostics=true │ │ │ │ │ │ └─cri-containerd-feb0b8b6bb17ac3fdd5946e4a75f306d531c347d9923d5f36691d6d272a7bb3f.scope │ │ │ │ │ │ └─24759 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-pod19708082_a0bb_4e1a_8b9f_8b55bade2a97.slice │ │ │ │ │ │ ├─cri-containerd-267658bf363448825ad8effc3144e2a05b7e99f48f9ef266949239dadf0e647f.scope │ │ │ │ │ │ │ └─24642 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,ClusterResourceSet=true,ClusterTopology=true,RuntimeSDK=false,MachineSetPreflightChecks=true,MachineWaitForVolumeDetachConsiderVolumeAttachments=true,PriorityQueue=false │ │ │ │ │ │ └─cri-containerd-f15e92eac3c699ac4091b22d11aebac4ddda90272e826d9430f8393eb08e6290.scope │ │ │ │ │ │ └─24363 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-pod2b65498f_1697_43c6_a06c_b95cb04050e9.slice │ │ │ │ │ │ ├─cri-containerd-61bac5c5b3476d4d45e83534a006e7e92b17979efcd672630913702862a801da.scope │ │ │ │ │ │ │ └─23224 /pause │ │ │ │ │ │ └─cri-containerd-7f8394ff1f96657dbbd8821483c8bd4d1c2cdd2a378ea35fe8be66375af9edd2.scope │ │ │ │ │ │ └─23605 local-path-provisioner --debug start --helper-image docker.io/kindest/local-path-helper:v20220607-9a4d8d2a --config /etc/config/config.json │ │ │ │ │ ├─kubelet-kubepods-besteffort-pod3c159520_1948_489d_96ad_e4753123a872.slice │ │ │ │ │ │ ├─cri-containerd-c43573aa54912bff9c0d74b6ef47458ab4fbfcd9e25f44f5250dba0ace166133.scope │ │ │ │ │ │ │ └─24161 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=kube-system │ │ │ │ │ │ └─cri-containerd-ccc37416b9aa55b0d540fe8325a804eb567d99c4e67550ef009f367d5d735c4a.scope │ │ │ │ │ │ └─23568 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-podcfd58e14_dd69_4f38_9746_b9c5983f57b6.slice │ │ │ │ │ │ ├─cri-containerd-2c4b62468f8906532289d834d3d20d746566a61df2d31dc105f87c9c243ba045.scope │ │ │ │ │ │ │ └─23866 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --acme-http01-solver-image=quay.io/jetstack/cert-manager-acmesolver:v1.18.1 --max-concurrent-challenges=60 │ │ │ │ │ │ └─cri-containerd-fd50372de23838501cfd1f9593f95c913e7d080f10f61d6628f9f6f454cca9d5.scope │ │ │ │ │ │ └─23400 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-podd87bcd89_9c5c_4cfb_a125_66fbb1d9c47d.slice │ │ │ │ │ │ ├─cri-containerd-88e90d76e68188293a25b8b8c51be0f3ccf15c5f821aa794f1b0cefabe0a94a3.scope │ │ │ │ │ │ │ └─22896 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=kind-control-plane │ │ │ │ │ │ └─cri-containerd-ff188e61a32f5f07cc3bf88092974b3b46868bbc83fdde14a2d301c6228e65e7.scope │ │ │ │ │ │ └─22841 /pause │ │ │ │ │ ├─kubelet-kubepods-besteffort-pode2c03ea4_9bb3_4184_b5e4_4e142c6fb653.slice │ │ │ │ │ │ ├─cri-containerd-0e3f59113c1c6e6512e47da5dbba7302a32112b2f9c352d5e9f73f4ab1595ca9.scope │ │ │ │ │ │ │ └─23560 /pause │ │ │ │ │ │ └─cri-containerd-68ceb73a5cea74ca45f26405a3bd47783a761566d102a7097661496ba37a5647.scope │ │ │ │ │ │ └─24073 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc │ │ │ │ │ └─kubelet-kubepods-besteffort-podf874c790_24e9_4b31_a814_05df2252c78f.slice │ │ │ │ │ ├─cri-containerd-be887502786ee7935d2add369d9825d38b897c67c3b50c15b89b6710ecae3e6c.scope │ │ │ │ │ │ └─24579 /pause │ │ │ │ │ └─cri-containerd-d1af59187bb078e68792bc9bfcee67bd66f3249aee55bbf56c703b117ecff25e.scope │ │ │ │ │ └─24953 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,ClusterTopology=true,KubeadmBootstrapFormatIgnition=true,PriorityQueue=false │ │ │ │ ├─kubelet-kubepods-burstable.slice │ │ │ │ │ ├─kubelet-kubepods-burstable-pod0656ab70da313d6449b17f099a2a3110.slice │ │ │ │ │ │ ├─cri-containerd-2ab87108ea1e6b8b64d509ef7c1c9342d77a7d272ccb9d4240e562e03b234ab1.scope │ │ │ │ │ │ │ └─22131 etcd --advertise-client-urls=https://172.18.0.2:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://172.18.0.2:2380 --initial-cluster=kind-control-plane=https://172.18.0.2:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://172.18.0.2:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://172.18.0.2:2380 --name=kind-control-plane --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ │ │ │ │ │ └─cri-containerd-6208997ab8fc91afafd91db67859c3b91025856a1f308a614267a5a3d0c37366.scope │ │ │ │ │ │ └─21804 /pause │ │ │ │ │ ├─kubelet-kubepods-burstable-pod53ff6c8abd472f64bc9a9afbd3a471a9.slice │ │ │ │ │ │ ├─cri-containerd-870df6abbc20c8d337ecc4c2bce2a50cd533c2789b8639678ed04c6f1d760e4c.scope │ │ │ │ │ │ │ └─21917 /pause │ │ │ │ │ │ └─cri-containerd-e8f0ee1d61dfec222520c595b1deb5719cdeebd6ece0c11d1ae8d92037973a09.scope │ │ │ │ │ │ └─22041 kube-controller-manager --allocate-node-cidrs=true --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=127.0.0.1 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-cidr=10.244.0.0/16 --cluster-name=kind --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key "--controllers=*,bootstrapsigner,tokencleaner" --enable-hostpath-provisioner=true --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/16 --use-service-account-credentials=true │ │ │ │ │ ├─kubelet-kubepods-burstable-pod7afc15c3_ead2_46e7_9a64_97d9b2aeff9e.slice │ │ │ │ │ │ ├─cri-containerd-19e68a60cb0b1f7aaff877e9aa2ce9e215bb133217cb6c6bec51e589245f9657.scope │ │ │ │ │ │ │ └─23978 /manager --metrics-bind-address=:8443 --leader-elect --health-probe-bind-address=:8081 │ │ │ │ │ │ └─cri-containerd-24a89f3f6c95c059079539dd5111752929e7c7739c4f76340c3501a5beba8343.scope │ │ │ │ │ │ └─23468 /pause │ │ │ │ │ ├─kubelet-kubepods-burstable-podaf8e1906_1a5b_4f97_902c_ca7d36e47926.slice │ │ │ │ │ │ ├─cri-containerd-48b1c569323512f743b5d77a02134ea29982e2225d3d1e74fb3f1f47fc7e0b24.scope │ │ │ │ │ │ │ └─23654 /coredns -conf /etc/coredns/Corefile │ │ │ │ │ │ └─cri-containerd-6472d35a6a9b786a04324c9f49d7f145bd3246b78321220d11fae0c5e43970c5.scope │ │ │ │ │ │ └─23217 /pause │ │ │ │ │ ├─kubelet-kubepods-burstable-podbee69ab63b6471d4da666ee970746eae.slice │ │ │ │ │ │ ├─cri-containerd-31d336081f7b38682b523a2347ccf8f412fd51b43947bcb6cc3201680c38d863.scope │ │ │ │ │ │ │ └─21975 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=127.0.0.1 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true │ │ │ │ │ │ └─cri-containerd-7c89215ac2295c29f7bd0d888c165a63c94d11eeb8c6a7292afd33a2f3cdd899.scope │ │ │ │ │ │ └─21907 /pause │ │ │ │ │ ├─kubelet-kubepods-burstable-podd25193d9_141b_4581_a1b0_bb10a32dc523.slice │ │ │ │ │ │ ├─cri-containerd-c21138d45276460b2a29141bd3bb697e08f97fec20a617b33be15c71bc61f54e.scope │ │ │ │ │ │ │ └─23679 /coredns -conf /etc/coredns/Corefile │ │ │ │ │ │ └─cri-containerd-f459ee43bf1a88a56ea2d80d0152986c09363cfb35460b03ee8f3a0f4424f49c.scope │ │ │ │ │ │ └─23475 /pause │ │ │ │ │ └─kubelet-kubepods-burstable-podef6ebc9842be361e05ebdb6790c540b6.slice │ │ │ │ │ ├─cri-containerd-1926238e4fdf69a50346ffe73b66ea67b7f75760bf715fa5fff86efedc94f171.scope │ │ │ │ │ │ └─21844 /pause │ │ │ │ │ └─cri-containerd-3c7a2ea72ddf5756767d9d3c9628ca18ef585748147673c7e3ac210cdce69837.scope │ │ │ │ │ └─22017 kube-apiserver --advertise-address=172.18.0.2 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --runtime-config= --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/16 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ │ │ │ └─kubelet-kubepods-pod30514ae6_c3de_43c7_85f3_bc4f3ac7154d.slice │ │ │ │ ├─cri-containerd-522d017c8e1da3feb0777e65b66d2de23ebc9a86b25a0adc7a5587008b794a64.scope │ │ │ │ │ └─22935 /bin/kindnetd │ │ │ │ └─cri-containerd-d01f22860fc3f4da6c86ca2a85f5326a74793725f56e3e7216be56ae80ef1f44.scope │ │ │ │ └─22849 /pause │ │ │ └─kubelet.service │ │ │ └─22229 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --container-runtime=remote --container-runtime-endpoint=unix:///run/containerd/containerd.sock --node-ip=172.18.0.2 --node-labels= --pod-infra-container-image=registry.k8s.io/pause:3.8 --provider-id=kind://docker/kind/kind-control-plane --fail-swap-on=false --cgroup-root=/kubelet │ │ └─system.slice │ │ ├─containerd.service │ │ │ ├─21331 /usr/local/bin/containerd │ │ │ ├─21782 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6208997ab8fc91afafd91db67859c3b91025856a1f308a614267a5a3d0c37366 -address /run/containerd/containerd.sock │ │ │ ├─21818 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1926238e4fdf69a50346ffe73b66ea67b7f75760bf715fa5fff86efedc94f171 -address /run/containerd/containerd.sock │ │ │ ├─21863 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7c89215ac2295c29f7bd0d888c165a63c94d11eeb8c6a7292afd33a2f3cdd899 -address /run/containerd/containerd.sock │ │ │ ├─21890 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 870df6abbc20c8d337ecc4c2bce2a50cd533c2789b8639678ed04c6f1d760e4c -address /run/containerd/containerd.sock │ │ │ ├─22800 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id ff188e61a32f5f07cc3bf88092974b3b46868bbc83fdde14a2d301c6228e65e7 -address /run/containerd/containerd.sock │ │ │ ├─22808 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id d01f22860fc3f4da6c86ca2a85f5326a74793725f56e3e7216be56ae80ef1f44 -address /run/containerd/containerd.sock │ │ │ ├─23169 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6472d35a6a9b786a04324c9f49d7f145bd3246b78321220d11fae0c5e43970c5 -address /run/containerd/containerd.sock │ │ │ ├─23188 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 61bac5c5b3476d4d45e83534a006e7e92b17979efcd672630913702862a801da -address /run/containerd/containerd.sock │ │ │ ├─23380 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id fd50372de23838501cfd1f9593f95c913e7d080f10f61d6628f9f6f454cca9d5 -address /run/containerd/containerd.sock │ │ │ ├─23427 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 24a89f3f6c95c059079539dd5111752929e7c7739c4f76340c3501a5beba8343 -address /run/containerd/containerd.sock │ │ │ ├─23435 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id f459ee43bf1a88a56ea2d80d0152986c09363cfb35460b03ee8f3a0f4424f49c -address /run/containerd/containerd.sock │ │ │ ├─23506 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0e3f59113c1c6e6512e47da5dbba7302a32112b2f9c352d5e9f73f4ab1595ca9 -address /run/containerd/containerd.sock │ │ │ ├─23523 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id ccc37416b9aa55b0d540fe8325a804eb567d99c4e67550ef009f367d5d735c4a -address /run/containerd/containerd.sock │ │ │ ├─24342 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id f15e92eac3c699ac4091b22d11aebac4ddda90272e826d9430f8393eb08e6290 -address /run/containerd/containerd.sock │ │ │ ├─24412 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id cab203df0fa479277115976d13436c61979f71791765859f21b6ef4d545e0a28 -address /run/containerd/containerd.sock │ │ │ ├─24559 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id be887502786ee7935d2add369d9825d38b897c67c3b50c15b89b6710ecae3e6c -address /run/containerd/containerd.sock │ │ │ └─24739 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id feb0b8b6bb17ac3fdd5946e4a75f306d531c347d9923d5f36691d6d272a7bb3f -address /run/containerd/containerd.sock │ │ └─systemd-journald.service │ │ └─21316 /lib/systemd/systemd-journald │ ├─docker.service │ │ ├─20379 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock │ │ └─21206 /usr/bin/docker-proxy -proto tcp -host-ip 127.0.0.1 -host-port 42841 -container-ip 172.18.0.2 -container-port 6443 -use-listen-fd │ ├─epmd.service │ │ └─25717 /usr/bin/epmd -systemd │ ├─fsidd.service │ │ └─53941 /usr/sbin/fsidd │ ├─haproxy.service │ │ ├─12955 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock │ │ └─12957 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock │ ├─iscsid.service │ │ ├─43375 /usr/sbin/iscsid │ │ └─43376 /usr/sbin/iscsid │ ├─ksmtuned.service │ │ ├─ 4941 /bin/bash /usr/sbin/ksmtuned │ │ └─128547 sleep 60 │ ├─libvirtd.service │ │ ├─ 42323 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/lib/libvirt/libvirt_leaseshelper │ │ ├─ 42324 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/lib/libvirt/libvirt_leaseshelper │ │ └─110870 /usr/sbin/libvirtd --timeout 120 │ ├─memcached.service │ │ └─65584 /usr/bin/memcached -m 64 -p 11211 -u memcache -l 127.0.0.1 -l ::1 -P /var/run/memcached/memcached.pid │ ├─mysql.service │ │ └─62164 /usr/sbin/mysqld │ ├─nfs-blkmap.service │ │ └─53946 /usr/sbin/blkmapd │ ├─nfs-idmapd.service │ │ └─53949 /usr/sbin/rpc.idmapd │ ├─nfs-mountd.service │ │ └─53960 /usr/sbin/rpc.mountd │ ├─nfsdcld.service │ │ └─53961 /usr/sbin/nfsdcld │ ├─nmbd.service │ │ └─54587 /usr/sbin/nmbd --foreground --no-process-group │ ├─ovn-controller-vtep.service │ │ └─99715 ovn-controller-vtep -vconsole:emer -vsyslog:err -vfile:info --vtep-db=/var/run/openvswitch/db.sock --ovnsb-db=/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-controller-vtep.log --pidfile=/var/run/ovn/ovn-controller-vtep.pid --detach │ ├─ovn-controller.service │ │ └─100402 ovn-controller unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --no-chdir --log-file=/var/log/ovn/ovn-controller.log --pidfile=/var/run/ovn/ovn-controller.pid --detach │ ├─ovn-northd.service │ │ └─100136 ovn-northd -vconsole:emer -vsyslog:err -vfile:info --ovnnb-db=unix:/var/run/ovn/ovnnb_db.sock --ovnsb-db=unix:/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-northd.log --pidfile=/var/run/ovn/ovn-northd.pid --detach │ ├─ovn-ovsdb-server-nb.service │ │ └─100056 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-nb.log --remote=punix:/var/run/ovn/ovnnb_db.sock --pidfile=/var/run/ovn/ovnnb_db.pid --unixctl=/var/run/ovn/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=db:OVN_Northbound,SSL,private_key --certificate=db:OVN_Northbound,SSL,certificate --ca-cert=db:OVN_Northbound,SSL,ca_cert --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers /var/lib/ovn/ovnnb_db.db │ ├─ovn-ovsdb-server-sb.service │ │ └─100058 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-sb.log --remote=punix:/var/run/ovn/ovnsb_db.sock --pidfile=/var/run/ovn/ovnsb_db.pid --unixctl=/var/run/ovn/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=db:OVN_Southbound,SSL,private_key --certificate=db:OVN_Southbound,SSL,certificate --ca-cert=db:OVN_Southbound,SSL,ca_cert --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers /var/lib/ovn/ovnsb_db.db │ ├─ovs-vswitchd.service │ │ └─99627 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach │ ├─ovsdb-server.service │ │ └─99578 ovsdb-server /etc/openvswitch/conf.db -vconsole:emer -vsyslog:err -vfile:info --remote=punix:/var/run/openvswitch/db.sock --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --no-chdir --log-file=/var/log/openvswitch/ovsdb-server.log --pidfile=/var/run/openvswitch/ovsdb-server.pid --detach │ ├─polkit.service │ │ └─742 /usr/lib/polkit-1/polkitd --no-debug │ ├─rabbitmq-server.service │ │ ├─25837 /usr/lib/erlang/erts-13.2.2.5/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -- -root /usr/lib/erlang -bindir /usr/lib/erlang/erts-13.2.2.5/bin -progname erl -- -home /var/lib/rabbitmq -- -pa "" -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger "[]" -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -enable-feature maybe_expr │ │ ├─25847 erl_child_setup 65536 │ │ ├─25955 /usr/lib/erlang/erts-13.2.2.5/bin/inet_gethost 4 │ │ ├─25956 /usr/lib/erlang/erts-13.2.2.5/bin/inet_gethost 4 │ │ └─25969 /bin/sh -s rabbit_disk_monitor │ ├─rpc-statd.service │ │ └─53951 /usr/sbin/rpc.statd │ ├─rpcbind.service │ │ └─53314 /sbin/rpcbind -f -w │ ├─rsyslog.service │ │ └─124176 /usr/sbin/rsyslogd -n -iNONE │ ├─smbd.service │ │ ├─54520 /usr/sbin/smbd --foreground --no-process-group │ │ ├─54523 "smbd: notifyd" . │ │ └─54524 "smbd: cleanupd " │ ├─ssh.service │ │ ├─ 744 "sshd: /usr/sbin/sshd -D [listener] 26 of 10-100 startups" │ │ ├─127825 "sshd: root [priv]" │ │ ├─127826 "sshd: root [net]" │ │ ├─127848 "sshd: [accepted]" │ │ ├─127849 "sshd: [net]" │ │ ├─127850 "sshd: [accepted]" │ │ ├─127851 "sshd: [net]" │ │ ├─127852 "sshd: [accepted]" │ │ ├─127853 "sshd: [net]" │ │ ├─127854 "sshd: [accepted]" │ │ ├─127855 "sshd: [net]" │ │ ├─127891 "sshd: [accepted]" │ │ ├─127893 "sshd: [net]" │ │ ├─127897 "sshd: [accepted]" │ │ ├─127898 "sshd: [net]" │ │ ├─127899 "sshd: [accepted]" │ │ ├─127900 "sshd: [net]" │ │ ├─127904 "sshd: [accepted]" │ │ ├─127905 "sshd: [net]" │ │ ├─127928 "sshd: [accepted]" │ │ ├─127988 "sshd: [accepted]" │ │ ├─127999 "sshd: [accepted]" │ │ ├─128090 "sshd: [accepted]" │ │ ├─128134 "sshd: [accepted]" │ │ ├─128135 "sshd: [net]" │ │ ├─128136 "sshd: [accepted]" │ │ ├─128137 "sshd: [accepted]" │ │ ├─128138 "sshd: [net]" │ │ ├─128141 "sshd: [accepted]" │ │ ├─128292 "sshd: [accepted]" │ │ ├─128364 "sshd: [accepted]" │ │ ├─128365 "sshd: [net]" │ │ ├─128604 "sshd: [accepted]" │ │ ├─129705 "sshd: [accepted]" │ │ ├─129861 "sshd: [accepted]" │ │ ├─129864 "sshd: [net]" │ │ ├─129915 "sshd: [accepted]" │ │ ├─130024 "sshd: [accepted]" │ │ ├─130095 "sshd: [accepted]" │ │ ├─130096 "sshd: [net]" │ │ ├─130127 "sshd: [accepted]" │ │ └─130128 "sshd: [net]" │ ├─system-devstack.slice │ │ ├─devstack@barbican-keystone-listener.service │ │ │ ├─117066 "barbican-keystone-listener: master process [/opt/stack/data/venv/bin/barbican-keystone-listener --config-file=/etc/barbican/barbican.conf]" │ │ │ └─117314 "barbican-keystone-listener: ServiceWrapper worker(0)" │ │ ├─devstack@barbican-retry.service │ │ │ ├─116544 "barbican-retry: master process [/opt/stack/data/venv/bin/barbican-retry --config-file=/etc/barbican/barbican.conf]" │ │ │ └─116844 "barbican-retry: ServiceWrapper worker(0)" │ │ ├─devstack@barbican-svc.service │ │ │ ├─115994 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─115995 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─115996 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─115997 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ │ └─115998 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ ├─devstack@c-api.service │ │ │ ├─111318 "cinder-apiuWSGI master" │ │ │ ├─111320 "cinder-apiuWSGI worker 1" │ │ │ ├─111321 "cinder-apiuWSGI worker 2" │ │ │ ├─111322 "cinder-apiuWSGI worker 3" │ │ │ └─111323 "cinder-apiuWSGI worker 4" │ │ ├─devstack@c-bak.service │ │ │ └─112602 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-backup --config-file /etc/cinder/cinder.conf │ │ ├─devstack@c-sch.service │ │ │ └─112032 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-scheduler --config-file /etc/cinder/cinder.conf │ │ ├─devstack@c-vol.service │ │ │ ├─113189 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ │ │ └─113469 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ │ ├─devstack@etcd.service │ │ │ └─63970 /opt/stack/bin/etcd --name np0000163014 --data-dir /opt/stack/data/etcd --initial-cluster-state new --initial-cluster-token etcd-cluster-01 --initial-cluster np0000163014=http://199.204.45.109:2380 --initial-advertise-peer-urls http://199.204.45.109:2380 --advertise-client-urls http://199.204.45.109:2379 --listen-peer-urls http://0.0.0.0:2380 --listen-client-urls http://199.204.45.109:2379 --log-level=debug │ │ ├─devstack@file_tracker.service │ │ │ ├─ 63325 /bin/bash /opt/stack/devstack/tools/file_tracker.sh │ │ │ └─129995 sleep 20 │ │ ├─devstack@g-api.service │ │ │ ├─114027 "glance-apiuWSGI master" │ │ │ ├─114029 "glance-apiuWSGI worker 1" │ │ │ ├─114030 "glance-apiuWSGI worker 2" │ │ │ ├─114031 "glance-apiuWSGI worker 3" │ │ │ └─114032 "glance-apiuWSGI worker 4" │ │ ├─devstack@keystone.service │ │ │ ├─65167 "keystoneuWSGI master" │ │ │ ├─65168 "keystoneuWSGI worker 1" │ │ │ ├─65169 "keystoneuWSGI worker 2" │ │ │ ├─65170 "keystoneuWSGI worker 3" │ │ │ └─65171 "keystoneuWSGI worker 4" │ │ ├─devstack@m-api.service │ │ │ ├─121134 "manila-apiuWSGI master" │ │ │ ├─121135 "manila-apiuWSGI worker 1" │ │ │ ├─121136 "manila-apiuWSGI worker 2" │ │ │ ├─121137 "manila-apiuWSGI worker 3" │ │ │ └─121138 "manila-apiuWSGI worker 4" │ │ ├─devstack@m-dat.service │ │ │ └─127317 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-data --config-file /etc/manila/manila.conf │ │ ├─devstack@m-sch.service │ │ │ └─126741 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-scheduler --config-file /etc/manila/manila.conf │ │ ├─devstack@m-shr.service │ │ │ ├─126208 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-share --config-file /etc/manila/manila.conf │ │ │ └─126626 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-share --config-file /etc/manila/manila.conf │ │ ├─devstack@magnum-api.service │ │ │ ├─118635 "magnum-apiuWSGI master" │ │ │ ├─118636 "magnum-apiuWSGI worker 1" │ │ │ ├─118637 "magnum-apiuWSGI worker 2" │ │ │ ├─118638 "magnum-apiuWSGI worker 3" │ │ │ └─118639 "magnum-apiuWSGI worker 4" │ │ ├─devstack@magnum-cond.service │ │ │ ├─119227 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119467 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119468 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119469 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119470 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119471 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119472 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119473 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119474 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119475 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119476 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119477 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119478 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119479 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119480 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ ├─119481 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ │ └─119482 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─devstack@memory_tracker.service │ │ │ ├─ 62827 /bin/bash /opt/stack/devstack/tools/memory_tracker.sh │ │ │ └─130003 sleep 20 │ │ ├─devstack@n-api-meta.service │ │ │ ├─107210 "nova-api-metauWSGI master" │ │ │ ├─107211 "nova-api-metauWSGI worker 1" │ │ │ ├─107212 "nova-api-metauWSGI worker 2" │ │ │ ├─107213 "nova-api-metauWSGI worker 3" │ │ │ ├─107214 "nova-api-metauWSGI worker 4" │ │ │ └─107215 "nova-api-metauWSGI http 1" │ │ ├─devstack@n-api.service │ │ │ ├─98732 "nova-apiuWSGI master" │ │ │ ├─98733 "nova-apiuWSGI worker 1" │ │ │ ├─98734 "nova-apiuWSGI worker 2" │ │ │ ├─98735 "nova-apiuWSGI worker 3" │ │ │ └─98736 "nova-apiuWSGI worker 4" │ │ ├─devstack@n-cond-cell1.service │ │ │ ├─109228 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova_cell1.conf]" │ │ │ ├─109927 "nova-conductor: ServiceWrapper worker(0)" │ │ │ ├─109935 "nova-conductor: ServiceWrapper worker(1)" │ │ │ ├─109944 "nova-conductor: ServiceWrapper worker(2)" │ │ │ └─109954 "nova-conductor: ServiceWrapper worker(3)" │ │ ├─devstack@n-cpu.service │ │ │ └─110381 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/nova-compute --config-file /etc/nova/nova-cpu.conf │ │ ├─devstack@n-novnc-cell1.service │ │ │ └─107911 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/nova-novncproxy --config-file /etc/nova/nova_cell1.conf --web /opt/stack/novnc │ │ ├─devstack@n-sch.service │ │ │ ├─106605 "nova-scheduler: master process [/opt/stack/data/venv/bin/nova-scheduler --config-file /etc/nova/nova.conf]" │ │ │ ├─107339 "nova-scheduler: ServiceWrapper worker(0)" │ │ │ ├─107345 "nova-scheduler: ServiceWrapper worker(1)" │ │ │ ├─107352 "nova-scheduler: ServiceWrapper worker(2)" │ │ │ └─107362 "nova-scheduler: ServiceWrapper worker(3)" │ │ ├─devstack@n-super-cond.service │ │ │ ├─108690 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova.conf]" │ │ │ ├─109512 "nova-conductor: ServiceWrapper worker(0)" │ │ │ ├─109517 "nova-conductor: ServiceWrapper worker(1)" │ │ │ ├─109523 "nova-conductor: ServiceWrapper worker(2)" │ │ │ └─109534 "nova-conductor: ServiceWrapper worker(3)" │ │ ├─devstack@neutron-api.service │ │ │ ├─102068 "neutron-apiuWSGI master" │ │ │ ├─102069 "neutron-apiuWSGI worker 1" │ │ │ ├─102070 "neutron-apiuWSGI worker 2" │ │ │ ├─102071 "neutron-apiuWSGI worker 3" │ │ │ └─102072 "neutron-apiuWSGI worker 4" │ │ ├─devstack@neutron-ovn-maintenance-worker.service │ │ │ ├─103560 "neutron-ovn-maintenance-worker: master process [/opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ └─104364 "neutron-server: maintenance worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@neutron-periodic-workers.service │ │ │ ├─103064 "neutron-periodic-workers: master process [/opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ ├─103784 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ ├─103796 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ ├─103807 "neutron-server: Periodic worker for \"DbQuotaNoLockDriver\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ └─103819 "neutron-server: Periodic worker for \"L3_NAT_dbonly_mixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@neutron-rpc-server.service │ │ │ ├─102556 "neutron-rpc-server: master process [/opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ │ ├─103752 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ │ └─103760 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─devstack@o-api.service │ │ │ ├─122913 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─122915 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─122916 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ │ ├─122917 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ │ └─122918 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ ├─devstack@o-da.service │ │ │ ├─123448 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/octavia-driver-agent --config-file /etc/octavia/octavia.conf │ │ │ ├─124243 "octavia-driver-agent - status_listener" │ │ │ ├─124245 "octavia-driver-agent - stats_listener" │ │ │ ├─124248 "octavia-driver-agent - get_listener" │ │ │ └─124363 "octavia-driver-agent - provider_agent -- ovn" │ │ ├─devstack@o-hk.service │ │ │ └─124055 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/octavia-housekeeping --config-file /etc/octavia/octavia.conf │ │ ├─devstack@openstack-cli-server.service │ │ │ └─61420 /opt/stack/data/venv/bin/python3 /opt/stack/devstack/files/openstack-cli-server/openstack-cli-server │ │ ├─devstack@placement-api.service │ │ │ ├─104337 "placementuWSGI master" │ │ │ ├─104338 "placementuWSGI worker 1" │ │ │ ├─104339 "placementuWSGI worker 2" │ │ │ ├─104340 "placementuWSGI worker 3" │ │ │ └─104341 "placementuWSGI worker 4" │ │ └─devstack@q-ovn-agent.service │ │ ├─100946 "neutron-ovn-agent: master process [/opt/stack/data/venv/bin/neutron-ovn-agent --config-file /etc/neutron/plugins/ml2/ovn_agent.ini]" │ │ ├─101518 "neutron-ovn-agent: ServiceWrapper worker(0)" │ │ ├─101743 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmpzexbsa_i/privsep.sock │ │ ├─105256 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmp2clb52x1/privsep.sock │ │ ├─127213 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.link_cmd --privsep_sock_path /tmp/tmpypayx3b_/privsep.sock │ │ ├─127750 sudo /opt/stack/data/venv/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf │ │ ├─127751 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf │ │ └─127791 haproxy -f /opt/stack/data/neutron/ovn-metadata-proxy/6fccf94a-88ef-49fd-a0c3-4d8e39aeea0c.conf │ ├─system-getty.slice │ │ └─getty@tty1.service │ │ └─714 /sbin/agetty -o "-p -- \\u" --noclear - linux │ ├─system-serial\x2dgetty.slice │ │ └─serial-getty@ttyS0.service │ │ └─715 /sbin/agetty -o "-p -- \\u" --keep-baud 115200,57600,38400,9600 - vt220 │ ├─systemd-journald.service │ │ └─18695 /usr/lib/systemd/systemd-journald │ ├─systemd-logind.service │ │ └─704 /usr/lib/systemd/systemd-logind │ ├─systemd-machined.service │ │ └─42225 /usr/lib/systemd/systemd-machined │ ├─systemd-networkd.service │ │ └─598 /usr/lib/systemd/systemd-networkd │ ├─systemd-resolved.service │ │ └─464 /usr/lib/systemd/systemd-resolved │ ├─systemd-timesyncd.service │ │ └─465 /usr/lib/systemd/systemd-timesyncd │ ├─systemd-udevd.service │ │ └─udev │ │ └─453 /usr/lib/systemd/systemd-udevd │ ├─virtlockd.service │ │ └─42437 /usr/sbin/virtlockd │ └─virtlogd.service │ └─48046 /usr/sbin/virtlogd └─user.slice └─user-1000.slice ├─session-1.scope │ ├─ 828 "sshd: zuul [priv]" │ ├─ 853 "sshd: zuul@notty" │ ├─ 1054 /usr/bin/python3 │ ├─130107 sh -c "/bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '\"'\"'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3'\"'\"' && sleep 0'" │ ├─130108 /bin/sh -c "sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3' && sleep 0" │ ├─130109 sudo -H -S -n -u root /bin/sh -c "echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3" │ ├─130110 /bin/sh -c "echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3" │ ├─130111 /usr/bin/python3 │ ├─130112 /bin/bash -c "sudo iptables-save > /home/zuul/iptables.txt\n\n# NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from\n# stale NFS mounts.\ntimeout -s 9 60s df -h > /home/zuul/df.txt || true\n# If 'df' times out, the mount output helps debug which NFS share\n# is unresponsive.\nmount > /home/zuul/mount.txt\n\nfor py_ver in 2 3; do\n if [[ \`which python\${py_ver}\` ]]; then\n python\${py_ver} -m pip freeze > /home/zuul/pip\${py_ver}-freeze.txt\n fi\ndone\n\nif [ \`command -v dpkg\` ]; then\n dpkg -l> /home/zuul/dpkg-l.txt\nfi\nif [ \`command -v rpm\` ]; then\n rpm -qa | sort > /home/zuul/rpm-qa.txt\nfi\n\n# Services status\nsudo systemctl status --all > services.txt 2>/dev/null\n\n# NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU\n# failed to start due to denials from SELinux — useful for CentOS\n# and Fedora machines. For Ubuntu (which runs AppArmor), DevStack\n# already captures the contents of /var/log/kern.log (via\n# \`journalctl -t kernel\` redirected into syslog.txt.gz), which\n# contains AppArmor-related messages.\nif [ -f /var/log/audit/audit.log ] ; then\n sudo cp /var/log/audit/audit.log /home/zuul/audit.log &&\n chmod +r /home/zuul/audit.log;\nfi\n\n# gzip and save any coredumps in /var/core\nif [ -d /var/core ]; then\n sudo gzip -r /var/core\n sudo cp -r /var/core /home/zuul/\nfi\n\nsudo ss -lntup | grep ':53' > /home/zuul/listen53.txt\n\n# NOTE(andreaf) Service logs are already in logs/ thanks for the\n# export-devstack-journal log. Apache logs are under apache/ thans to the\n# apache-logs-conf role.\ngrep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}\\.[0-9]{1,3}/ /g' | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}/ /g' | \\\n sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' |\n sed -r 's/\\[.*\\]/ /g' | \\\n sed -r 's/\\s[0-9]+\\s/ /g' | \\\n awk '{if (\$0 in seen) {seen[\$0]++} else {out[++n]=\$0;seen[\$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]\" :: \" out[i] }' > /home/zuul/deprecations.log\n" │ ├─130125 sudo systemctl status --all │ └─130126 systemctl status --all └─user@1000.service └─init.scope ├─833 /usr/lib/systemd/systemd --user └─834 "(sd-pam)" Apr 02 14:50:39 np0000163014 systemd[1]: Reloading... Apr 02 14:50:39 np0000163014 systemd[1]: Reloading finished in 321 ms. Apr 02 14:50:39 np0000163014 systemd[1]: Started devstack@m-sch.service - Devstack devstack@m-sch.service. Apr 02 14:50:41 np0000163014 systemd[1]: Reloading requested from client PID 127127 ('systemctl') (unit session-1.scope)... Apr 02 14:50:41 np0000163014 systemd[1]: Reloading... Apr 02 14:50:41 np0000163014 systemd[1]: Reloading finished in 294 ms. Apr 02 14:50:41 np0000163014 systemd[1]: Reloading requested from client PID 127219 ('systemctl') (unit session-1.scope)... Apr 02 14:50:41 np0000163014 systemd[1]: Reloading... Apr 02 14:50:41 np0000163014 systemd[1]: Reloading finished in 302 ms. Apr 02 14:50:42 np0000163014 systemd[1]: Started devstack@m-dat.service - Devstack devstack@m-dat.service. ● machine.slice - Virtual Machine and Container Slice Loaded: loaded (/usr/lib/systemd/system/machine.slice; static) Active: active since Thu 2026-04-02 14:34:41 UTC; 17min ago Docs: man:systemd.special(7) Tasks: 0 Memory: 0B (peak: 0B) CPU: 0 CGroup: /machine.slice Apr 02 14:34:41 np0000163014 systemd[1]: Created slice machine.slice - Virtual Machine and Container Slice. ● system-devstack.slice - Slice /system/devstack Loaded: loaded Active: active since Thu 2026-04-02 14:38:52 UTC; 13min ago Tasks: 777 Memory: 10.3G (peak: 10.3G) CPU: 6min 9.465s CGroup: /system.slice/system-devstack.slice ├─devstack@barbican-keystone-listener.service │ ├─117066 "barbican-keystone-listener: master process [/opt/stack/data/venv/bin/barbican-keystone-listener --config-file=/etc/barbican/barbican.conf]" │ └─117314 "barbican-keystone-listener: ServiceWrapper worker(0)" ├─devstack@barbican-retry.service │ ├─116544 "barbican-retry: master process [/opt/stack/data/venv/bin/barbican-retry --config-file=/etc/barbican/barbican.conf]" │ └─116844 "barbican-retry: ServiceWrapper worker(0)" ├─devstack@barbican-svc.service │ ├─115994 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ ├─115995 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ ├─115996 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ ├─115997 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ └─115998 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv ├─devstack@c-api.service │ ├─111318 "cinder-apiuWSGI master" │ ├─111320 "cinder-apiuWSGI worker 1" │ ├─111321 "cinder-apiuWSGI worker 2" │ ├─111322 "cinder-apiuWSGI worker 3" │ └─111323 "cinder-apiuWSGI worker 4" ├─devstack@c-bak.service │ └─112602 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-backup --config-file /etc/cinder/cinder.conf ├─devstack@c-sch.service │ └─112032 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-scheduler --config-file /etc/cinder/cinder.conf ├─devstack@c-vol.service │ ├─113189 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ └─113469 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf ├─devstack@etcd.service │ └─63970 /opt/stack/bin/etcd --name np0000163014 --data-dir /opt/stack/data/etcd --initial-cluster-state new --initial-cluster-token etcd-cluster-01 --initial-cluster np0000163014=http://199.204.45.109:2380 --initial-advertise-peer-urls http://199.204.45.109:2380 --advertise-client-urls http://199.204.45.109:2379 --listen-peer-urls http://0.0.0.0:2380 --listen-client-urls http://199.204.45.109:2379 --log-level=debug ├─devstack@file_tracker.service │ ├─ 63325 /bin/bash /opt/stack/devstack/tools/file_tracker.sh │ └─129995 sleep 20 ├─devstack@g-api.service │ ├─114027 "glance-apiuWSGI master" │ ├─114029 "glance-apiuWSGI worker 1" │ ├─114030 "glance-apiuWSGI worker 2" │ ├─114031 "glance-apiuWSGI worker 3" │ └─114032 "glance-apiuWSGI worker 4" ├─devstack@keystone.service │ ├─65167 "keystoneuWSGI master" │ ├─65168 "keystoneuWSGI worker 1" │ ├─65169 "keystoneuWSGI worker 2" │ ├─65170 "keystoneuWSGI worker 3" │ └─65171 "keystoneuWSGI worker 4" ├─devstack@m-api.service │ ├─121134 "manila-apiuWSGI master" │ ├─121135 "manila-apiuWSGI worker 1" │ ├─121136 "manila-apiuWSGI worker 2" │ ├─121137 "manila-apiuWSGI worker 3" │ └─121138 "manila-apiuWSGI worker 4" ├─devstack@m-dat.service │ └─127317 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-data --config-file /etc/manila/manila.conf ├─devstack@m-sch.service │ └─126741 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-scheduler --config-file /etc/manila/manila.conf ├─devstack@m-shr.service │ ├─126208 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-share --config-file /etc/manila/manila.conf │ └─126626 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-share --config-file /etc/manila/manila.conf ├─devstack@magnum-api.service │ ├─118635 "magnum-apiuWSGI master" │ ├─118636 "magnum-apiuWSGI worker 1" │ ├─118637 "magnum-apiuWSGI worker 2" │ ├─118638 "magnum-apiuWSGI worker 3" │ └─118639 "magnum-apiuWSGI worker 4" ├─devstack@magnum-cond.service │ ├─119227 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119467 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119468 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119469 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119470 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119471 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119472 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119473 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119474 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119475 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119476 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119477 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119478 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119479 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119480 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─119481 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ └─119482 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor ├─devstack@memory_tracker.service │ ├─ 62827 /bin/bash /opt/stack/devstack/tools/memory_tracker.sh │ └─130003 sleep 20 ├─devstack@n-api-meta.service │ ├─107210 "nova-api-metauWSGI master" │ ├─107211 "nova-api-metauWSGI worker 1" │ ├─107212 "nova-api-metauWSGI worker 2" │ ├─107213 "nova-api-metauWSGI worker 3" │ ├─107214 "nova-api-metauWSGI worker 4" │ └─107215 "nova-api-metauWSGI http 1" ├─devstack@n-api.service │ ├─98732 "nova-apiuWSGI master" │ ├─98733 "nova-apiuWSGI worker 1" │ ├─98734 "nova-apiuWSGI worker 2" │ ├─98735 "nova-apiuWSGI worker 3" │ └─98736 "nova-apiuWSGI worker 4" ├─devstack@n-cond-cell1.service │ ├─109228 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova_cell1.conf]" │ ├─109927 "nova-conductor: ServiceWrapper worker(0)" │ ├─109935 "nova-conductor: ServiceWrapper worker(1)" │ ├─109944 "nova-conductor: ServiceWrapper worker(2)" │ └─109954 "nova-conductor: ServiceWrapper worker(3)" ├─devstack@n-cpu.service │ └─110381 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/nova-compute --config-file /etc/nova/nova-cpu.conf ├─devstack@n-novnc-cell1.service │ └─107911 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/nova-novncproxy --config-file /etc/nova/nova_cell1.conf --web /opt/stack/novnc ├─devstack@n-sch.service │ ├─106605 "nova-scheduler: master process [/opt/stack/data/venv/bin/nova-scheduler --config-file /etc/nova/nova.conf]" │ ├─107339 "nova-scheduler: ServiceWrapper worker(0)" │ ├─107345 "nova-scheduler: ServiceWrapper worker(1)" │ ├─107352 "nova-scheduler: ServiceWrapper worker(2)" │ └─107362 "nova-scheduler: ServiceWrapper worker(3)" ├─devstack@n-super-cond.service │ ├─108690 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova.conf]" │ ├─109512 "nova-conductor: ServiceWrapper worker(0)" │ ├─109517 "nova-conductor: ServiceWrapper worker(1)" │ ├─109523 "nova-conductor: ServiceWrapper worker(2)" │ └─109534 "nova-conductor: ServiceWrapper worker(3)" ├─devstack@neutron-api.service │ ├─102068 "neutron-apiuWSGI master" │ ├─102069 "neutron-apiuWSGI worker 1" │ ├─102070 "neutron-apiuWSGI worker 2" │ ├─102071 "neutron-apiuWSGI worker 3" │ └─102072 "neutron-apiuWSGI worker 4" ├─devstack@neutron-ovn-maintenance-worker.service │ ├─103560 "neutron-ovn-maintenance-worker: master process [/opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ └─104364 "neutron-server: maintenance worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" ├─devstack@neutron-periodic-workers.service │ ├─103064 "neutron-periodic-workers: master process [/opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ ├─103784 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ ├─103796 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ ├─103807 "neutron-server: Periodic worker for \"DbQuotaNoLockDriver\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ └─103819 "neutron-server: Periodic worker for \"L3_NAT_dbonly_mixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" ├─devstack@neutron-rpc-server.service │ ├─102556 "neutron-rpc-server: master process [/opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ ├─103752 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ └─103760 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" ├─devstack@o-api.service │ ├─122913 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ ├─122915 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ ├─122916 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ ├─122917 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ └─122918 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv ├─devstack@o-da.service │ ├─123448 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/octavia-driver-agent --config-file /etc/octavia/octavia.conf │ ├─124243 "octavia-driver-agent - status_listener" │ ├─124245 "octavia-driver-agent - stats_listener" │ ├─124248 "octavia-driver-agent - get_listener" │ └─124363 "octavia-driver-agent - provider_agent -- ovn" ├─devstack@o-hk.service │ └─124055 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/octavia-housekeeping --config-file /etc/octavia/octavia.conf ├─devstack@openstack-cli-server.service │ └─61420 /opt/stack/data/venv/bin/python3 /opt/stack/devstack/files/openstack-cli-server/openstack-cli-server ├─devstack@placement-api.service │ ├─104337 "placementuWSGI master" │ ├─104338 "placementuWSGI worker 1" │ ├─104339 "placementuWSGI worker 2" │ ├─104340 "placementuWSGI worker 3" │ └─104341 "placementuWSGI worker 4" └─devstack@q-ovn-agent.service ├─100946 "neutron-ovn-agent: master process [/opt/stack/data/venv/bin/neutron-ovn-agent --config-file /etc/neutron/plugins/ml2/ovn_agent.ini]" ├─101518 "neutron-ovn-agent: ServiceWrapper worker(0)" ├─101743 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmpzexbsa_i/privsep.sock ├─105256 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmp2clb52x1/privsep.sock ├─127213 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.link_cmd --privsep_sock_path /tmp/tmpypayx3b_/privsep.sock ├─127750 sudo /opt/stack/data/venv/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf ├─127751 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf └─127791 haproxy -f /opt/stack/data/neutron/ovn-metadata-proxy/6fccf94a-88ef-49fd-a0c3-4d8e39aeea0c.conf Apr 02 14:51:51 np0000163014 manila-scheduler[126741]: DEBUG dbcounter [-] [126741] Writing DB stats manila:SELECT=30,manila:UPDATE=7,manila:DELETE=1 {{(pid=126741) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:52 np0000163014 nova-conductor[109512]: DEBUG oslo.service.backend._threading.loopingcall [None req-22d1cb6a-d38f-45ce-abe0-5fb7d2956fc8 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=109512) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:52 np0000163014 nova-conductor[109927]: DEBUG oslo.service.backend._threading.loopingcall [None req-d09c5126-d665-43ce-8a4a-f61444a0d106 None None] Dynamic interval looping call 'nova.service.Service.periodic_tasks' sleeping for 60.00 seconds {{(pid=109927) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:52 np0000163014 octavia-housekeeping[124055]: DEBUG octavia.cmd.house_keeping [-] Initiating the cleanup of old resources... {{(pid=124055) db_cleanup /opt/stack/octavia/octavia/cmd/house_keeping.py:49}} Apr 02 14:51:53 np0000163014 manila-data[127317]: DEBUG dbcounter [-] [127317] Writing DB stats manila:SELECT=27,manila:UPDATE=6 {{(pid=127317) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} Apr 02 14:51:54 np0000163014 nova-conductor[109935]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.99 seconds {{(pid=109935) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:54 np0000163014 nova-conductor[109927]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.99 seconds {{(pid=109927) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:54 np0000163014 nova-conductor[109954]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 10.00 seconds {{(pid=109954) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:55 np0000163014 nova-conductor[109944]: DEBUG oslo.service.backend._threading.loopingcall [-] Fixed interval looping call 'nova.servicegroup.drivers.db.DbDriver._report_state' sleeping for 9.99 seconds {{(pid=109944) _run_loop /opt/stack/data/venv/lib/python3.12/site-packages/oslo_service/backend/_threading/loopingcall.py:125}} Apr 02 14:51:55 np0000163014 devstack@neutron-api.service[102072]: DEBUG dbcounter [-] [102072] Writing DB stats neutron:UPDATE=1,neutron:SELECT=1 {{(pid=102072) stat_writer /opt/stack/data/venv/lib/python3.12/site-packages/dbcounter.py:115}} ● system-getty.slice - Slice /system/getty Loaded: loaded Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Tasks: 1 Memory: 328.0K (peak: 1.8M) CPU: 17ms CGroup: /system.slice/system-getty.slice └─getty@tty1.service └─714 /sbin/agetty -o "-p -- \\u" --noclear - linux Notice: journal has been rotated since unit was started, output may be incomplete. ● system-modprobe.slice - Slice /system/modprobe Loaded: loaded Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Tasks: 0 Memory: 80.0K (peak: 3.9M) CPU: 73ms CGroup: /system.slice/system-modprobe.slice Notice: journal has been rotated since unit was started, output may be incomplete. ● system-serial\x2dgetty.slice - Slice /system/serial-getty Loaded: loaded Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Tasks: 1 Memory: 280.0K (peak: 1.8M) CPU: 26ms CGroup: /system.slice/system-serial\x2dgetty.slice └─serial-getty@ttyS0.service └─715 /sbin/agetty -o "-p -- \\u" --keep-baud 115200,57600,38400,9600 - vt220 Notice: journal has been rotated since unit was started, output may be incomplete. ● system.slice - System Slice Loaded: loaded Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Docs: man:systemd.special(7) Tasks: 1991 Memory: 16.4G (peak: 16.4G) CPU: 18min 11.081s CGroup: /system.slice ├─apache-htcacheclean.service │ └─13911 /usr/bin/htcacheclean -d 120 -p /var/cache/apache2/mod_cache_disk -l 300M -n ├─apache2.service │ ├─121459 /usr/sbin/apache2 -k start │ ├─121463 /usr/sbin/apache2 -k start │ └─121465 /usr/sbin/apache2 -k start ├─containerd.service │ ├─20255 /usr/bin/containerd │ └─21112 /usr/bin/containerd-shim-runc-v2 -namespace moby -id 54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8 -address /run/containerd/containerd.sock ├─dbus.service │ └─699 @dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only ├─dm-event.service │ └─113941 /usr/sbin/dmeventd -f ├─docker-54a8734f02f9d8cb158fac8ea74ed4fcdf61ba683190e3da178dfef6f63e1ee8.scope │ ├─init.scope │ │ └─21135 /sbin/init │ ├─kubelet.slice │ │ ├─kubelet-kubepods.slice │ │ │ ├─kubelet-kubepods-besteffort.slice │ │ │ │ ├─kubelet-kubepods-besteffort-pod0252a4bd_02fc_4e45_b96b_87fd60aae118.slice │ │ │ │ │ ├─cri-containerd-38b467f0d9256afa3b38e32675958de195df595af5b33c93ef4ad091fd7ca2ce.scope │ │ │ │ │ │ └─24819 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,KubeadmBootstrapFormatIgnition=true,PriorityQueue=false --bootstrap-token-ttl=15m │ │ │ │ │ └─cri-containerd-cab203df0fa479277115976d13436c61979f71791765859f21b6ef4d545e0a28.scope │ │ │ │ │ └─24432 /pause │ │ │ │ ├─kubelet-kubepods-besteffort-pod0b3e7747_7f3a_4a37_880c_7609beb2bec2.slice │ │ │ │ │ ├─cri-containerd-0d1c8a0dc4444ab42de6f1edc1a3d7f9d15e8257c27b683b19c06074509a75fe.scope │ │ │ │ │ │ └─25117 /manager --leader-elect --v=2 --diagnostics-address=127.0.0.1:8080 --insecure-diagnostics=true │ │ │ │ │ └─cri-containerd-feb0b8b6bb17ac3fdd5946e4a75f306d531c347d9923d5f36691d6d272a7bb3f.scope │ │ │ │ │ └─24759 /pause │ │ │ │ ├─kubelet-kubepods-besteffort-pod19708082_a0bb_4e1a_8b9f_8b55bade2a97.slice │ │ │ │ │ ├─cri-containerd-267658bf363448825ad8effc3144e2a05b7e99f48f9ef266949239dadf0e647f.scope │ │ │ │ │ │ └─24642 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,ClusterResourceSet=true,ClusterTopology=true,RuntimeSDK=false,MachineSetPreflightChecks=true,MachineWaitForVolumeDetachConsiderVolumeAttachments=true,PriorityQueue=false │ │ │ │ │ └─cri-containerd-f15e92eac3c699ac4091b22d11aebac4ddda90272e826d9430f8393eb08e6290.scope │ │ │ │ │ └─24363 /pause │ │ │ │ ├─kubelet-kubepods-besteffort-pod2b65498f_1697_43c6_a06c_b95cb04050e9.slice │ │ │ │ │ ├─cri-containerd-61bac5c5b3476d4d45e83534a006e7e92b17979efcd672630913702862a801da.scope │ │ │ │ │ │ └─23224 /pause │ │ │ │ │ └─cri-containerd-7f8394ff1f96657dbbd8821483c8bd4d1c2cdd2a378ea35fe8be66375af9edd2.scope │ │ │ │ │ └─23605 local-path-provisioner --debug start --helper-image docker.io/kindest/local-path-helper:v20220607-9a4d8d2a --config /etc/config/config.json │ │ │ │ ├─kubelet-kubepods-besteffort-pod3c159520_1948_489d_96ad_e4753123a872.slice │ │ │ │ │ ├─cri-containerd-c43573aa54912bff9c0d74b6ef47458ab4fbfcd9e25f44f5250dba0ace166133.scope │ │ │ │ │ │ └─24161 /app/cmd/cainjector/cainjector --v=2 --leader-election-namespace=kube-system │ │ │ │ │ └─cri-containerd-ccc37416b9aa55b0d540fe8325a804eb567d99c4e67550ef009f367d5d735c4a.scope │ │ │ │ │ └─23568 /pause │ │ │ │ ├─kubelet-kubepods-besteffort-podcfd58e14_dd69_4f38_9746_b9c5983f57b6.slice │ │ │ │ │ ├─cri-containerd-2c4b62468f8906532289d834d3d20d746566a61df2d31dc105f87c9c243ba045.scope │ │ │ │ │ │ └─23866 /app/cmd/controller/controller --v=2 --cluster-resource-namespace=cert-manager --leader-election-namespace=kube-system --acme-http01-solver-image=quay.io/jetstack/cert-manager-acmesolver:v1.18.1 --max-concurrent-challenges=60 │ │ │ │ │ └─cri-containerd-fd50372de23838501cfd1f9593f95c913e7d080f10f61d6628f9f6f454cca9d5.scope │ │ │ │ │ └─23400 /pause │ │ │ │ ├─kubelet-kubepods-besteffort-podd87bcd89_9c5c_4cfb_a125_66fbb1d9c47d.slice │ │ │ │ │ ├─cri-containerd-88e90d76e68188293a25b8b8c51be0f3ccf15c5f821aa794f1b0cefabe0a94a3.scope │ │ │ │ │ │ └─22896 /usr/local/bin/kube-proxy --config=/var/lib/kube-proxy/config.conf --hostname-override=kind-control-plane │ │ │ │ │ └─cri-containerd-ff188e61a32f5f07cc3bf88092974b3b46868bbc83fdde14a2d301c6228e65e7.scope │ │ │ │ │ └─22841 /pause │ │ │ │ ├─kubelet-kubepods-besteffort-pode2c03ea4_9bb3_4184_b5e4_4e142c6fb653.slice │ │ │ │ │ ├─cri-containerd-0e3f59113c1c6e6512e47da5dbba7302a32112b2f9c352d5e9f73f4ab1595ca9.scope │ │ │ │ │ │ └─23560 /pause │ │ │ │ │ └─cri-containerd-68ceb73a5cea74ca45f26405a3bd47783a761566d102a7097661496ba37a5647.scope │ │ │ │ │ └─24073 /app/cmd/webhook/webhook --v=2 --secure-port=10250 --dynamic-serving-ca-secret-namespace=cert-manager --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.cert-manager --dynamic-serving-dns-names=cert-manager-webhook.cert-manager.svc │ │ │ │ └─kubelet-kubepods-besteffort-podf874c790_24e9_4b31_a814_05df2252c78f.slice │ │ │ │ ├─cri-containerd-be887502786ee7935d2add369d9825d38b897c67c3b50c15b89b6710ecae3e6c.scope │ │ │ │ │ └─24579 /pause │ │ │ │ └─cri-containerd-d1af59187bb078e68792bc9bfcee67bd66f3249aee55bbf56c703b117ecff25e.scope │ │ │ │ └─24953 /manager --leader-elect --diagnostics-address=:8443 --insecure-diagnostics=false --feature-gates=MachinePool=true,ClusterTopology=true,KubeadmBootstrapFormatIgnition=true,PriorityQueue=false │ │ │ ├─kubelet-kubepods-burstable.slice │ │ │ │ ├─kubelet-kubepods-burstable-pod0656ab70da313d6449b17f099a2a3110.slice │ │ │ │ │ ├─cri-containerd-2ab87108ea1e6b8b64d509ef7c1c9342d77a7d272ccb9d4240e562e03b234ab1.scope │ │ │ │ │ │ └─22131 etcd --advertise-client-urls=https://172.18.0.2:2379 --cert-file=/etc/kubernetes/pki/etcd/server.crt --client-cert-auth=true --data-dir=/var/lib/etcd --experimental-initial-corrupt-check=true --experimental-watch-progress-notify-interval=5s --initial-advertise-peer-urls=https://172.18.0.2:2380 --initial-cluster=kind-control-plane=https://172.18.0.2:2380 --key-file=/etc/kubernetes/pki/etcd/server.key --listen-client-urls=https://127.0.0.1:2379,https://172.18.0.2:2379 --listen-metrics-urls=http://127.0.0.1:2381 --listen-peer-urls=https://172.18.0.2:2380 --name=kind-control-plane --peer-cert-file=/etc/kubernetes/pki/etcd/peer.crt --peer-client-cert-auth=true --peer-key-file=/etc/kubernetes/pki/etcd/peer.key --peer-trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt --snapshot-count=10000 --trusted-ca-file=/etc/kubernetes/pki/etcd/ca.crt │ │ │ │ │ └─cri-containerd-6208997ab8fc91afafd91db67859c3b91025856a1f308a614267a5a3d0c37366.scope │ │ │ │ │ └─21804 /pause │ │ │ │ ├─kubelet-kubepods-burstable-pod53ff6c8abd472f64bc9a9afbd3a471a9.slice │ │ │ │ │ ├─cri-containerd-870df6abbc20c8d337ecc4c2bce2a50cd533c2789b8639678ed04c6f1d760e4c.scope │ │ │ │ │ │ └─21917 /pause │ │ │ │ │ └─cri-containerd-e8f0ee1d61dfec222520c595b1deb5719cdeebd6ece0c11d1ae8d92037973a09.scope │ │ │ │ │ └─22041 kube-controller-manager --allocate-node-cidrs=true --authentication-kubeconfig=/etc/kubernetes/controller-manager.conf --authorization-kubeconfig=/etc/kubernetes/controller-manager.conf --bind-address=127.0.0.1 --client-ca-file=/etc/kubernetes/pki/ca.crt --cluster-cidr=10.244.0.0/16 --cluster-name=kind --cluster-signing-cert-file=/etc/kubernetes/pki/ca.crt --cluster-signing-key-file=/etc/kubernetes/pki/ca.key "--controllers=*,bootstrapsigner,tokencleaner" --enable-hostpath-provisioner=true --kubeconfig=/etc/kubernetes/controller-manager.conf --leader-elect=true --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --root-ca-file=/etc/kubernetes/pki/ca.crt --service-account-private-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/16 --use-service-account-credentials=true │ │ │ │ ├─kubelet-kubepods-burstable-pod7afc15c3_ead2_46e7_9a64_97d9b2aeff9e.slice │ │ │ │ │ ├─cri-containerd-19e68a60cb0b1f7aaff877e9aa2ce9e215bb133217cb6c6bec51e589245f9657.scope │ │ │ │ │ │ └─23978 /manager --metrics-bind-address=:8443 --leader-elect --health-probe-bind-address=:8081 │ │ │ │ │ └─cri-containerd-24a89f3f6c95c059079539dd5111752929e7c7739c4f76340c3501a5beba8343.scope │ │ │ │ │ └─23468 /pause │ │ │ │ ├─kubelet-kubepods-burstable-podaf8e1906_1a5b_4f97_902c_ca7d36e47926.slice │ │ │ │ │ ├─cri-containerd-48b1c569323512f743b5d77a02134ea29982e2225d3d1e74fb3f1f47fc7e0b24.scope │ │ │ │ │ │ └─23654 /coredns -conf /etc/coredns/Corefile │ │ │ │ │ └─cri-containerd-6472d35a6a9b786a04324c9f49d7f145bd3246b78321220d11fae0c5e43970c5.scope │ │ │ │ │ └─23217 /pause │ │ │ │ ├─kubelet-kubepods-burstable-podbee69ab63b6471d4da666ee970746eae.slice │ │ │ │ │ ├─cri-containerd-31d336081f7b38682b523a2347ccf8f412fd51b43947bcb6cc3201680c38d863.scope │ │ │ │ │ │ └─21975 kube-scheduler --authentication-kubeconfig=/etc/kubernetes/scheduler.conf --authorization-kubeconfig=/etc/kubernetes/scheduler.conf --bind-address=127.0.0.1 --kubeconfig=/etc/kubernetes/scheduler.conf --leader-elect=true │ │ │ │ │ └─cri-containerd-7c89215ac2295c29f7bd0d888c165a63c94d11eeb8c6a7292afd33a2f3cdd899.scope │ │ │ │ │ └─21907 /pause │ │ │ │ ├─kubelet-kubepods-burstable-podd25193d9_141b_4581_a1b0_bb10a32dc523.slice │ │ │ │ │ ├─cri-containerd-c21138d45276460b2a29141bd3bb697e08f97fec20a617b33be15c71bc61f54e.scope │ │ │ │ │ │ └─23679 /coredns -conf /etc/coredns/Corefile │ │ │ │ │ └─cri-containerd-f459ee43bf1a88a56ea2d80d0152986c09363cfb35460b03ee8f3a0f4424f49c.scope │ │ │ │ │ └─23475 /pause │ │ │ │ └─kubelet-kubepods-burstable-podef6ebc9842be361e05ebdb6790c540b6.slice │ │ │ │ ├─cri-containerd-1926238e4fdf69a50346ffe73b66ea67b7f75760bf715fa5fff86efedc94f171.scope │ │ │ │ │ └─21844 /pause │ │ │ │ └─cri-containerd-3c7a2ea72ddf5756767d9d3c9628ca18ef585748147673c7e3ac210cdce69837.scope │ │ │ │ └─22017 kube-apiserver --advertise-address=172.18.0.2 --allow-privileged=true --authorization-mode=Node,RBAC --client-ca-file=/etc/kubernetes/pki/ca.crt --enable-admission-plugins=NodeRestriction --enable-bootstrap-token-auth=true --etcd-cafile=/etc/kubernetes/pki/etcd/ca.crt --etcd-certfile=/etc/kubernetes/pki/apiserver-etcd-client.crt --etcd-keyfile=/etc/kubernetes/pki/apiserver-etcd-client.key --etcd-servers=https://127.0.0.1:2379 --kubelet-client-certificate=/etc/kubernetes/pki/apiserver-kubelet-client.crt --kubelet-client-key=/etc/kubernetes/pki/apiserver-kubelet-client.key --kubelet-preferred-address-types=InternalIP,ExternalIP,Hostname --proxy-client-cert-file=/etc/kubernetes/pki/front-proxy-client.crt --proxy-client-key-file=/etc/kubernetes/pki/front-proxy-client.key --requestheader-allowed-names=front-proxy-client --requestheader-client-ca-file=/etc/kubernetes/pki/front-proxy-ca.crt --requestheader-extra-headers-prefix=X-Remote-Extra- --requestheader-group-headers=X-Remote-Group --requestheader-username-headers=X-Remote-User --runtime-config= --secure-port=6443 --service-account-issuer=https://kubernetes.default.svc.cluster.local --service-account-key-file=/etc/kubernetes/pki/sa.pub --service-account-signing-key-file=/etc/kubernetes/pki/sa.key --service-cluster-ip-range=10.96.0.0/16 --tls-cert-file=/etc/kubernetes/pki/apiserver.crt --tls-private-key-file=/etc/kubernetes/pki/apiserver.key │ │ │ └─kubelet-kubepods-pod30514ae6_c3de_43c7_85f3_bc4f3ac7154d.slice │ │ │ ├─cri-containerd-522d017c8e1da3feb0777e65b66d2de23ebc9a86b25a0adc7a5587008b794a64.scope │ │ │ │ └─22935 /bin/kindnetd │ │ │ └─cri-containerd-d01f22860fc3f4da6c86ca2a85f5326a74793725f56e3e7216be56ae80ef1f44.scope │ │ │ └─22849 /pause │ │ └─kubelet.service │ │ └─22229 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf --config=/var/lib/kubelet/config.yaml --container-runtime=remote --container-runtime-endpoint=unix:///run/containerd/containerd.sock --node-ip=172.18.0.2 --node-labels= --pod-infra-container-image=registry.k8s.io/pause:3.8 --provider-id=kind://docker/kind/kind-control-plane --fail-swap-on=false --cgroup-root=/kubelet │ └─system.slice │ ├─containerd.service │ │ ├─21331 /usr/local/bin/containerd │ │ ├─21782 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6208997ab8fc91afafd91db67859c3b91025856a1f308a614267a5a3d0c37366 -address /run/containerd/containerd.sock │ │ ├─21818 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 1926238e4fdf69a50346ffe73b66ea67b7f75760bf715fa5fff86efedc94f171 -address /run/containerd/containerd.sock │ │ ├─21863 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 7c89215ac2295c29f7bd0d888c165a63c94d11eeb8c6a7292afd33a2f3cdd899 -address /run/containerd/containerd.sock │ │ ├─21890 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 870df6abbc20c8d337ecc4c2bce2a50cd533c2789b8639678ed04c6f1d760e4c -address /run/containerd/containerd.sock │ │ ├─22800 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id ff188e61a32f5f07cc3bf88092974b3b46868bbc83fdde14a2d301c6228e65e7 -address /run/containerd/containerd.sock │ │ ├─22808 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id d01f22860fc3f4da6c86ca2a85f5326a74793725f56e3e7216be56ae80ef1f44 -address /run/containerd/containerd.sock │ │ ├─23169 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 6472d35a6a9b786a04324c9f49d7f145bd3246b78321220d11fae0c5e43970c5 -address /run/containerd/containerd.sock │ │ ├─23188 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 61bac5c5b3476d4d45e83534a006e7e92b17979efcd672630913702862a801da -address /run/containerd/containerd.sock │ │ ├─23380 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id fd50372de23838501cfd1f9593f95c913e7d080f10f61d6628f9f6f454cca9d5 -address /run/containerd/containerd.sock │ │ ├─23427 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 24a89f3f6c95c059079539dd5111752929e7c7739c4f76340c3501a5beba8343 -address /run/containerd/containerd.sock │ │ ├─23435 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id f459ee43bf1a88a56ea2d80d0152986c09363cfb35460b03ee8f3a0f4424f49c -address /run/containerd/containerd.sock │ │ ├─23506 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id 0e3f59113c1c6e6512e47da5dbba7302a32112b2f9c352d5e9f73f4ab1595ca9 -address /run/containerd/containerd.sock │ │ ├─23523 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id ccc37416b9aa55b0d540fe8325a804eb567d99c4e67550ef009f367d5d735c4a -address /run/containerd/containerd.sock │ │ ├─24342 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id f15e92eac3c699ac4091b22d11aebac4ddda90272e826d9430f8393eb08e6290 -address /run/containerd/containerd.sock │ │ ├─24412 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id cab203df0fa479277115976d13436c61979f71791765859f21b6ef4d545e0a28 -address /run/containerd/containerd.sock │ │ ├─24559 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id be887502786ee7935d2add369d9825d38b897c67c3b50c15b89b6710ecae3e6c -address /run/containerd/containerd.sock │ │ └─24739 /usr/local/bin/containerd-shim-runc-v2 -namespace k8s.io -id feb0b8b6bb17ac3fdd5946e4a75f306d531c347d9923d5f36691d6d272a7bb3f -address /run/containerd/containerd.sock │ └─systemd-journald.service │ └─21316 /lib/systemd/systemd-journald ├─docker.service │ ├─20379 /usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock │ └─21206 /usr/bin/docker-proxy -proto tcp -host-ip 127.0.0.1 -host-port 42841 -container-ip 172.18.0.2 -container-port 6443 -use-listen-fd ├─epmd.service │ └─25717 /usr/bin/epmd -systemd ├─fsidd.service │ └─53941 /usr/sbin/fsidd ├─haproxy.service │ ├─12955 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock │ └─12957 /usr/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /run/haproxy.pid -S /run/haproxy-master.sock ├─iscsid.service │ ├─43375 /usr/sbin/iscsid │ └─43376 /usr/sbin/iscsid ├─ksmtuned.service │ ├─ 4941 /bin/bash /usr/sbin/ksmtuned │ └─128547 sleep 60 ├─libvirtd.service │ ├─ 42323 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/lib/libvirt/libvirt_leaseshelper │ ├─ 42324 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/lib/libvirt/libvirt_leaseshelper │ └─110870 /usr/sbin/libvirtd --timeout 120 ├─memcached.service │ └─65584 /usr/bin/memcached -m 64 -p 11211 -u memcache -l 127.0.0.1 -l ::1 -P /var/run/memcached/memcached.pid ├─mysql.service │ └─62164 /usr/sbin/mysqld ├─nfs-blkmap.service │ └─53946 /usr/sbin/blkmapd ├─nfs-idmapd.service │ └─53949 /usr/sbin/rpc.idmapd ├─nfs-mountd.service │ └─53960 /usr/sbin/rpc.mountd ├─nfsdcld.service │ └─53961 /usr/sbin/nfsdcld ├─nmbd.service │ └─54587 /usr/sbin/nmbd --foreground --no-process-group ├─ovn-controller-vtep.service │ └─99715 ovn-controller-vtep -vconsole:emer -vsyslog:err -vfile:info --vtep-db=/var/run/openvswitch/db.sock --ovnsb-db=/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-controller-vtep.log --pidfile=/var/run/ovn/ovn-controller-vtep.pid --detach ├─ovn-controller.service │ └─100402 ovn-controller unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --no-chdir --log-file=/var/log/ovn/ovn-controller.log --pidfile=/var/run/ovn/ovn-controller.pid --detach ├─ovn-northd.service │ └─100136 ovn-northd -vconsole:emer -vsyslog:err -vfile:info --ovnnb-db=unix:/var/run/ovn/ovnnb_db.sock --ovnsb-db=unix:/var/run/ovn/ovnsb_db.sock --no-chdir --log-file=/var/log/ovn/ovn-northd.log --pidfile=/var/run/ovn/ovn-northd.pid --detach ├─ovn-ovsdb-server-nb.service │ └─100056 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-nb.log --remote=punix:/var/run/ovn/ovnnb_db.sock --pidfile=/var/run/ovn/ovnnb_db.pid --unixctl=/var/run/ovn/ovnnb_db.ctl --remote=db:OVN_Northbound,NB_Global,connections --private-key=db:OVN_Northbound,SSL,private_key --certificate=db:OVN_Northbound,SSL,certificate --ca-cert=db:OVN_Northbound,SSL,ca_cert --ssl-protocols=db:OVN_Northbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Northbound,SSL,ssl_ciphers /var/lib/ovn/ovnnb_db.db ├─ovn-ovsdb-server-sb.service │ └─100058 ovsdb-server -vconsole:off -vfile:info --log-file=/var/log/ovn/ovsdb-server-sb.log --remote=punix:/var/run/ovn/ovnsb_db.sock --pidfile=/var/run/ovn/ovnsb_db.pid --unixctl=/var/run/ovn/ovnsb_db.ctl --remote=db:OVN_Southbound,SB_Global,connections --private-key=db:OVN_Southbound,SSL,private_key --certificate=db:OVN_Southbound,SSL,certificate --ca-cert=db:OVN_Southbound,SSL,ca_cert --ssl-protocols=db:OVN_Southbound,SSL,ssl_protocols --ssl-ciphers=db:OVN_Southbound,SSL,ssl_ciphers /var/lib/ovn/ovnsb_db.db ├─ovs-vswitchd.service │ └─99627 ovs-vswitchd unix:/var/run/openvswitch/db.sock -vconsole:emer -vsyslog:err -vfile:info --mlockall --no-chdir --log-file=/var/log/openvswitch/ovs-vswitchd.log --pidfile=/var/run/openvswitch/ovs-vswitchd.pid --detach ├─ovsdb-server.service │ └─99578 ovsdb-server /etc/openvswitch/conf.db -vconsole:emer -vsyslog:err -vfile:info --remote=punix:/var/run/openvswitch/db.sock --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --no-chdir --log-file=/var/log/openvswitch/ovsdb-server.log --pidfile=/var/run/openvswitch/ovsdb-server.pid --detach ├─polkit.service │ └─742 /usr/lib/polkit-1/polkitd --no-debug ├─rabbitmq-server.service │ ├─25837 /usr/lib/erlang/erts-13.2.2.5/bin/beam.smp -W w -MBas ageffcbf -MHas ageffcbf -MBlmbcs 512 -MHlmbcs 512 -MMmcs 30 -pc unicode -P 1048576 -t 5000000 -stbt db -zdbbl 128000 -sbwt none -sbwtdcpu none -sbwtdio none -- -root /usr/lib/erlang -bindir /usr/lib/erlang/erts-13.2.2.5/bin -progname erl -- -home /var/lib/rabbitmq -- -pa "" -noshell -noinput -s rabbit boot -boot start_sasl -syslog logger "[]" -syslog syslog_error_logger false -kernel prevent_overlapping_partitions false -enable-feature maybe_expr │ ├─25847 erl_child_setup 65536 │ ├─25955 /usr/lib/erlang/erts-13.2.2.5/bin/inet_gethost 4 │ ├─25956 /usr/lib/erlang/erts-13.2.2.5/bin/inet_gethost 4 │ └─25969 /bin/sh -s rabbit_disk_monitor ├─rpc-statd.service │ └─53951 /usr/sbin/rpc.statd ├─rpcbind.service │ └─53314 /sbin/rpcbind -f -w ├─rsyslog.service │ └─124176 /usr/sbin/rsyslogd -n -iNONE ├─smbd.service │ ├─54520 /usr/sbin/smbd --foreground --no-process-group │ ├─54523 "smbd: notifyd" . │ └─54524 "smbd: cleanupd " ├─ssh.service │ ├─ 744 "sshd: /usr/sbin/sshd -D [listener] 26 of 10-100 startups" │ ├─127825 "sshd: root [priv]" │ ├─127826 "sshd: root [net]" │ ├─127848 "sshd: [accepted]" │ ├─127849 "sshd: [net]" │ ├─127850 "sshd: [accepted]" │ ├─127851 "sshd: [net]" │ ├─127852 "sshd: [accepted]" │ ├─127853 "sshd: [net]" │ ├─127854 "sshd: [accepted]" │ ├─127855 "sshd: [net]" │ ├─127891 "sshd: [accepted]" │ ├─127893 "sshd: [net]" │ ├─127897 "sshd: [accepted]" │ ├─127898 "sshd: [net]" │ ├─127899 "sshd: [accepted]" │ ├─127900 "sshd: [net]" │ ├─127904 "sshd: [accepted]" │ ├─127905 "sshd: [net]" │ ├─127928 "sshd: [accepted]" │ ├─127988 "sshd: [accepted]" │ ├─127999 "sshd: [accepted]" │ ├─128090 "sshd: [accepted]" │ ├─128134 "sshd: [accepted]" │ ├─128135 "sshd: [net]" │ ├─128136 "sshd: [accepted]" │ ├─128137 "sshd: [accepted]" │ ├─128138 "sshd: [net]" │ ├─128141 "sshd: [accepted]" │ ├─128292 "sshd: [accepted]" │ ├─128364 "sshd: [accepted]" │ ├─128365 "sshd: [net]" │ ├─128604 "sshd: [accepted]" │ ├─129705 "sshd: [accepted]" │ ├─129861 "sshd: [accepted]" │ ├─129864 "sshd: [net]" │ ├─129915 "sshd: [accepted]" │ ├─130024 "sshd: [accepted]" │ ├─130095 "sshd: [accepted]" │ ├─130096 "sshd: [net]" │ ├─130127 "sshd: [accepted]" │ └─130128 "sshd: [net]" ├─system-devstack.slice │ ├─devstack@barbican-keystone-listener.service │ │ ├─117066 "barbican-keystone-listener: master process [/opt/stack/data/venv/bin/barbican-keystone-listener --config-file=/etc/barbican/barbican.conf]" │ │ └─117314 "barbican-keystone-listener: ServiceWrapper worker(0)" │ ├─devstack@barbican-retry.service │ │ ├─116544 "barbican-retry: master process [/opt/stack/data/venv/bin/barbican-retry --config-file=/etc/barbican/barbican.conf]" │ │ └─116844 "barbican-retry: ServiceWrapper worker(0)" │ ├─devstack@barbican-svc.service │ │ ├─115994 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ ├─115995 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ ├─115996 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ ├─115997 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ │ └─115998 /opt/stack/data/venv/bin/uwsgi --ini /etc/barbican/barbican-uwsgi.ini --venv /opt/stack/data/venv │ ├─devstack@c-api.service │ │ ├─111318 "cinder-apiuWSGI master" │ │ ├─111320 "cinder-apiuWSGI worker 1" │ │ ├─111321 "cinder-apiuWSGI worker 2" │ │ ├─111322 "cinder-apiuWSGI worker 3" │ │ └─111323 "cinder-apiuWSGI worker 4" │ ├─devstack@c-bak.service │ │ └─112602 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-backup --config-file /etc/cinder/cinder.conf │ ├─devstack@c-sch.service │ │ └─112032 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-scheduler --config-file /etc/cinder/cinder.conf │ ├─devstack@c-vol.service │ │ ├─113189 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ │ └─113469 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/cinder-volume --config-file /etc/cinder/cinder.conf │ ├─devstack@etcd.service │ │ └─63970 /opt/stack/bin/etcd --name np0000163014 --data-dir /opt/stack/data/etcd --initial-cluster-state new --initial-cluster-token etcd-cluster-01 --initial-cluster np0000163014=http://199.204.45.109:2380 --initial-advertise-peer-urls http://199.204.45.109:2380 --advertise-client-urls http://199.204.45.109:2379 --listen-peer-urls http://0.0.0.0:2380 --listen-client-urls http://199.204.45.109:2379 --log-level=debug │ ├─devstack@file_tracker.service │ │ ├─ 63325 /bin/bash /opt/stack/devstack/tools/file_tracker.sh │ │ └─129995 sleep 20 │ ├─devstack@g-api.service │ │ ├─114027 "glance-apiuWSGI master" │ │ ├─114029 "glance-apiuWSGI worker 1" │ │ ├─114030 "glance-apiuWSGI worker 2" │ │ ├─114031 "glance-apiuWSGI worker 3" │ │ └─114032 "glance-apiuWSGI worker 4" │ ├─devstack@keystone.service │ │ ├─65167 "keystoneuWSGI master" │ │ ├─65168 "keystoneuWSGI worker 1" │ │ ├─65169 "keystoneuWSGI worker 2" │ │ ├─65170 "keystoneuWSGI worker 3" │ │ └─65171 "keystoneuWSGI worker 4" │ ├─devstack@m-api.service │ │ ├─121134 "manila-apiuWSGI master" │ │ ├─121135 "manila-apiuWSGI worker 1" │ │ ├─121136 "manila-apiuWSGI worker 2" │ │ ├─121137 "manila-apiuWSGI worker 3" │ │ └─121138 "manila-apiuWSGI worker 4" │ ├─devstack@m-dat.service │ │ └─127317 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-data --config-file /etc/manila/manila.conf │ ├─devstack@m-sch.service │ │ └─126741 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-scheduler --config-file /etc/manila/manila.conf │ ├─devstack@m-shr.service │ │ ├─126208 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-share --config-file /etc/manila/manila.conf │ │ └─126626 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/manila-share --config-file /etc/manila/manila.conf │ ├─devstack@magnum-api.service │ │ ├─118635 "magnum-apiuWSGI master" │ │ ├─118636 "magnum-apiuWSGI worker 1" │ │ ├─118637 "magnum-apiuWSGI worker 2" │ │ ├─118638 "magnum-apiuWSGI worker 3" │ │ └─118639 "magnum-apiuWSGI worker 4" │ ├─devstack@magnum-cond.service │ │ ├─119227 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119467 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119468 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119469 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119470 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119471 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119472 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119473 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119474 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119475 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119476 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119477 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119478 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119479 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119480 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ ├─119481 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ │ └─119482 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/magnum-conductor │ ├─devstack@memory_tracker.service │ │ ├─ 62827 /bin/bash /opt/stack/devstack/tools/memory_tracker.sh │ │ └─130003 sleep 20 │ ├─devstack@n-api-meta.service │ │ ├─107210 "nova-api-metauWSGI master" │ │ ├─107211 "nova-api-metauWSGI worker 1" │ │ ├─107212 "nova-api-metauWSGI worker 2" │ │ ├─107213 "nova-api-metauWSGI worker 3" │ │ ├─107214 "nova-api-metauWSGI worker 4" │ │ └─107215 "nova-api-metauWSGI http 1" │ ├─devstack@n-api.service │ │ ├─98732 "nova-apiuWSGI master" │ │ ├─98733 "nova-apiuWSGI worker 1" │ │ ├─98734 "nova-apiuWSGI worker 2" │ │ ├─98735 "nova-apiuWSGI worker 3" │ │ └─98736 "nova-apiuWSGI worker 4" │ ├─devstack@n-cond-cell1.service │ │ ├─109228 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova_cell1.conf]" │ │ ├─109927 "nova-conductor: ServiceWrapper worker(0)" │ │ ├─109935 "nova-conductor: ServiceWrapper worker(1)" │ │ ├─109944 "nova-conductor: ServiceWrapper worker(2)" │ │ └─109954 "nova-conductor: ServiceWrapper worker(3)" │ ├─devstack@n-cpu.service │ │ └─110381 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/nova-compute --config-file /etc/nova/nova-cpu.conf │ ├─devstack@n-novnc-cell1.service │ │ └─107911 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/nova-novncproxy --config-file /etc/nova/nova_cell1.conf --web /opt/stack/novnc │ ├─devstack@n-sch.service │ │ ├─106605 "nova-scheduler: master process [/opt/stack/data/venv/bin/nova-scheduler --config-file /etc/nova/nova.conf]" │ │ ├─107339 "nova-scheduler: ServiceWrapper worker(0)" │ │ ├─107345 "nova-scheduler: ServiceWrapper worker(1)" │ │ ├─107352 "nova-scheduler: ServiceWrapper worker(2)" │ │ └─107362 "nova-scheduler: ServiceWrapper worker(3)" │ ├─devstack@n-super-cond.service │ │ ├─108690 "nova-conductor: master process [/opt/stack/data/venv/bin/nova-conductor --config-file /etc/nova/nova.conf]" │ │ ├─109512 "nova-conductor: ServiceWrapper worker(0)" │ │ ├─109517 "nova-conductor: ServiceWrapper worker(1)" │ │ ├─109523 "nova-conductor: ServiceWrapper worker(2)" │ │ └─109534 "nova-conductor: ServiceWrapper worker(3)" │ ├─devstack@neutron-api.service │ │ ├─102068 "neutron-apiuWSGI master" │ │ ├─102069 "neutron-apiuWSGI worker 1" │ │ ├─102070 "neutron-apiuWSGI worker 2" │ │ ├─102071 "neutron-apiuWSGI worker 3" │ │ └─102072 "neutron-apiuWSGI worker 4" │ ├─devstack@neutron-ovn-maintenance-worker.service │ │ ├─103560 "neutron-ovn-maintenance-worker: master process [/opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ └─104364 "neutron-server: maintenance worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-ovn-maintenance-worker --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ ├─devstack@neutron-periodic-workers.service │ │ ├─103064 "neutron-periodic-workers: master process [/opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ ├─103784 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─103796 "neutron-server: Periodic worker for \"AgentSchedulerDbMixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ ├─103807 "neutron-server: Periodic worker for \"DbQuotaNoLockDriver\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ └─103819 "neutron-server: Periodic worker for \"L3_NAT_dbonly_mixin\" (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-periodic-workers --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ ├─devstack@neutron-rpc-server.service │ │ ├─102556 "neutron-rpc-server: master process [/opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini]" │ │ ├─103752 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ │ └─103760 "neutron-server: rpc worker (/opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rpc-server --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini)" │ ├─devstack@o-api.service │ │ ├─122913 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ ├─122915 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ ├─122916 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ ├─122917 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ │ └─122918 /bin/uwsgi --ini /etc/octavia/octavia-uwsgi.ini --venv /opt/stack/data/venv │ ├─devstack@o-da.service │ │ ├─123448 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/octavia-driver-agent --config-file /etc/octavia/octavia.conf │ │ ├─124243 "octavia-driver-agent - status_listener" │ │ ├─124245 "octavia-driver-agent - stats_listener" │ │ ├─124248 "octavia-driver-agent - get_listener" │ │ └─124363 "octavia-driver-agent - provider_agent -- ovn" │ ├─devstack@o-hk.service │ │ └─124055 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/octavia-housekeeping --config-file /etc/octavia/octavia.conf │ ├─devstack@openstack-cli-server.service │ │ └─61420 /opt/stack/data/venv/bin/python3 /opt/stack/devstack/files/openstack-cli-server/openstack-cli-server │ ├─devstack@placement-api.service │ │ ├─104337 "placementuWSGI master" │ │ ├─104338 "placementuWSGI worker 1" │ │ ├─104339 "placementuWSGI worker 2" │ │ ├─104340 "placementuWSGI worker 3" │ │ └─104341 "placementuWSGI worker 4" │ └─devstack@q-ovn-agent.service │ ├─100946 "neutron-ovn-agent: master process [/opt/stack/data/venv/bin/neutron-ovn-agent --config-file /etc/neutron/plugins/ml2/ovn_agent.ini]" │ ├─101518 "neutron-ovn-agent: ServiceWrapper worker(0)" │ ├─101743 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.namespace_cmd --privsep_sock_path /tmp/tmpzexbsa_i/privsep.sock │ ├─105256 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.default --privsep_sock_path /tmp/tmp2clb52x1/privsep.sock │ ├─127213 /opt/stack/data/venv/bin/python3.12 /usr/local/bin/privsep-helper --config-file /etc/neutron/plugins/ml2/ovn_agent.ini --privsep_context neutron.privileged.link_cmd --privsep_sock_path /tmp/tmpypayx3b_/privsep.sock │ ├─127750 sudo /opt/stack/data/venv/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf │ ├─127751 /opt/stack/data/venv/bin/python3.12 /opt/stack/data/venv/bin/neutron-rootwrap-daemon /etc/neutron/rootwrap.conf │ └─127791 haproxy -f /opt/stack/data/neutron/ovn-metadata-proxy/6fccf94a-88ef-49fd-a0c3-4d8e39aeea0c.conf ├─system-getty.slice │ └─getty@tty1.service │ └─714 /sbin/agetty -o "-p -- \\u" --noclear - linux ├─system-serial\x2dgetty.slice │ └─serial-getty@ttyS0.service │ └─715 /sbin/agetty -o "-p -- \\u" --keep-baud 115200,57600,38400,9600 - vt220 ├─systemd-journald.service │ └─18695 /usr/lib/systemd/systemd-journald ├─systemd-logind.service │ └─704 /usr/lib/systemd/systemd-logind ├─systemd-machined.service │ └─42225 /usr/lib/systemd/systemd-machined ├─systemd-networkd.service │ └─598 /usr/lib/systemd/systemd-networkd ├─systemd-resolved.service │ └─464 /usr/lib/systemd/systemd-resolved ├─systemd-timesyncd.service │ └─465 /usr/lib/systemd/systemd-timesyncd ├─systemd-udevd.service │ └─udev │ └─453 /usr/lib/systemd/systemd-udevd ├─virtlockd.service │ └─42437 /usr/sbin/virtlockd └─virtlogd.service └─48046 /usr/sbin/virtlogd Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05762|poll_loop|DBG|wakeup due to [POLLIN] on fd 31 (199.204.45.109:6641<->199.204.45.109:51512) at ../lib/stream-ssl.c:842 (0% CPU usage) Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05763|stream_ssl|DBG|server27<--ssl:199.204.45.109:51512 type 256 (5 bytes) Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05764|stream_ssl|DBG|server27<--ssl:199.204.45.109:51512 type 257 (1 bytes) Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05765|jsonrpc|DBG|ssl:199.204.45.109:51512: received request, method="echo", params=[], id="echo" Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05766|jsonrpc|DBG|ssl:199.204.45.109:51512: send reply, result=[], id="echo" Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05767|stream_ssl|DBG|server27-->ssl:199.204.45.109:51512 type 256 (5 bytes) Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05768|stream_ssl|DBG|server27-->ssl:199.204.45.109:51512 type 257 (1 bytes) Apr 02 14:51:53 np0000163014 ovsdb-server[100056]: ovs|05769|poll_loop|DBG|wakeup due to 0-ms timeout at ../lib/stream-ssl.c:844 (0% CPU usage) Apr 02 14:51:55 np0000163014 ovsdb-server[100056]: ovs|05770|poll_loop|DBG|wakeup due to 1537-ms timeout at ../ovsdb/ovsdb-server.c:400 (0% CPU usage) Apr 02 14:51:55 np0000163014 ovsdb-server[100058]: ovs|04176|poll_loop|DBG|wakeup due to 2218-ms timeout at ../ovsdb/ovsdb-server.c:400 (0% CPU usage) ● user-1000.slice - User Slice of UID 1000 Loaded: loaded Drop-In: /usr/lib/systemd/system/user-.slice.d └─10-defaults.conf Active: active since Thu 2026-04-02 14:21:59 UTC; 29min ago Docs: man:user@.service(5) Tasks: 15 (limit: 169565) Memory: 23.5G (peak: 23.6G) CPU: 23min 35.721s CGroup: /user.slice/user-1000.slice ├─session-1.scope │ ├─ 828 "sshd: zuul [priv]" │ ├─ 853 "sshd: zuul@notty" │ ├─ 1054 /usr/bin/python3 │ ├─130107 sh -c "/bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '\"'\"'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3'\"'\"' && sleep 0'" │ ├─130108 /bin/sh -c "sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3' && sleep 0" │ ├─130109 sudo -H -S -n -u root /bin/sh -c "echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3" │ ├─130110 /bin/sh -c "echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3" │ ├─130111 /usr/bin/python3 │ ├─130112 /bin/bash -c "sudo iptables-save > /home/zuul/iptables.txt\n\n# NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from\n# stale NFS mounts.\ntimeout -s 9 60s df -h > /home/zuul/df.txt || true\n# If 'df' times out, the mount output helps debug which NFS share\n# is unresponsive.\nmount > /home/zuul/mount.txt\n\nfor py_ver in 2 3; do\n if [[ \`which python\${py_ver}\` ]]; then\n python\${py_ver} -m pip freeze > /home/zuul/pip\${py_ver}-freeze.txt\n fi\ndone\n\nif [ \`command -v dpkg\` ]; then\n dpkg -l> /home/zuul/dpkg-l.txt\nfi\nif [ \`command -v rpm\` ]; then\n rpm -qa | sort > /home/zuul/rpm-qa.txt\nfi\n\n# Services status\nsudo systemctl status --all > services.txt 2>/dev/null\n\n# NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU\n# failed to start due to denials from SELinux — useful for CentOS\n# and Fedora machines. For Ubuntu (which runs AppArmor), DevStack\n# already captures the contents of /var/log/kern.log (via\n# \`journalctl -t kernel\` redirected into syslog.txt.gz), which\n# contains AppArmor-related messages.\nif [ -f /var/log/audit/audit.log ] ; then\n sudo cp /var/log/audit/audit.log /home/zuul/audit.log &&\n chmod +r /home/zuul/audit.log;\nfi\n\n# gzip and save any coredumps in /var/core\nif [ -d /var/core ]; then\n sudo gzip -r /var/core\n sudo cp -r /var/core /home/zuul/\nfi\n\nsudo ss -lntup | grep ':53' > /home/zuul/listen53.txt\n\n# NOTE(andreaf) Service logs are already in logs/ thanks for the\n# export-devstack-journal log. Apache logs are under apache/ thans to the\n# apache-logs-conf role.\ngrep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}\\.[0-9]{1,3}/ /g' | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}/ /g' | \\\n sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' |\n sed -r 's/\\[.*\\]/ /g' | \\\n sed -r 's/\\s[0-9]+\\s/ /g' | \\\n awk '{if (\$0 in seen) {seen[\$0]++} else {out[++n]=\$0;seen[\$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]\" :: \" out[i] }' > /home/zuul/deprecations.log\n" │ ├─130125 sudo systemctl status --all │ └─130126 systemctl status --all └─user@1000.service └─init.scope ├─833 /usr/lib/systemd/systemd --user └─834 "(sd-pam)" Apr 02 14:51:52 np0000163014 python3[130101]: ansible-ansible.legacy.command Invoked with _raw_params=cp -pRL /etc/openstack /home/zuul/etc/ zuul_no_log=False zuul_log_id=0242ac17-0011-42b5-8cef-00000000002f-1-controller zuul_output_max_bytes=1073741824 zuul_ansible_split_streams=False _uses_shell=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Apr 02 14:51:52 np0000163014 sudo[130099]: pam_unix(sudo:session): session closed for user root Apr 02 14:51:53 np0000163014 sudo[130109]: zuul : PWD=/home/zuul ; USER=root ; COMMAND=/bin/sh -c 'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3' Apr 02 14:51:53 np0000163014 sudo[130109]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=1000) Apr 02 14:51:53 np0000163014 python3[130111]: ansible-ansible.legacy.command Invoked with executable=/bin/bash _raw_params=sudo iptables-save > /home/zuul/iptables.txt # NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from # stale NFS mounts. timeout -s 9 60s df -h > /home/zuul/df.txt || true # If 'df' times out, the mount output helps debug which NFS share # is unresponsive. mount > /home/zuul/mount.txt for py_ver in 2 3; do if [[ `which python${py_ver}` ]]; then python${py_ver} -m pip freeze > /home/zuul/pip${py_ver}-freeze.txt fi done if [ `command -v dpkg` ]; then dpkg -l> /home/zuul/dpkg-l.txt fi if [ `command -v rpm` ]; then rpm -qa | sort > /home/zuul/rpm-qa.txt fi # Services status sudo systemctl status --all > services.txt 2>/dev/null # NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU # failed to start due to denials from SELinux — useful for CentOS # and Fedora machines. For Ubuntu (which runs AppArmor), DevStack # already captures the contents of /var/log/kern.log (via # `journalctl -t kernel` redirected into syslog.txt.gz), which # contains AppArmor-related messages. if [ -f /var/log/audit/audit.log ] ; then sudo cp /var/log/audit/audit.log /home/zuul/audit.log && chmod +r /home/zuul/audit.log; fi # gzip and save any coredumps in /var/core if [ -d /var/core ]; then sudo gzip -r /var/core sudo cp -r /var/core /home/zuul/ fi sudo ss -lntup | grep ':53' > /home/zuul/listen53.txt # NOTE(andreaf) Service logs are already in logs/ thanks for the # export-devstack-journal log. Apache logs are under apache/ thans to the # apache-logs-conf role. grep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \ sed -r 's/[0-9]{1,2}\:[0-9]{1,2}\:[0-9]{1,2}\.[0-9]{1,3}/ /g' | \ sed -r 's/[0-9]{1,2}\:[0-9]{1,2}\:[0-9]{1,2}/ /g' | \ sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' | sed -r 's/\[.*\]/ /g' | \ sed -r 's/\s[0-9]+\s/ /g' | \ awk '{if ($0 in seen) {seen[$0]++} else {out[++n]=$0;seen[$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]" :: " out[i] }' > /home/zuul/deprecations.log _uses_shell=True zuul_no_log=False zuul_log_id=0242ac17-0011-42b5-8cef-000000000033-1-controller zuul_output_max_bytes=1073741824 zuul_ansible_split_streams=False expand_argument_vars=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None creates=None removes=None stdin=None Apr 02 14:51:53 np0000163014 sudo[130114]: root : PWD=/home/zuul ; USER=root ; COMMAND=/usr/sbin/iptables-save Apr 02 14:51:53 np0000163014 sudo[130114]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=0) Apr 02 14:51:53 np0000163014 sudo[130114]: pam_unix(sudo:session): session closed for user root Apr 02 14:51:53 np0000163014 sudo[130125]: root : PWD=/home/zuul ; USER=root ; COMMAND=/usr/bin/systemctl status --all Apr 02 14:51:53 np0000163014 sudo[130125]: pam_unix(sudo:session): session opened for user root(uid=0) by zuul(uid=0) ● user.slice - User and Session Slice Loaded: loaded (/usr/lib/systemd/system/user.slice; static) Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Docs: man:systemd.special(7) Tasks: 15 Memory: 23.5G (peak: 23.6G) CPU: 23min 35.725s CGroup: /user.slice └─user-1000.slice ├─session-1.scope │ ├─ 828 "sshd: zuul [priv]" │ ├─ 853 "sshd: zuul@notty" │ ├─ 1054 /usr/bin/python3 │ ├─130107 sh -c "/bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '\"'\"'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3'\"'\"' && sleep 0'" │ ├─130108 /bin/sh -c "sudo -H -S -n -u root /bin/sh -c 'echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3' && sleep 0" │ ├─130109 sudo -H -S -n -u root /bin/sh -c "echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3" │ ├─130110 /bin/sh -c "echo BECOME-SUCCESS-rtcuibluluifzclfhmxzemcbwtngcwjb ; /usr/bin/python3" │ ├─130111 /usr/bin/python3 │ ├─130112 /bin/bash -c "sudo iptables-save > /home/zuul/iptables.txt\n\n# NOTE(sfernand): Run 'df' with a 60s timeout to prevent hangs from\n# stale NFS mounts.\ntimeout -s 9 60s df -h > /home/zuul/df.txt || true\n# If 'df' times out, the mount output helps debug which NFS share\n# is unresponsive.\nmount > /home/zuul/mount.txt\n\nfor py_ver in 2 3; do\n if [[ \`which python\${py_ver}\` ]]; then\n python\${py_ver} -m pip freeze > /home/zuul/pip\${py_ver}-freeze.txt\n fi\ndone\n\nif [ \`command -v dpkg\` ]; then\n dpkg -l> /home/zuul/dpkg-l.txt\nfi\nif [ \`command -v rpm\` ]; then\n rpm -qa | sort > /home/zuul/rpm-qa.txt\nfi\n\n# Services status\nsudo systemctl status --all > services.txt 2>/dev/null\n\n# NOTE(kchamart) The 'audit.log' can be useful in cases when QEMU\n# failed to start due to denials from SELinux — useful for CentOS\n# and Fedora machines. For Ubuntu (which runs AppArmor), DevStack\n# already captures the contents of /var/log/kern.log (via\n# \`journalctl -t kernel\` redirected into syslog.txt.gz), which\n# contains AppArmor-related messages.\nif [ -f /var/log/audit/audit.log ] ; then\n sudo cp /var/log/audit/audit.log /home/zuul/audit.log &&\n chmod +r /home/zuul/audit.log;\nfi\n\n# gzip and save any coredumps in /var/core\nif [ -d /var/core ]; then\n sudo gzip -r /var/core\n sudo cp -r /var/core /home/zuul/\nfi\n\nsudo ss -lntup | grep ':53' > /home/zuul/listen53.txt\n\n# NOTE(andreaf) Service logs are already in logs/ thanks for the\n# export-devstack-journal log. Apache logs are under apache/ thans to the\n# apache-logs-conf role.\ngrep -i deprecat /home/zuul/logs/*.txt /home/zuul/apache/*.log | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}\\.[0-9]{1,3}/ /g' | \\\n sed -r 's/[0-9]{1,2}\\:[0-9]{1,2}\\:[0-9]{1,2}/ /g' | \\\n sed -r 's/[0-9]{1,4}-[0-9]{1,2}-[0-9]{1,4}/ /g' |\n sed -r 's/\\[.*\\]/ /g' | \\\n sed -r 's/\\s[0-9]+\\s/ /g' | \\\n awk '{if (\$0 in seen) {seen[\$0]++} else {out[++n]=\$0;seen[\$0]=1}} END { for (i=1; i<=n; i++) print seen[out[i]]\" :: \" out[i] }' > /home/zuul/deprecations.log\n" │ ├─130125 sudo systemctl status --all │ └─130126 systemctl status --all └─user@1000.service └─init.scope ├─833 /usr/lib/systemd/systemd --user └─834 "(sd-pam)" Notice: journal has been rotated since unit was started, output may be incomplete. ● cloud-init-hotplugd.socket - cloud-init hotplug hook socket Loaded: loaded (/usr/lib/systemd/system/cloud-init-hotplugd.socket; enabled; preset: enabled) Active: active (listening) since Thu 2026-04-02 14:21:10 UTC; 30min ago Triggers: ● cloud-init-hotplugd.service Listen: /run/cloud-init/share/hook-hotplug-cmd (FIFO) CGroup: /system.slice/cloud-init-hotplugd.socket Apr 02 14:21:10 np0000163014 systemd[1]: Listening on cloud-init-hotplugd.socket - cloud-init hotplug hook socket. ● dbus.socket - D-Bus System Message Bus Socket Loaded: loaded (/usr/lib/systemd/system/dbus.socket; static) Active: active (running) since Thu 2026-04-02 14:21:10 UTC; 30min ago Triggers: ● dbus.service Listen: /run/dbus/system_bus_socket (Stream) CGroup: /system.slice/dbus.socket Apr 02 14:21:10 np0000163014 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. ● dm-event.socket - Device-mapper event daemon FIFOs Loaded: loaded (/usr/lib/systemd/system/dm-event.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:28:23 UTC; 23min ago Triggers: ● dm-event.service Docs: man:dmeventd(8) Listen: /run/dmeventd-server (FIFO) /run/dmeventd-client (FIFO) CGroup: /system.slice/dm-event.socket Apr 02 14:28:23 np0000163014 systemd[1]: Listening on dm-event.socket - Device-mapper event daemon FIFOs. ● docker.socket - Docker Socket for the API Loaded: loaded (/usr/lib/systemd/system/docker.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:29:34 UTC; 22min ago Triggers: ● docker.service Listen: /run/docker.sock (Stream) Tasks: 0 (limit: 77075) Memory: 0B (peak: 256.0K) CPU: 1ms CGroup: /system.slice/docker.socket Apr 02 14:29:34 np0000163014 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 02 14:29:34 np0000163014 systemd[1]: Listening on docker.socket - Docker Socket for the API. ● epmd.socket - Erlang Port Mapper Daemon Activation Socket Loaded: loaded (/usr/lib/systemd/system/epmd.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:31:29 UTC; 20min ago Triggers: ● epmd.service Listen: [::]:4369 (Stream) Tasks: 0 (limit: 77075) Memory: 8.0K (peak: 256.0K) CPU: 807us CGroup: /system.slice/epmd.socket Apr 02 14:31:29 np0000163014 systemd[1]: Listening on epmd.socket - Erlang Port Mapper Daemon Activation Socket. ● iscsid.socket - Open-iSCSI iscsid Socket Loaded: loaded (/usr/lib/systemd/system/iscsid.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:21:10 UTC; 30min ago Triggers: ● iscsid.service Docs: man:iscsid(8) man:iscsiadm(8) Listen: @ISCSIADM_ABSTRACT_NAMESPACE (Stream) CGroup: /system.slice/iscsid.socket Apr 02 14:21:10 np0000163014 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. ● libvirtd-admin.socket - libvirt legacy monolithic daemon admin socket Loaded: loaded (/usr/lib/systemd/system/libvirtd-admin.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:34:41 UTC; 17min ago Triggers: ● libvirtd.service Listen: /run/libvirt/libvirt-admin-sock (Stream) CGroup: /system.slice/libvirtd-admin.socket Apr 02 14:34:41 np0000163014 systemd[1]: Listening on libvirtd-admin.socket - libvirt legacy monolithic daemon admin socket. ● libvirtd-ro.socket - libvirt legacy monolithic daemon read-only socket Loaded: loaded (/usr/lib/systemd/system/libvirtd-ro.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:34:41 UTC; 17min ago Triggers: ● libvirtd.service Listen: /run/libvirt/libvirt-sock-ro (Stream) CGroup: /system.slice/libvirtd-ro.socket Apr 02 14:34:41 np0000163014 systemd[1]: Listening on libvirtd-ro.socket - libvirt legacy monolithic daemon read-only socket. ● libvirtd.socket - libvirt legacy monolithic daemon socket Loaded: loaded (/usr/lib/systemd/system/libvirtd.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:34:41 UTC; 17min ago Triggers: ● libvirtd.service Listen: /run/libvirt/libvirt-sock (Stream) Tasks: 0 (limit: 77075) Memory: 0B (peak: 256.0K) CPU: 1ms CGroup: /system.slice/libvirtd.socket Apr 02 14:34:41 np0000163014 systemd[1]: Starting libvirtd.socket - libvirt legacy monolithic daemon socket... Apr 02 14:34:41 np0000163014 systemd[1]: Listening on libvirtd.socket - libvirt legacy monolithic daemon socket. ● lvm2-lvmpolld.socket - LVM2 poll daemon socket Loaded: loaded (/usr/lib/systemd/system/lvm2-lvmpolld.socket; enabled; preset: enabled) Active: active (listening) since Thu 2026-04-02 14:28:25 UTC; 23min ago Triggers: ● lvm2-lvmpolld.service Docs: man:lvmpolld(8) Listen: /run/lvm/lvmpolld.socket (Stream) CGroup: /system.slice/lvm2-lvmpolld.socket Apr 02 14:28:25 np0000163014 systemd[1]: Listening on lvm2-lvmpolld.socket - LVM2 poll daemon socket. ● rpcbind.socket - RPCbind Server Activation Socket Loaded: loaded (/usr/lib/systemd/system/rpcbind.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:36:32 UTC; 15min ago Triggers: ● rpcbind.service Listen: /run/rpcbind.sock (Stream) 0.0.0.0:111 (Stream) 0.0.0.0:111 (Datagram) [::]:111 (Stream) [::]:111 (Datagram) Tasks: 0 (limit: 77075) Memory: 20.0K (peak: 272.0K) CPU: 5ms CGroup: /system.slice/rpcbind.socket Apr 02 14:36:32 np0000163014 systemd[1]: Listening on rpcbind.socket - RPCbind Server Activation Socket. ● ssh.socket - OpenBSD Secure Shell server socket Loaded: loaded (/usr/lib/systemd/system/ssh.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:21:10 UTC; 30min ago Triggers: ● ssh.service Listen: 0.0.0.0:22 (Stream) [::]:22 (Stream) Tasks: 0 (limit: 77075) Memory: 12.0K (peak: 264.0K) CPU: 2ms CGroup: /system.slice/ssh.socket Apr 02 14:21:10 np0000163014 systemd[1]: Listening on ssh.socket - OpenBSD Secure Shell server socket. ● syslog.socket - Syslog Socket Loaded: loaded (/usr/lib/systemd/system/syslog.socket; static) Active: active (running) since Thu 2026-04-02 14:27:55 UTC; 23min ago Triggers: ● rsyslog.service Docs: man:systemd.special(7) https://www.freedesktop.org/wiki/Software/systemd/syslog Listen: /run/systemd/journal/syslog (Datagram) CGroup: /system.slice/syslog.socket Apr 02 14:27:55 np0000163014 systemd[1]: Listening on syslog.socket - Syslog Socket. ● systemd-coredump.socket - Process Core Dump Socket Loaded: loaded (/usr/lib/systemd/system/systemd-coredump.socket; static) Active: active (listening) since Thu 2026-04-02 14:34:36 UTC; 17min ago Docs: man:systemd-coredump(8) Listen: /run/systemd/coredump (SequentialPacket) Accepted: 0; Connected: 0; CGroup: /system.slice/systemd-coredump.socket Apr 02 14:34:36 np0000163014 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. ● systemd-fsckd.socket - fsck to fsckd communication Socket Loaded: loaded (/usr/lib/systemd/system/systemd-fsckd.socket; static) Active: active (listening) since Thu 2026-04-02 14:21:03 UTC; 30min ago Triggers: ● systemd-fsckd.service Docs: man:systemd-fsckd.service(8) man:systemd-fsck@.service(8) man:systemd-fsck-root.service(8) Listen: /run/systemd/fsck.progress (Stream) CGroup: /system.slice/systemd-fsckd.socket Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-initctl.socket - initctl Compatibility Named Pipe Loaded: loaded (/usr/lib/systemd/system/systemd-initctl.socket; static) Active: active (listening) since Thu 2026-04-02 14:21:03 UTC; 30min ago Triggers: ● systemd-initctl.service Docs: man:systemd-initctl.socket(8) Listen: /run/initctl (FIFO) CGroup: /system.slice/systemd-initctl.socket Notice: journal has been rotated since unit was started, output may be incomplete. ○ systemd-journald-audit.socket - Journal Audit Socket Loaded: loaded (/usr/lib/systemd/system/systemd-journald-audit.socket; disabled; preset: enabled) Active: inactive (dead) Triggers: ● systemd-journald.service Docs: man:systemd-journald.service(8) man:journald.conf(5) Listen: audit 1 (Netlink) ● systemd-journald-dev-log.socket - Journal Socket (/dev/log) Loaded: loaded (/usr/lib/systemd/system/systemd-journald-dev-log.socket; static) Active: active (running) since Thu 2026-04-02 14:21:03 UTC; 30min ago Triggers: ● systemd-journald.service Docs: man:systemd-journald.service(8) man:journald.conf(5) Listen: /run/systemd/journal/dev-log (Datagram) CGroup: /system.slice/systemd-journald-dev-log.socket Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-journald.socket - Journal Socket Loaded: loaded (/usr/lib/systemd/system/systemd-journald.socket; static) Active: active (running) since Thu 2026-04-02 14:21:04 UTC; 30min ago Triggers: ● systemd-journald.service Docs: man:systemd-journald.service(8) man:journald.conf(5) Listen: /run/systemd/journal/socket (Datagram) /run/systemd/journal/stdout (Stream) CGroup: /system.slice/systemd-journald.socket Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-networkd.socket - Network Service Netlink Socket Loaded: loaded (/usr/lib/systemd/system/systemd-networkd.socket; disabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:21:05 UTC; 30min ago Triggers: ● systemd-networkd.service Docs: man:systemd-networkd.service(8) man:rtnetlink(7) Listen: route 1361 (Netlink) CGroup: /system.slice/systemd-networkd.socket Apr 02 14:21:05 np0000163014 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. ○ systemd-pcrextend.socket - TPM2 PCR Extension (Varlink) Loaded: loaded (/usr/lib/systemd/system/systemd-pcrextend.socket; disabled; preset: enabled) Active: inactive (dead) Condition: start condition unmet at Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-pcrextend(8) Listen: /run/systemd/io.systemd.PCRExtend (Stream) Accepted: 0; Connected: 0; ● systemd-rfkill.socket - Load/Save RF Kill Switch Status /dev/rfkill Watch Loaded: loaded (/usr/lib/systemd/system/systemd-rfkill.socket; static) Active: active (listening) since Thu 2026-04-02 14:21:04 UTC; 30min ago Triggers: ● systemd-rfkill.service Docs: man:systemd-rfkill.socket(8) Listen: /dev/rfkill (Special) CGroup: /system.slice/systemd-rfkill.socket Apr 02 14:21:04 ubuntu systemd[1]: Listening on systemd-rfkill.socket - Load/Save RF Kill Switch Status /dev/rfkill Watch. ● systemd-sysext.socket - System Extension Image Management (Varlink) Loaded: loaded (/usr/lib/systemd/system/systemd-sysext.socket; disabled; preset: enabled) Active: active (listening) since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd-sysext(8) Listen: /run/systemd/io.systemd.sysext (Stream) Accepted: 0; Connected: 0; CGroup: /system.slice/systemd-sysext.socket Apr 02 14:21:04 ubuntu systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). ● systemd-udevd-control.socket - udev Control Socket Loaded: loaded (/usr/lib/systemd/system/systemd-udevd-control.socket; static) Active: active (running) since Thu 2026-04-02 14:21:04 UTC; 30min ago Triggers: ● systemd-udevd.service Docs: man:systemd-udevd-control.socket(8) man:udev(7) Listen: /run/udev/control (SequentialPacket) CGroup: /system.slice/systemd-udevd-control.socket Notice: journal has been rotated since unit was started, output may be incomplete. ● systemd-udevd-kernel.socket - udev Kernel Socket Loaded: loaded (/usr/lib/systemd/system/systemd-udevd-kernel.socket; static) Active: active (running) since Thu 2026-04-02 14:21:04 UTC; 30min ago Triggers: ● systemd-udevd.service Docs: man:systemd-udevd-kernel.socket(8) man:udev(7) Listen: kobject-uevent 1 (Netlink) CGroup: /system.slice/systemd-udevd-kernel.socket Notice: journal has been rotated since unit was started, output may be incomplete. ● uuidd.socket - UUID daemon activation socket Loaded: loaded (/usr/lib/systemd/system/uuidd.socket; enabled; preset: enabled) Active: active (listening) since Thu 2026-04-02 14:21:10 UTC; 30min ago Triggers: ● uuidd.service Listen: /run/uuidd/request (Stream) CGroup: /system.slice/uuidd.socket Apr 02 14:21:10 np0000163014 systemd[1]: Listening on uuidd.socket - UUID daemon activation socket. ● virtlockd-admin.socket - libvirt locking daemon admin socket Loaded: loaded (/usr/lib/systemd/system/virtlockd-admin.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:34:43 UTC; 17min ago Triggers: ● virtlockd.service Listen: /run/libvirt/virtlockd-admin-sock (Stream) CGroup: /system.slice/virtlockd-admin.socket Apr 02 14:34:43 np0000163014 systemd[1]: Listening on virtlockd-admin.socket - libvirt locking daemon admin socket. ● virtlockd.socket - libvirt locking daemon socket Loaded: loaded (/usr/lib/systemd/system/virtlockd.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:34:41 UTC; 17min ago Triggers: ● virtlockd.service Listen: /run/libvirt/virtlockd-sock (Stream) CGroup: /system.slice/virtlockd.socket Apr 02 14:34:41 np0000163014 systemd[1]: Listening on virtlockd.socket - libvirt locking daemon socket. ● virtlogd-admin.socket - libvirt logging daemon admin socket Loaded: loaded (/usr/lib/systemd/system/virtlogd-admin.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:34:43 UTC; 17min ago Triggers: ● virtlogd.service Listen: /run/libvirt/virtlogd-admin-sock (Stream) CGroup: /system.slice/virtlogd-admin.socket Apr 02 14:34:43 np0000163014 systemd[1]: Listening on virtlogd-admin.socket - libvirt logging daemon admin socket. ● virtlogd.socket - libvirt logging daemon socket Loaded: loaded (/usr/lib/systemd/system/virtlogd.socket; enabled; preset: enabled) Active: active (running) since Thu 2026-04-02 14:34:41 UTC; 17min ago Triggers: ● virtlogd.service Listen: /run/libvirt/virtlogd-sock (Stream) CGroup: /system.slice/virtlogd.socket Apr 02 14:34:41 np0000163014 systemd[1]: Listening on virtlogd.socket - libvirt logging daemon socket. ● root-swapfile.swap - /root/swapfile Loaded: loaded (/etc/fstab; generated) Active: active since Thu 2026-04-02 14:24:36 UTC; 27min ago What: /root/swapfile Docs: man:fstab(5) man:systemd-fstab-generator(8) ● basic.target - Basic System Loaded: loaded (/usr/lib/systemd/system/basic.target; static) Active: active since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:10 np0000163014 systemd[1]: Reached target basic.target - Basic System. ○ blockdev@dev-disk-by\x2dlabel-cloudimg\x2drootfs.target - Block Device Preparation for /dev/disk/by-label/cloudimg-rootfs Loaded: loaded (/usr/lib/systemd/system/blockdev@.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● cloud-config.target - Cloud-config availability Loaded: loaded (/usr/lib/systemd/system/cloud-config.target; static) Active: active since Thu 2026-04-02 14:21:10 UTC; 30min ago Apr 02 14:21:10 np0000163014 systemd[1]: Reached target cloud-config.target - Cloud-config availability. ● cloud-init.target - Cloud-init target Loaded: loaded (/usr/lib/systemd/system/cloud-init.target; enabled-runtime; preset: enabled) Active: active since Thu 2026-04-02 14:21:11 UTC; 30min ago Apr 02 14:21:11 np0000163014 systemd[1]: Reached target cloud-init.target - Cloud-init target. ○ cryptsetup-pre.target - Local Encrypted Volumes (Pre) Loaded: loaded (/usr/lib/systemd/system/cryptsetup-pre.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● cryptsetup.target - Local Encrypted Volumes Loaded: loaded (/usr/lib/systemd/system/cryptsetup.target; static) Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Docs: man:systemd.special(7) Notice: journal has been rotated since unit was started, output may be incomplete. ○ emergency.target - Emergency Mode Loaded: loaded (/usr/lib/systemd/system/emergency.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ final.target - Late Shutdown Services Loaded: loaded (/usr/lib/systemd/system/final.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● first-boot-complete.target - First Boot Complete Loaded: loaded (/usr/lib/systemd/system/first-boot-complete.target; static) Active: active since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:04 ubuntu systemd[1]: Reached target first-boot-complete.target - First Boot Complete. ○ getty-pre.target - Preparation for Logins Loaded: loaded (/usr/lib/systemd/system/getty-pre.target; static) Active: inactive (dead) Docs: man:systemd.special(7) man:systemd-getty-generator(8) https://0pointer.de/blog/projects/serial-console.html ● getty.target - Login Prompts Loaded: loaded (/usr/lib/systemd/system/getty.target; static) Active: active since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd.special(7) man:systemd-getty-generator(8) https://0pointer.de/blog/projects/serial-console.html Apr 02 14:21:10 np0000163014 systemd[1]: Reached target getty.target - Login Prompts. ● graphical.target - Graphical Interface Loaded: loaded (/usr/lib/systemd/system/graphical.target; static) Active: active since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:10 np0000163014 systemd[1]: Reached target graphical.target - Graphical Interface. ○ hibernate.target - System Hibernation Loaded: loaded (/usr/lib/systemd/system/hibernate.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ hybrid-sleep.target - Hybrid Suspend+Hibernate Loaded: loaded (/usr/lib/systemd/system/hybrid-sleep.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ initrd-fs.target - Initrd File Systems Loaded: loaded (/usr/lib/systemd/system/initrd-fs.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ initrd-root-device.target - Initrd Root Device Loaded: loaded (/usr/lib/systemd/system/initrd-root-device.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ initrd-root-fs.target - Initrd Root File System Loaded: loaded (/usr/lib/systemd/system/initrd-root-fs.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ initrd-switch-root.target - Switch Root Loaded: loaded (/usr/lib/systemd/system/initrd-switch-root.target; static) Active: inactive (dead) ○ initrd-usr-fs.target - Initrd /usr File System Loaded: loaded (/usr/lib/systemd/system/initrd-usr-fs.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ initrd.target - Initrd Default Target Loaded: loaded (/usr/lib/systemd/system/initrd.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● integritysetup.target - Local Integrity Protected Volumes Loaded: loaded (/usr/lib/systemd/system/integritysetup.target; static) Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Docs: man:systemd.special(7) Notice: journal has been rotated since unit was started, output may be incomplete. ● local-fs-pre.target - Preparation for Local File Systems Loaded: loaded (/usr/lib/systemd/system/local-fs-pre.target; static) Active: active since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:04 ubuntu systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. ● local-fs.target - Local File Systems Loaded: loaded (/usr/lib/systemd/system/local-fs.target; static) Active: active since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:04 ubuntu systemd[1]: Reached target local-fs.target - Local File Systems. ● machines.target - Containers Loaded: loaded (/usr/lib/systemd/system/machines.target; enabled; preset: enabled) Active: active since Thu 2026-04-02 14:34:37 UTC; 17min ago Docs: man:systemd.special(7) Apr 02 14:34:37 np0000163014 systemd[1]: Reached target machines.target - Containers. ● multi-user.target - Multi-User System Loaded: loaded (/usr/lib/systemd/system/multi-user.target; static) Active: active since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:10 np0000163014 systemd[1]: Reached target multi-user.target - Multi-User System. ● network-online.target - Network is Online Loaded: loaded (/usr/lib/systemd/system/network-online.target; static) Active: active since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd.special(7) https://systemd.io/NETWORK_ONLINE Apr 02 14:21:10 np0000163014 systemd[1]: Reached target network-online.target - Network is Online. ● network-pre.target - Preparation for Network Loaded: loaded (/usr/lib/systemd/system/network-pre.target; static) Active: active since Thu 2026-04-02 14:21:05 UTC; 30min ago Docs: man:systemd.special(7) https://systemd.io/NETWORK_ONLINE Apr 02 14:21:05 np0000163014 systemd[1]: Reached target network-pre.target - Preparation for Network. ● network.target - Network Loaded: loaded (/usr/lib/systemd/system/network.target; static) Active: active since Thu 2026-04-02 14:21:05 UTC; 30min ago Docs: man:systemd.special(7) https://systemd.io/NETWORK_ONLINE Apr 02 14:21:05 np0000163014 systemd[1]: Reached target network.target - Network. ● nfs-client.target - NFS client services Loaded: loaded (/usr/lib/systemd/system/nfs-client.target; enabled; preset: enabled) Active: active since Thu 2026-04-02 14:36:35 UTC; 15min ago Apr 02 14:36:35 np0000163014 systemd[1]: Reached target nfs-client.target - NFS client services. ● nss-lookup.target - Host and Network Name Lookups Loaded: loaded (/usr/lib/systemd/system/nss-lookup.target; static) Active: active since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:04 ubuntu systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. ○ nss-user-lookup.target - User and Group Name Lookups Loaded: loaded (/usr/lib/systemd/system/nss-user-lookup.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● paths.target - Path Units Loaded: loaded (/usr/lib/systemd/system/paths.target; static) Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Docs: man:systemd.special(7) Notice: journal has been rotated since unit was started, output may be incomplete. ○ remote-cryptsetup.target - Remote Encrypted Volumes Loaded: loaded (/usr/lib/systemd/system/remote-cryptsetup.target; disabled; preset: enabled) Active: inactive (dead) Docs: man:systemd.special(7) ● remote-fs-pre.target - Preparation for Remote File Systems Loaded: loaded (/usr/lib/systemd/system/remote-fs-pre.target; static) Active: active since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:10 np0000163014 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. ● remote-fs.target - Remote File Systems Loaded: loaded (/usr/lib/systemd/system/remote-fs.target; enabled; preset: enabled) Active: active since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:10 np0000163014 systemd[1]: Reached target remote-fs.target - Remote File Systems. ○ remote-veritysetup.target - Remote Verity Protected Volumes Loaded: loaded (/usr/lib/systemd/system/remote-veritysetup.target; disabled; preset: enabled) Active: inactive (dead) Docs: man:systemd.special(7) ○ rescue.target - Rescue Mode Loaded: loaded (/usr/lib/systemd/system/rescue.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● rpc_pipefs.target Loaded: loaded (/run/systemd/generator/rpc_pipefs.target; generated) Active: active since Thu 2026-04-02 14:36:35 UTC; 15min ago Apr 02 14:36:35 np0000163014 systemd[1]: Reached target rpc_pipefs.target. ● rpcbind.target - RPC Port Mapper Loaded: loaded (/usr/lib/systemd/system/rpcbind.target; static) Active: active since Thu 2026-04-02 14:36:32 UTC; 15min ago Docs: man:systemd.special(7) Apr 02 14:36:32 np0000163014 systemd[1]: Reached target rpcbind.target - RPC Port Mapper. ○ shutdown.target - System Shutdown Loaded: loaded (/usr/lib/systemd/system/shutdown.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ sleep.target - Sleep Loaded: loaded (/usr/lib/systemd/system/sleep.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● slices.target - Slice Units Loaded: loaded (/usr/lib/systemd/system/slices.target; static) Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Docs: man:systemd.special(7) Notice: journal has been rotated since unit was started, output may be incomplete. ● sockets.target - Socket Units Loaded: loaded (/usr/lib/systemd/system/sockets.target; static) Active: active since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:10 np0000163014 systemd[1]: Reached target sockets.target - Socket Units. ○ soft-reboot.target - Reboot System Userspace Loaded: loaded (/usr/lib/systemd/system/soft-reboot.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ suspend-then-hibernate.target - Suspend; Hibernate if not used for a period of time Loaded: loaded (/usr/lib/systemd/system/suspend-then-hibernate.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ suspend.target - Suspend Loaded: loaded (/usr/lib/systemd/system/suspend.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● swap.target - Swaps Loaded: loaded (/usr/lib/systemd/system/swap.target; static) Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Docs: man:systemd.special(7) Notice: journal has been rotated since unit was started, output may be incomplete. ● sysinit.target - System Initialization Loaded: loaded (/usr/lib/systemd/system/sysinit.target; static) Active: active since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:10 np0000163014 systemd[1]: Reached target sysinit.target - System Initialization. ● time-set.target - System Time Set Loaded: loaded (/usr/lib/systemd/system/time-set.target; static) Active: active since Thu 2026-04-02 14:21:04 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:04 ubuntu systemd[1]: Reached target time-set.target - System Time Set. ○ time-sync.target - System Time Synchronized Loaded: loaded (/usr/lib/systemd/system/time-sync.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● timers.target - Timer Units Loaded: loaded (/usr/lib/systemd/system/timers.target; static) Active: active since Thu 2026-04-02 14:21:10 UTC; 30min ago Docs: man:systemd.special(7) Apr 02 14:21:10 np0000163014 systemd[1]: Reached target timers.target - Timer Units. ○ umount.target - Unmount All Filesystems Loaded: loaded (/usr/lib/systemd/system/umount.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ○ veritysetup-pre.target - Local Verity Protected Volumes (Pre) Loaded: loaded (/usr/lib/systemd/system/veritysetup-pre.target; static) Active: inactive (dead) Docs: man:systemd.special(7) ● veritysetup.target - Local Verity Protected Volumes Loaded: loaded (/usr/lib/systemd/system/veritysetup.target; static) Active: active since Thu 2026-04-02 14:21:03 UTC; 30min ago Docs: man:systemd.special(7) Notice: journal has been rotated since unit was started, output may be incomplete. ● virt-guest-shutdown.target - libvirt guests shutdown target Loaded: loaded (/usr/lib/systemd/system/virt-guest-shutdown.target; static) Active: active since Thu 2026-04-02 14:34:43 UTC; 17min ago Docs: https://libvirt.org/ Apr 02 14:34:43 np0000163014 systemd[1]: Reached target virt-guest-shutdown.target - libvirt guests shutdown target. ● apt-daily-upgrade.timer - Daily apt upgrade and clean activities Loaded: loaded (/usr/lib/systemd/system/apt-daily-upgrade.timer; enabled; preset: enabled) Active: active (waiting) since Thu 2026-04-02 14:21:10 UTC; 30min ago Trigger: Fri 2026-04-03 06:33:50 UTC; 15h left Triggers: ● apt-daily-upgrade.service Apr 02 14:21:10 np0000163014 systemd[1]: Started apt-daily-upgrade.timer - Daily apt upgrade and clean activities. ● apt-daily.timer - Daily apt download activities Loaded: loaded (/usr/lib/systemd/system/apt-daily.timer; enabled; preset: enabled) Active: active (waiting) since Thu 2026-04-02 14:21:10 UTC; 30min ago Trigger: Thu 2026-04-02 19:47:34 UTC; 4h 55min left Triggers: ● apt-daily.service Apr 02 14:21:10 np0000163014 systemd[1]: Started apt-daily.timer - Daily apt download activities. ● dpkg-db-backup.timer - Daily dpkg database backup timer Loaded: loaded (/usr/lib/systemd/system/dpkg-db-backup.timer; enabled; preset: enabled) Active: active (waiting) since Thu 2026-04-02 14:21:10 UTC; 30min ago Trigger: Fri 2026-04-03 00:00:00 UTC; 9h left Triggers: ● dpkg-db-backup.service Docs: man:dpkg(1) Apr 02 14:21:10 np0000163014 systemd[1]: Started dpkg-db-backup.timer - Daily dpkg database backup timer. ● e2scrub_all.timer - Periodic ext4 Online Metadata Check for All Filesystems Loaded: loaded (/usr/lib/systemd/system/e2scrub_all.timer; enabled; preset: enabled) Active: active (waiting) since Thu 2026-04-02 14:21:10 UTC; 30min ago Trigger: Sun 2026-04-05 03:10:20 UTC; 2 days left Triggers: ● e2scrub_all.service Apr 02 14:21:10 np0000163014 systemd[1]: Started e2scrub_all.timer - Periodic ext4 Online Metadata Check for All Filesystems. ● fstrim.timer - Discard unused filesystem blocks once a week Loaded: loaded (/usr/lib/systemd/system/fstrim.timer; enabled; preset: enabled) Active: active (waiting) since Thu 2026-04-02 14:21:10 UTC; 30min ago Trigger: Mon 2026-04-06 00:41:04 UTC; 3 days left Triggers: ● fstrim.service Docs: man:fstrim Apr 02 14:21:10 np0000163014 systemd[1]: Started fstrim.timer - Discard unused filesystem blocks once a week. ● logrotate.timer - Daily rotation of log files Loaded: loaded (/usr/lib/systemd/system/logrotate.timer; enabled; preset: enabled) Active: active (waiting) since Thu 2026-04-02 14:31:28 UTC; 20min ago Trigger: Fri 2026-04-03 00:00:00 UTC; 9h left Triggers: ● logrotate.service Docs: man:logrotate(8) man:logrotate.conf(5) Apr 02 14:31:28 np0000163014 systemd[1]: Started logrotate.timer - Daily rotation of log files. ● man-db.timer - Daily man-db regeneration Loaded: loaded (/usr/lib/systemd/system/man-db.timer; enabled; preset: enabled) Active: active (waiting) since Thu 2026-04-02 14:28:18 UTC; 23min ago Trigger: Fri 2026-04-03 01:42:59 UTC; 10h left Triggers: ● man-db.service Docs: man:mandb(8) Apr 02 14:28:18 np0000163014 systemd[1]: Started man-db.timer - Daily man-db regeneration. ● motd-news.timer - Message of the Day Loaded: loaded (/usr/lib/systemd/system/motd-news.timer; enabled; preset: enabled) Active: active (waiting) since Thu 2026-04-02 14:21:10 UTC; 30min ago Trigger: Fri 2026-04-03 07:34:55 UTC; 16h left Triggers: ● motd-news.service Apr 02 14:21:10 np0000163014 systemd[1]: Started motd-news.timer - Message of the Day. ● systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories Loaded: loaded (/usr/lib/systemd/system/systemd-tmpfiles-clean.timer; static) Active: active (waiting) since Thu 2026-04-02 14:21:10 UTC; 30min ago Trigger: Fri 2026-04-03 14:36:07 UTC; 23h left Triggers: ● systemd-tmpfiles-clean.service Docs: man:tmpfiles.d(5) man:systemd-tmpfiles(8) Apr 02 14:21:10 np0000163014 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories.