2026-04-17 08:19:59.018354 | Job console starting 2026-04-17 08:19:59.032352 | Updating git repos 2026-04-17 08:19:59.123281 | Cloning repos into workspace 2026-04-17 08:19:59.338040 | Restoring repo states 2026-04-17 08:19:59.371629 | Merging changes 2026-04-17 08:20:01.080874 | Checking out repos 2026-04-17 08:20:01.215466 | Preparing playbooks 2026-04-17 08:20:05.739789 | Running Ansible setup 2026-04-17 08:20:09.445213 | PRE-RUN START: [trusted : github.com/vexxhost/zuul-config/playbooks/base/pre.yaml@main] 2026-04-17 08:20:10.063340 | 2026-04-17 08:20:10.063460 | PLAY [localhost] 2026-04-17 08:20:10.071136 | 2026-04-17 08:20:10.071222 | TASK [Gathering Facts] 2026-04-17 08:20:10.972014 | localhost | ok 2026-04-17 08:20:10.983771 | 2026-04-17 08:20:10.983887 | TASK [Setup log path fact] 2026-04-17 08:20:11.003459 | localhost | ok 2026-04-17 08:20:11.017688 | 2026-04-17 08:20:11.017800 | TASK [set-zuul-log-path-fact : Set log path for a build] 2026-04-17 08:20:11.048461 | localhost | ok 2026-04-17 08:20:11.057882 | 2026-04-17 08:20:11.057979 | TASK [emit-job-header : Print job information] 2026-04-17 08:20:11.098746 | # Job Information 2026-04-17 08:20:11.156878 | Ansible Version: 2.16.16 2026-04-17 08:20:11.157102 | Job: atmosphere-molecule-aio-openvswitch 2026-04-17 08:20:11.157160 | Pipeline: check 2026-04-17 08:20:11.157206 | Executor: 0a8996d2b663 2026-04-17 08:20:11.157249 | Triggered by: https://github.com/vexxhost/atmosphere/pull/3834 2026-04-17 08:20:11.157299 | Event ID: 22e760f0-3a36-11f1-880d-f7c84e8af0b2 2026-04-17 08:20:11.163890 | 2026-04-17 08:20:11.165968 | LOOP [emit-job-header : Print node information] 2026-04-17 08:20:11.265024 | localhost | ok: 2026-04-17 08:20:11.265234 | localhost | # Node Information 2026-04-17 08:20:11.265263 | localhost | Inventory Hostname: instance 2026-04-17 08:20:11.265286 | localhost | Hostname: np0000167010 2026-04-17 08:20:11.265306 | localhost | Username: zuul 2026-04-17 08:20:11.265330 | localhost | Distro: Ubuntu 22.04 2026-04-17 08:20:11.265349 | localhost | Provider: yul1 2026-04-17 08:20:11.265368 | localhost | Region: ca-ymq-1 2026-04-17 08:20:11.265387 | localhost | Label: ubuntu-jammy-16 2026-04-17 08:20:11.265405 | localhost | Product Name: OpenStack Nova 2026-04-17 08:20:11.265423 | localhost | Interface IP: 199.204.45.242 2026-04-17 08:20:11.282248 | 2026-04-17 08:20:11.282429 | TASK [log-inventory : Ensure Zuul Ansible directory exists] 2026-04-17 08:20:11.703601 | localhost -> localhost | changed 2026-04-17 08:20:11.712204 | 2026-04-17 08:20:11.712322 | TASK [log-inventory : Copy ansible inventory to logs dir] 2026-04-17 08:20:12.639383 | localhost -> localhost | changed 2026-04-17 08:20:12.650027 | 2026-04-17 08:20:12.650129 | PLAY [all] 2026-04-17 08:20:12.661392 | 2026-04-17 08:20:12.661538 | TASK [add-build-sshkey : Check to see if ssh key was already created for this build] 2026-04-17 08:20:12.897628 | instance -> localhost | ok 2026-04-17 08:20:12.910065 | 2026-04-17 08:20:12.910199 | TASK [add-build-sshkey : Create a new key in workspace based on build UUID] 2026-04-17 08:20:12.944730 | instance | ok 2026-04-17 08:20:12.959928 | instance | included: /var/lib/zuul/builds/b48b679401b84230b69e09a70b891bfd/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey/tasks/create-key-and-replace.yaml 2026-04-17 08:20:12.965478 | 2026-04-17 08:20:12.965537 | TASK [add-build-sshkey : Create Temp SSH key] 2026-04-17 08:20:14.102355 | instance -> localhost | Generating public/private rsa key pair. 2026-04-17 08:20:14.102519 | instance -> localhost | Your identification has been saved in /var/lib/zuul/builds/b48b679401b84230b69e09a70b891bfd/work/b48b679401b84230b69e09a70b891bfd_id_rsa 2026-04-17 08:20:14.102549 | instance -> localhost | Your public key has been saved in /var/lib/zuul/builds/b48b679401b84230b69e09a70b891bfd/work/b48b679401b84230b69e09a70b891bfd_id_rsa.pub 2026-04-17 08:20:14.102573 | instance -> localhost | The key fingerprint is: 2026-04-17 08:20:14.102595 | instance -> localhost | SHA256:FiIlNISDwU/WGgKAVMAtsKQHTlWt4wJ2jjjFwbNtn+E zuul-build-sshkey 2026-04-17 08:20:14.102628 | instance -> localhost | The key's randomart image is: 2026-04-17 08:20:14.102650 | instance -> localhost | +---[RSA 3072]----+ 2026-04-17 08:20:14.102673 | instance -> localhost | |#O=*B.o | 2026-04-17 08:20:14.102695 | instance -> localhost | |BB*= = . | 2026-04-17 08:20:14.102715 | instance -> localhost | |ooB*+ o . | 2026-04-17 08:20:14.102736 | instance -> localhost | | +++o+.. . | 2026-04-17 08:20:14.102756 | instance -> localhost | |o.=..o.oS | 2026-04-17 08:20:14.102775 | instance -> localhost | |o. o .E. | 2026-04-17 08:20:14.102794 | instance -> localhost | | . . | 2026-04-17 08:20:14.102814 | instance -> localhost | | | 2026-04-17 08:20:14.102836 | instance -> localhost | | | 2026-04-17 08:20:14.102856 | instance -> localhost | +----[SHA256]-----+ 2026-04-17 08:20:14.102901 | instance -> localhost | ok: Runtime: 0:00:00.734587 2026-04-17 08:20:14.107816 | 2026-04-17 08:20:14.107877 | TASK [add-build-sshkey : Remote setup ssh keys (linux)] 2026-04-17 08:20:14.137839 | instance | ok 2026-04-17 08:20:14.150544 | instance | included: /var/lib/zuul/builds/b48b679401b84230b69e09a70b891bfd/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey/tasks/remote-linux.yaml 2026-04-17 08:20:14.161135 | 2026-04-17 08:20:14.453242 | TASK [add-build-sshkey : Remove previously added zuul-build-sshkey] 2026-04-17 08:20:14.480035 | instance | skipping: Conditional result was False 2026-04-17 08:20:14.493158 | 2026-04-17 08:20:14.493257 | TASK [add-build-sshkey : Enable access via build key on all nodes] 2026-04-17 08:20:14.930322 | instance | changed 2026-04-17 08:20:14.936110 | 2026-04-17 08:20:14.936180 | TASK [add-build-sshkey : Make sure user has a .ssh] 2026-04-17 08:20:15.121123 | instance | ok 2026-04-17 08:20:15.128065 | 2026-04-17 08:20:15.128158 | TASK [add-build-sshkey : Install build private key as SSH key on all nodes] 2026-04-17 08:20:15.611752 | instance | changed 2026-04-17 08:20:15.693858 | 2026-04-17 08:20:15.694094 | TASK [add-build-sshkey : Install build public key as SSH key on all nodes] 2026-04-17 08:20:16.154926 | instance | changed 2026-04-17 08:20:16.163954 | 2026-04-17 08:20:16.164058 | TASK [add-build-sshkey : Remote setup ssh keys (windows)] 2026-04-17 08:20:16.189361 | instance | skipping: Conditional result was False 2026-04-17 08:20:16.198820 | 2026-04-17 08:20:16.198952 | TASK [remove-zuul-sshkey : Remove master key from local agent] 2026-04-17 08:20:16.577287 | instance -> localhost | changed 2026-04-17 08:20:16.635976 | 2026-04-17 08:20:16.636130 | TASK [add-build-sshkey : Add back temp key] 2026-04-17 08:20:17.192906 | instance -> localhost | Identity added: /var/lib/zuul/builds/b48b679401b84230b69e09a70b891bfd/work/b48b679401b84230b69e09a70b891bfd_id_rsa (zuul-build-sshkey) 2026-04-17 08:20:17.193165 | instance -> localhost | ok: Runtime: 0:00:00.014461 2026-04-17 08:20:17.198769 | 2026-04-17 08:20:17.198838 | TASK [add-build-sshkey : Verify we can still SSH to all nodes] 2026-04-17 08:20:17.498085 | instance | ok 2026-04-17 08:20:17.502996 | 2026-04-17 08:20:17.503080 | TASK [add-build-sshkey : Verify we can still SSH to all nodes (windows)] 2026-04-17 08:20:17.526803 | instance | skipping: Conditional result was False 2026-04-17 08:20:17.544219 | 2026-04-17 08:20:17.544361 | TASK [prepare-workspace : Start zuul_console daemon.] 2026-04-17 08:20:17.819460 | instance | ok 2026-04-17 08:20:17.882936 | 2026-04-17 08:20:17.883033 | TASK [prepare-workspace : Synchronize src repos to workspace directory.] 2026-04-17 08:20:19.869107 | instance | Output suppressed because no_log was given 2026-04-17 08:20:19.878044 | 2026-04-17 08:20:19.878130 | LOOP [ensure-output-dirs : Empty Zuul Output directories by removing them] 2026-04-17 08:20:20.065919 | instance | ok: "logs" 2026-04-17 08:20:20.066204 | instance | ok: All items complete 2026-04-17 08:20:20.066245 | 2026-04-17 08:20:20.214412 | instance | ok: "artifacts" 2026-04-17 08:20:20.373865 | instance | ok: "docs" 2026-04-17 08:20:20.546435 | 2026-04-17 08:20:20.546569 | LOOP [ensure-output-dirs : Ensure Zuul Output directories exist] 2026-04-17 08:20:20.734009 | instance | changed: "logs" 2026-04-17 08:20:20.886087 | instance | changed: "artifacts" 2026-04-17 08:20:21.036921 | instance | changed: "docs" 2026-04-17 08:20:21.118448 | 2026-04-17 08:20:21.118547 | PLAY RECAP 2026-04-17 08:20:21.118596 | instance | ok: 15 changed: 8 unreachable: 0 failed: 0 skipped: 3 rescued: 0 ignored: 0 2026-04-17 08:20:21.118626 | localhost | ok: 6 changed: 2 unreachable: 0 failed: 0 skipped: 0 rescued: 0 ignored: 0 2026-04-17 08:20:21.118648 | 2026-04-17 08:20:21.256581 | PRE-RUN END RESULT_NORMAL: [trusted : github.com/vexxhost/zuul-config/playbooks/base/pre.yaml@main] 2026-04-17 08:20:21.268944 | PRE-RUN START: [untrusted : github.com/vexxhost/zuul-jobs/playbooks/molecule/pre.yaml@main] 2026-04-17 08:20:21.867711 | 2026-04-17 08:20:21.867896 | PLAY [all] 2026-04-17 08:20:21.881959 | 2026-04-17 08:20:21.882099 | TASK [setup-uv : Extract archive] 2026-04-17 08:20:24.171046 | instance | changed 2026-04-17 08:20:24.177066 | 2026-04-17 08:20:24.177168 | TASK [setup-uv : Print version] 2026-04-17 08:20:24.282277 | instance | uv 0.8.13 2026-04-17 08:20:24.714007 | instance | ok: Runtime: 0:00:00.014462 2026-04-17 08:20:24.722238 | 2026-04-17 08:20:24.722302 | PLAY RECAP 2026-04-17 08:20:24.722362 | instance | ok: 2 changed: 2 unreachable: 0 failed: 0 skipped: 0 rescued: 0 ignored: 0 2026-04-17 08:20:24.722395 | 2026-04-17 08:20:24.869161 | PRE-RUN END RESULT_NORMAL: [untrusted : github.com/vexxhost/zuul-jobs/playbooks/molecule/pre.yaml@main] 2026-04-17 08:20:24.880059 | PRE-RUN START: [untrusted : github.com/vexxhost/atmosphere/test-playbooks/molecule/pre.yml@main] 2026-04-17 08:20:25.474713 | 2026-04-17 08:20:25.474838 | PLAY [all] 2026-04-17 08:20:25.486076 | 2026-04-17 08:20:25.486154 | TASK [Install "jq" for log collection] 2026-04-17 08:21:53.361521 | instance | changed 2026-04-17 08:21:53.367265 | 2026-04-17 08:21:53.367343 | TASK [Install pip3 for Python package management] 2026-04-17 08:21:57.954992 | instance | changed 2026-04-17 08:21:57.960219 | 2026-04-17 08:21:57.960285 | TASK [Install Python "kubernetes" library for kubernetes.core modules] 2026-04-17 08:22:01.045269 | instance | changed 2026-04-17 08:22:01.048908 | 2026-04-17 08:22:01.048980 | PLAY [all] 2026-04-17 08:22:01.058312 | 2026-04-17 08:22:01.058380 | TASK [ensure-go : Check installed go version] 2026-04-17 08:22:01.604657 | instance | ok: ERROR (ignored) 2026-04-17 08:22:01.605259 | instance | { 2026-04-17 08:22:01.605342 | instance | "failed_when_result": false, 2026-04-17 08:22:01.605392 | instance | "msg": "[Errno 2] No such file or directory: b'go'", 2026-04-17 08:22:01.605440 | instance | "rc": 2 2026-04-17 08:22:01.605488 | instance | } 2026-04-17 08:22:01.611315 | 2026-04-17 08:22:01.611389 | TASK [ensure-go : Skip if correct version of go is installed] 2026-04-17 08:22:01.665612 | instance | ok 2026-04-17 08:22:01.675166 | instance | included: /var/lib/zuul/builds/b48b679401b84230b69e09a70b891bfd/untrusted/project_2/opendev.org/zuul/zuul-jobs/roles/ensure-go/tasks/install-go.yaml 2026-04-17 08:22:01.680520 | 2026-04-17 08:22:01.680596 | TASK [ensure-go : Create temp directory] 2026-04-17 08:22:01.982488 | instance | changed 2026-04-17 08:22:01.989237 | 2026-04-17 08:22:01.989308 | TASK [ensure-go : Get archive checksum] 2026-04-17 08:22:02.619282 | instance | ok: OK (64 bytes) 2026-04-17 08:22:02.625277 | 2026-04-17 08:22:02.625344 | TASK [ensure-go : Download go archive] 2026-04-17 08:22:03.886590 | instance | changed: OK (78559214 bytes) 2026-04-17 08:22:03.896895 | 2026-04-17 08:22:03.896969 | TASK [ensure-go : Install go] 2026-04-17 08:22:09.830910 | instance | changed 2026-04-17 08:22:09.838366 | 2026-04-17 08:22:09.838428 | PLAY [all] 2026-04-17 08:22:09.844976 | 2026-04-17 08:22:09.845040 | TASK [Copy inventory file for Zuul] 2026-04-17 08:22:10.582273 | instance | changed 2026-04-17 08:22:10.588062 | 2026-04-17 08:22:10.588136 | TASK [Switch "ansible_host" to private IP] 2026-04-17 08:22:10.935782 | instance | changed: 1 replacements made 2026-04-17 08:22:10.971062 | 2026-04-17 08:22:10.971236 | TASK [Run molecule prepare] 2026-04-17 08:22:11.313757 | instance | Using CPython 3.10.12 interpreter at: /usr/bin/python3 2026-04-17 08:22:11.313864 | instance | Creating virtual environment at: .venv 2026-04-17 08:22:11.340205 | instance | Building atmosphere @ file:///home/zuul/src/github.com/vexxhost/atmosphere 2026-04-17 08:22:11.364046 | instance | Downloading ansible-core (2.1MiB) 2026-04-17 08:22:11.364216 | instance | Downloading openstacksdk (1.7MiB) 2026-04-17 08:22:11.364423 | instance | Downloading pydantic-core (2.0MiB) 2026-04-17 08:22:11.364808 | instance | Downloading rjsonnet (1.2MiB) 2026-04-17 08:22:11.365357 | instance | Downloading pygments (1.2MiB) 2026-04-17 08:22:11.367822 | instance | Downloading netaddr (2.2MiB) 2026-04-17 08:22:11.368072 | instance | Downloading cryptography (4.2MiB) 2026-04-17 08:22:11.368744 | instance | Downloading kubernetes (1.9MiB) 2026-04-17 08:22:11.369038 | instance | Downloading setuptools (1.1MiB) 2026-04-17 08:22:11.681440 | instance | Building pyperclip==1.9.0 2026-04-17 08:22:11.769393 | instance | Downloading rjsonnet 2026-04-17 08:22:11.798044 | instance | Downloading pydantic-core 2026-04-17 08:22:11.846460 | instance | Downloading netaddr 2026-04-17 08:22:11.854924 | instance | Downloading pygments 2026-04-17 08:22:11.916066 | instance | Downloading setuptools 2026-04-17 08:22:11.931370 | instance | Downloading cryptography 2026-04-17 08:22:11.976897 | instance | Downloading kubernetes 2026-04-17 08:22:12.012703 | instance | Downloading ansible-core 2026-04-17 08:22:12.052333 | instance | Downloading openstacksdk 2026-04-17 08:22:12.391058 | instance | Built pyperclip==1.9.0 2026-04-17 08:22:12.580494 | instance | Built atmosphere @ file:///home/zuul/src/github.com/vexxhost/atmosphere 2026-04-17 08:22:12.623352 | instance | Installed 83 packages in 40ms 2026-04-17 08:22:13.301865 | instance | WARNING Molecule scenarios should migrate to 'extensions/molecule' 2026-04-17 08:22:13.945088 | instance | INFO [aio > discovery] scenario test matrix: prepare 2026-04-17 08:22:13.945183 | instance | INFO [aio > prerun] Performing prerun with role_name_check=0... 2026-04-17 08:23:03.119401 | instance | INFO [aio > prepare] Executing 2026-04-17 08:23:04.148282 | instance | 2026-04-17 08:23:04.148820 | instance | PLAY [Prepare] ***************************************************************** 2026-04-17 08:23:04.149145 | instance | 2026-04-17 08:23:04.149442 | instance | TASK [Gathering Facts] ********************************************************* 2026-04-17 08:23:04.149780 | instance | Friday 17 April 2026 08:23:04 +0000 (0:00:00.029) 0:00:00.029 ********** 2026-04-17 08:23:05.328627 | instance | [WARNING]: Platform linux on host instance is using the discovered Python 2026-04-17 08:23:05.328854 | instance | interpreter at /usr/bin/python3.10, but future installation of another Python 2026-04-17 08:23:05.329185 | instance | interpreter could change the meaning of that path. See 2026-04-17 08:23:05.329478 | instance | https://docs.ansible.com/ansible- 2026-04-17 08:23:05.329759 | instance | core/2.17/reference_appendices/interpreter_discovery.html for more information. 2026-04-17 08:23:05.337367 | instance | ok: [instance] 2026-04-17 08:23:05.337622 | instance | 2026-04-17 08:23:05.337918 | instance | TASK [Configure short hostname] ************************************************ 2026-04-17 08:23:05.338189 | instance | Friday 17 April 2026 08:23:05 +0000 (0:00:01.189) 0:00:01.219 ********** 2026-04-17 08:23:05.995787 | instance | changed: [instance] 2026-04-17 08:23:05.995973 | instance | 2026-04-17 08:23:05.996314 | instance | TASK [Ensure hostname inside hosts file] *************************************** 2026-04-17 08:23:05.996598 | instance | Friday 17 April 2026 08:23:05 +0000 (0:00:00.657) 0:00:01.877 ********** 2026-04-17 08:23:06.244986 | instance | [WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created 2026-04-17 08:23:06.245212 | instance | with a mode of 0700, this may cause issues when running as another user. To 2026-04-17 08:23:06.245465 | instance | avoid this, create the remote_tmp dir with the correct permissions manually 2026-04-17 08:23:06.254347 | instance | changed: [instance] 2026-04-17 08:23:06.254625 | instance | 2026-04-17 08:23:06.254936 | instance | TASK [Install "dirmngr" for GPG keyserver operations] ************************** 2026-04-17 08:23:06.255195 | instance | Friday 17 April 2026 08:23:06 +0000 (0:00:00.259) 0:00:02.136 ********** 2026-04-17 08:23:07.499872 | instance | ok: [instance] 2026-04-17 08:23:07.500110 | instance | 2026-04-17 08:23:07.500394 | instance | TASK [Purge "snapd" package] *************************************************** 2026-04-17 08:23:07.500685 | instance | Friday 17 April 2026 08:23:07 +0000 (0:00:01.245) 0:00:03.381 ********** 2026-04-17 08:23:08.280443 | instance | ok: [instance] 2026-04-17 08:23:08.280689 | instance | 2026-04-17 08:23:08.280963 | instance | PLAY [Generate workspace for Atmosphere] *************************************** 2026-04-17 08:23:08.281209 | instance | 2026-04-17 08:23:08.281504 | instance | TASK [Create folders for workspace] ******************************************** 2026-04-17 08:23:08.281774 | instance | Friday 17 April 2026 08:23:08 +0000 (0:00:00.780) 0:00:04.162 ********** 2026-04-17 08:23:09.365635 | instance | changed: [localhost] => (item=group_vars) 2026-04-17 08:23:09.365918 | instance | changed: [localhost] => (item=group_vars/all) 2026-04-17 08:23:09.366243 | instance | changed: [localhost] => (item=group_vars/controllers) 2026-04-17 08:23:09.366560 | instance | changed: [localhost] => (item=group_vars/cephs) 2026-04-17 08:23:09.366857 | instance | changed: [localhost] => (item=group_vars/computes) 2026-04-17 08:23:09.367139 | instance | changed: [localhost] => (item=host_vars) 2026-04-17 08:23:09.367424 | instance | 2026-04-17 08:23:09.367719 | instance | PLAY [Generate Ceph control plane configuration for workspace] ***************** 2026-04-17 08:23:09.367982 | instance | 2026-04-17 08:23:09.368263 | instance | TASK [Ensure the Ceph control plane configuration file exists] ***************** 2026-04-17 08:23:09.368565 | instance | Friday 17 April 2026 08:23:09 +0000 (0:00:01.084) 0:00:05.247 ********** 2026-04-17 08:23:09.559988 | instance | changed: [localhost] 2026-04-17 08:23:09.560222 | instance | 2026-04-17 08:23:09.560504 | instance | TASK [Load the current Ceph control plane configuration into a variable] ******* 2026-04-17 08:23:09.560774 | instance | Friday 17 April 2026 08:23:09 +0000 (0:00:00.194) 0:00:05.442 ********** 2026-04-17 08:23:09.590020 | instance | ok: [localhost] 2026-04-17 08:23:09.590327 | instance | 2026-04-17 08:23:09.590645 | instance | TASK [Generate Ceph control plane values for missing variables] **************** 2026-04-17 08:23:09.590939 | instance | Friday 17 April 2026 08:23:09 +0000 (0:00:00.030) 0:00:05.472 ********** 2026-04-17 08:23:09.644925 | instance | ok: [localhost] => (item={'key': 'ceph_fsid', 'value': 'cdde4b55-116f-5e4e-a8ba-4cc121f65bad'}) 2026-04-17 08:23:09.645211 | instance | ok: [localhost] => (item={'key': 'ceph_mon_public_network', 'value': '10.96.240.0/24'}) 2026-04-17 08:23:09.645473 | instance | 2026-04-17 08:23:09.645741 | instance | TASK [Write new Ceph control plane configuration file to disk] ***************** 2026-04-17 08:23:09.646014 | instance | Friday 17 April 2026 08:23:09 +0000 (0:00:00.054) 0:00:05.527 ********** 2026-04-17 08:23:10.200264 | instance | changed: [localhost] 2026-04-17 08:23:10.200474 | instance | 2026-04-17 08:23:10.200802 | instance | PLAY [Generate Ceph OSD configuration for workspace] *************************** 2026-04-17 08:23:10.201061 | instance | 2026-04-17 08:23:10.201339 | instance | TASK [Ensure the Ceph OSDs configuration file exists] ************************** 2026-04-17 08:23:10.201627 | instance | Friday 17 April 2026 08:23:10 +0000 (0:00:00.554) 0:00:06.082 ********** 2026-04-17 08:23:10.375184 | instance | changed: [localhost] 2026-04-17 08:23:10.375407 | instance | 2026-04-17 08:23:10.375708 | instance | TASK [Load the current Ceph OSDs configuration into a variable] **************** 2026-04-17 08:23:10.376019 | instance | Friday 17 April 2026 08:23:10 +0000 (0:00:00.174) 0:00:06.256 ********** 2026-04-17 08:23:10.403104 | instance | ok: [localhost] 2026-04-17 08:23:10.403345 | instance | 2026-04-17 08:23:10.403647 | instance | TASK [Generate Ceph OSDs values for missing variables] ************************* 2026-04-17 08:23:10.403898 | instance | Friday 17 April 2026 08:23:10 +0000 (0:00:00.028) 0:00:06.285 ********** 2026-04-17 08:23:10.440445 | instance | ok: [localhost] => (item={'key': 'ceph_osd_devices', 'value': ['/dev/vdb', '/dev/vdc', '/dev/vdd']}) 2026-04-17 08:23:10.440681 | instance | 2026-04-17 08:23:10.440951 | instance | TASK [Write new Ceph OSDs configuration file to disk] ************************** 2026-04-17 08:23:10.441220 | instance | Friday 17 April 2026 08:23:10 +0000 (0:00:00.036) 0:00:06.322 ********** 2026-04-17 08:23:10.808833 | instance | changed: [localhost] 2026-04-17 08:23:10.809106 | instance | 2026-04-17 08:23:10.809477 | instance | PLAY [Generate Kubernetes configuration for workspace] ************************* 2026-04-17 08:23:10.809774 | instance | 2026-04-17 08:23:10.810100 | instance | TASK [Ensure the Kubernetes configuration file exists] ************************* 2026-04-17 08:23:10.810664 | instance | Friday 17 April 2026 08:23:10 +0000 (0:00:00.367) 0:00:06.690 ********** 2026-04-17 08:23:11.008121 | instance | changed: [localhost] 2026-04-17 08:23:11.008360 | instance | 2026-04-17 08:23:11.008657 | instance | TASK [Load the current Kubernetes configuration into a variable] *************** 2026-04-17 08:23:11.008894 | instance | Friday 17 April 2026 08:23:10 +0000 (0:00:00.199) 0:00:06.890 ********** 2026-04-17 08:23:11.038190 | instance | ok: [localhost] 2026-04-17 08:23:11.038427 | instance | 2026-04-17 08:23:11.038743 | instance | TASK [Generate Kubernetes values for missing variables] ************************ 2026-04-17 08:23:11.039088 | instance | Friday 17 April 2026 08:23:11 +0000 (0:00:00.030) 0:00:06.920 ********** 2026-04-17 08:23:11.084738 | instance | ok: [localhost] => (item={'key': 'kubernetes_hostname', 'value': '10.96.240.10'}) 2026-04-17 08:23:11.085035 | instance | ok: [localhost] => (item={'key': 'kubernetes_keepalived_vrid', 'value': 42}) 2026-04-17 08:23:11.085329 | instance | ok: [localhost] => (item={'key': 'kubernetes_keepalived_vip', 'value': '10.96.240.10'}) 2026-04-17 08:23:11.085590 | instance | 2026-04-17 08:23:11.085871 | instance | TASK [Write new Kubernetes configuration file to disk] ************************* 2026-04-17 08:23:11.086097 | instance | Friday 17 April 2026 08:23:11 +0000 (0:00:00.046) 0:00:06.967 ********** 2026-04-17 08:23:11.444441 | instance | changed: [localhost] 2026-04-17 08:23:11.444671 | instance | 2026-04-17 08:23:11.444950 | instance | PLAY [Generate Keepalived configuration for workspace] ************************* 2026-04-17 08:23:11.445203 | instance | 2026-04-17 08:23:11.445496 | instance | TASK [Ensure the Keeaplived configuration file exists] ************************* 2026-04-17 08:23:11.445773 | instance | Friday 17 April 2026 08:23:11 +0000 (0:00:00.359) 0:00:07.326 ********** 2026-04-17 08:23:11.624594 | instance | changed: [localhost] 2026-04-17 08:23:11.624816 | instance | 2026-04-17 08:23:11.625121 | instance | TASK [Load the current Keepalived configuration into a variable] *************** 2026-04-17 08:23:11.625377 | instance | Friday 17 April 2026 08:23:11 +0000 (0:00:00.179) 0:00:07.506 ********** 2026-04-17 08:23:11.651667 | instance | ok: [localhost] 2026-04-17 08:23:11.651886 | instance | 2026-04-17 08:23:11.652175 | instance | TASK [Generate Keepalived values for missing variables] ************************ 2026-04-17 08:23:11.652447 | instance | Friday 17 April 2026 08:23:11 +0000 (0:00:00.027) 0:00:07.533 ********** 2026-04-17 08:23:11.687248 | instance | ok: [localhost] => (item={'key': 'keepalived_interface', 'value': 'br-ex'}) 2026-04-17 08:23:11.687531 | instance | ok: [localhost] => (item={'key': 'keepalived_vip', 'value': '10.96.250.10'}) 2026-04-17 08:23:11.687801 | instance | 2026-04-17 08:23:11.688079 | instance | TASK [Write new Keepalived configuration file to disk] ************************* 2026-04-17 08:23:11.688364 | instance | Friday 17 April 2026 08:23:11 +0000 (0:00:00.035) 0:00:07.569 ********** 2026-04-17 08:23:12.041941 | instance | changed: [localhost] 2026-04-17 08:23:12.042205 | instance | 2026-04-17 08:23:12.042619 | instance | PLAY [Generate endpoints for workspace] **************************************** 2026-04-17 08:23:12.042885 | instance | 2026-04-17 08:23:12.043178 | instance | TASK [Gathering Facts] ********************************************************* 2026-04-17 08:23:12.043468 | instance | Friday 17 April 2026 08:23:12 +0000 (0:00:00.354) 0:00:07.923 ********** 2026-04-17 08:23:12.725761 | instance | ok: [localhost] 2026-04-17 08:23:12.726053 | instance | 2026-04-17 08:23:12.726303 | instance | TASK [Ensure the endpoints file exists] **************************************** 2026-04-17 08:23:12.726640 | instance | Friday 17 April 2026 08:23:12 +0000 (0:00:00.683) 0:00:08.607 ********** 2026-04-17 08:23:12.918832 | instance | changed: [localhost] 2026-04-17 08:23:12.919042 | instance | 2026-04-17 08:23:12.919327 | instance | TASK [Load the current endpoints into a variable] ****************************** 2026-04-17 08:23:12.919656 | instance | Friday 17 April 2026 08:23:12 +0000 (0:00:00.193) 0:00:08.801 ********** 2026-04-17 08:23:12.949933 | instance | ok: [localhost] 2026-04-17 08:23:12.950177 | instance | 2026-04-17 08:23:12.950451 | instance | TASK [Generate endpoint skeleton for missing variables] ************************ 2026-04-17 08:23:12.950788 | instance | Friday 17 April 2026 08:23:12 +0000 (0:00:00.031) 0:00:08.832 ********** 2026-04-17 08:23:13.743326 | instance | ok: [localhost] => (item=keycloak_host) 2026-04-17 08:23:13.743456 | instance | ok: [localhost] => (item=kube_prometheus_stack_grafana_host) 2026-04-17 08:23:13.743749 | instance | ok: [localhost] => (item=kube_prometheus_stack_alertmanager_host) 2026-04-17 08:23:13.744014 | instance | ok: [localhost] => (item=kube_prometheus_stack_prometheus_host) 2026-04-17 08:23:13.744281 | instance | ok: [localhost] => (item=openstack_helm_endpoints_region_name) 2026-04-17 08:23:13.744582 | instance | ok: [localhost] => (item=openstack_helm_endpoints_keystone_api_host) 2026-04-17 08:23:13.744846 | instance | ok: [localhost] => (item=openstack_helm_endpoints_glance_api_host) 2026-04-17 08:23:13.745108 | instance | ok: [localhost] => (item=openstack_helm_endpoints_cinder_api_host) 2026-04-17 08:23:13.745313 | instance | ok: [localhost] => (item=openstack_helm_endpoints_placement_api_host) 2026-04-17 08:23:13.745613 | instance | ok: [localhost] => (item=openstack_helm_endpoints_barbican_api_host) 2026-04-17 08:23:13.745849 | instance | ok: [localhost] => (item=openstack_helm_endpoints_neutron_api_host) 2026-04-17 08:23:13.746109 | instance | ok: [localhost] => (item=openstack_helm_endpoints_nova_api_host) 2026-04-17 08:23:13.746370 | instance | ok: [localhost] => (item=openstack_helm_endpoints_nova_novnc_host) 2026-04-17 08:23:13.746729 | instance | ok: [localhost] => (item=openstack_helm_endpoints_ironic_api_host) 2026-04-17 08:23:13.746994 | instance | ok: [localhost] => (item=openstack_helm_endpoints_designate_api_host) 2026-04-17 08:23:13.747259 | instance | ok: [localhost] => (item=openstack_helm_endpoints_octavia_api_host) 2026-04-17 08:23:13.747520 | instance | ok: [localhost] => (item=openstack_helm_endpoints_magnum_api_host) 2026-04-17 08:23:13.747739 | instance | ok: [localhost] => (item=openstack_helm_endpoints_magnum_registry_host) 2026-04-17 08:23:13.748003 | instance | ok: [localhost] => (item=openstack_helm_endpoints_heat_api_host) 2026-04-17 08:23:13.748266 | instance | ok: [localhost] => (item=openstack_helm_endpoints_heat_cfn_api_host) 2026-04-17 08:23:13.748611 | instance | ok: [localhost] => (item=openstack_helm_endpoints_horizon_api_host) 2026-04-17 08:23:13.748874 | instance | ok: [localhost] => (item=openstack_helm_endpoints_rgw_host) 2026-04-17 08:23:13.749138 | instance | ok: [localhost] => (item=openstack_helm_endpoints_manila_api_host) 2026-04-17 08:23:13.749388 | instance | 2026-04-17 08:23:13.749651 | instance | TASK [Write new endpoints file to disk] **************************************** 2026-04-17 08:23:13.749921 | instance | Friday 17 April 2026 08:23:13 +0000 (0:00:00.792) 0:00:09.625 ********** 2026-04-17 08:23:14.132662 | instance | changed: [localhost] 2026-04-17 08:23:14.132872 | instance | 2026-04-17 08:23:14.133144 | instance | TASK [Ensure the endpoints file exists] **************************************** 2026-04-17 08:23:14.133427 | instance | Friday 17 April 2026 08:23:14 +0000 (0:00:00.389) 0:00:10.014 ********** 2026-04-17 08:23:14.318847 | instance | changed: [localhost] 2026-04-17 08:23:14.318944 | instance | 2026-04-17 08:23:14.318958 | instance | PLAY [Generate Neutron configuration for workspace] **************************** 2026-04-17 08:23:14.318968 | instance | 2026-04-17 08:23:14.318976 | instance | TASK [Ensure the Neutron configuration file exists] **************************** 2026-04-17 08:23:14.318986 | instance | Friday 17 April 2026 08:23:14 +0000 (0:00:00.184) 0:00:10.199 ********** 2026-04-17 08:23:14.502077 | instance | changed: [localhost] 2026-04-17 08:23:14.502215 | instance | 2026-04-17 08:23:14.502463 | instance | TASK [Load the current Neutron configuration into a variable] ****************** 2026-04-17 08:23:14.502869 | instance | Friday 17 April 2026 08:23:14 +0000 (0:00:00.184) 0:00:10.383 ********** 2026-04-17 08:23:14.530604 | instance | ok: [localhost] 2026-04-17 08:23:14.530852 | instance | 2026-04-17 08:23:14.531261 | instance | TASK [Generate Neutron values for missing variables] *************************** 2026-04-17 08:23:14.531522 | instance | Friday 17 April 2026 08:23:14 +0000 (0:00:00.029) 0:00:10.413 ********** 2026-04-17 08:23:14.577713 | instance | ok: [localhost] => (item={'key': 'neutron_networks', 'value': [{'name': 'public', 'external': True, 'shared': True, 'mtu_size': 1500, 'port_security_enabled': True, 'provider_network_type': 'flat', 'provider_physical_network': 'external', 'subnets': [{'name': 'public-subnet', 'cidr': '10.96.250.0/24', 'gateway_ip': '10.96.250.10', 'allocation_pool_start': '10.96.250.200', 'allocation_pool_end': '10.96.250.220', 'enable_dhcp': True}]}]}) 2026-04-17 08:23:14.577919 | instance | 2026-04-17 08:23:14.578201 | instance | TASK [Write new Neutron configuration file to disk] **************************** 2026-04-17 08:23:14.578464 | instance | Friday 17 April 2026 08:23:14 +0000 (0:00:00.046) 0:00:10.459 ********** 2026-04-17 08:23:14.979758 | instance | changed: [localhost] 2026-04-17 08:23:14.980018 | instance | 2026-04-17 08:23:14.980326 | instance | PLAY [Generate Nova configuration for workspace] ******************************* 2026-04-17 08:23:14.980598 | instance | 2026-04-17 08:23:14.980895 | instance | TASK [Ensure the Nova configuration file exists] ******************************* 2026-04-17 08:23:14.981265 | instance | Friday 17 April 2026 08:23:14 +0000 (0:00:00.401) 0:00:10.861 ********** 2026-04-17 08:23:15.182555 | instance | changed: [localhost] 2026-04-17 08:23:15.182778 | instance | 2026-04-17 08:23:15.183061 | instance | TASK [Load the current Nova configuration into a variable] ********************* 2026-04-17 08:23:15.183337 | instance | Friday 17 April 2026 08:23:15 +0000 (0:00:00.202) 0:00:11.064 ********** 2026-04-17 08:23:15.213418 | instance | ok: [localhost] 2026-04-17 08:23:15.213659 | instance | 2026-04-17 08:23:15.213954 | instance | TASK [Generate Nova values for missing variables] ****************************** 2026-04-17 08:23:15.214199 | instance | Friday 17 April 2026 08:23:15 +0000 (0:00:00.031) 0:00:11.095 ********** 2026-04-17 08:23:15.259396 | instance | ok: [localhost] => (item={'key': 'nova_flavors', 'value': [{'name': 'm1.tiny', 'ram': 512, 'disk': 1, 'vcpus': 1}, {'name': 'm1.small', 'ram': 2048, 'disk': 20, 'vcpus': 1}, {'name': 'm1.medium', 'ram': 4096, 'disk': 40, 'vcpus': 2}, {'name': 'm1.large', 'ram': 8192, 'disk': 80, 'vcpus': 4}, {'name': 'm1.xlarge', 'ram': 16384, 'disk': 160, 'vcpus': 8}]}) 2026-04-17 08:23:15.259617 | instance | 2026-04-17 08:23:15.259888 | instance | TASK [Write new Nova configuration file to disk] ******************************* 2026-04-17 08:23:15.260161 | instance | Friday 17 April 2026 08:23:15 +0000 (0:00:00.046) 0:00:11.141 ********** 2026-04-17 08:23:15.626040 | instance | changed: [localhost] 2026-04-17 08:23:15.626140 | instance | 2026-04-17 08:23:15.626175 | instance | PLAY [Generate secrets for workspace] ****************************************** 2026-04-17 08:23:15.626608 | instance | 2026-04-17 08:23:15.626956 | instance | TASK [Ensure the secrets file exists] ****************************************** 2026-04-17 08:23:15.627262 | instance | Friday 17 April 2026 08:23:15 +0000 (0:00:00.365) 0:00:11.507 ********** 2026-04-17 08:23:15.804489 | instance | changed: [localhost] 2026-04-17 08:23:15.804780 | instance | 2026-04-17 08:23:15.805107 | instance | TASK [Load the current secrets into a variable] ******************************** 2026-04-17 08:23:15.805413 | instance | Friday 17 April 2026 08:23:15 +0000 (0:00:00.178) 0:00:11.686 ********** 2026-04-17 08:23:15.841995 | instance | ok: [localhost] 2026-04-17 08:23:15.842282 | instance | 2026-04-17 08:23:15.842661 | instance | TASK [Generate secrets for missing variables] ********************************** 2026-04-17 08:23:15.842987 | instance | Friday 17 April 2026 08:23:15 +0000 (0:00:00.030) 0:00:11.717 ********** 2026-04-17 08:23:16.252704 | instance | ok: [localhost] => (item=heat_auth_encryption_key) 2026-04-17 08:23:16.252960 | instance | ok: [localhost] => (item=keepalived_password) 2026-04-17 08:23:16.253152 | instance | ok: [localhost] => (item=keycloak_admin_password) 2026-04-17 08:23:16.253339 | instance | ok: [localhost] => (item=keycloak_database_password) 2026-04-17 08:23:16.253522 | instance | ok: [localhost] => (item=keystone_keycloak_client_secret) 2026-04-17 08:23:16.253737 | instance | ok: [localhost] => (item=keystone_oidc_crypto_passphrase) 2026-04-17 08:23:16.253926 | instance | ok: [localhost] => (item=kube_prometheus_stack_grafana_admin_password) 2026-04-17 08:23:16.254112 | instance | ok: [localhost] => (item=octavia_heartbeat_key) 2026-04-17 08:23:16.254296 | instance | ok: [localhost] => (item=openstack_helm_endpoints_rabbitmq_admin_password) 2026-04-17 08:23:16.254478 | instance | ok: [localhost] => (item=openstack_helm_endpoints_memcached_secret_key) 2026-04-17 08:23:16.254709 | instance | ok: [localhost] => (item=openstack_helm_endpoints_keystone_admin_password) 2026-04-17 08:23:16.254902 | instance | ok: [localhost] => (item=openstack_helm_endpoints_keystone_mariadb_password) 2026-04-17 08:23:16.255085 | instance | ok: [localhost] => (item=openstack_helm_endpoints_keystone_rabbitmq_password) 2026-04-17 08:23:16.255267 | instance | ok: [localhost] => (item=openstack_helm_endpoints_glance_keystone_password) 2026-04-17 08:23:16.255449 | instance | ok: [localhost] => (item=openstack_helm_endpoints_glance_mariadb_password) 2026-04-17 08:23:16.255633 | instance | ok: [localhost] => (item=openstack_helm_endpoints_glance_rabbitmq_password) 2026-04-17 08:23:16.255818 | instance | ok: [localhost] => (item=openstack_helm_endpoints_cinder_keystone_password) 2026-04-17 08:23:16.256001 | instance | ok: [localhost] => (item=openstack_helm_endpoints_cinder_mariadb_password) 2026-04-17 08:23:16.256183 | instance | ok: [localhost] => (item=openstack_helm_endpoints_cinder_rabbitmq_password) 2026-04-17 08:23:16.256364 | instance | ok: [localhost] => (item=openstack_helm_endpoints_placement_keystone_password) 2026-04-17 08:23:16.256549 | instance | ok: [localhost] => (item=openstack_helm_endpoints_placement_mariadb_password) 2026-04-17 08:23:16.256732 | instance | ok: [localhost] => (item=openstack_helm_endpoints_barbican_keystone_password) 2026-04-17 08:23:16.256914 | instance | ok: [localhost] => (item=openstack_helm_endpoints_barbican_mariadb_password) 2026-04-17 08:23:16.257105 | instance | ok: [localhost] => (item=openstack_helm_endpoints_neutron_keystone_password) 2026-04-17 08:23:16.257288 | instance | ok: [localhost] => (item=openstack_helm_endpoints_neutron_mariadb_password) 2026-04-17 08:23:16.257473 | instance | ok: [localhost] => (item=openstack_helm_endpoints_neutron_rabbitmq_password) 2026-04-17 08:23:16.257656 | instance | ok: [localhost] => (item=openstack_helm_endpoints_neutron_metadata_secret) 2026-04-17 08:23:16.257869 | instance | ok: [localhost] => (item=openstack_helm_endpoints_nova_keystone_password) 2026-04-17 08:23:16.258053 | instance | ok: [localhost] => (item=openstack_helm_endpoints_nova_mariadb_password) 2026-04-17 08:23:16.258239 | instance | ok: [localhost] => (item=openstack_helm_endpoints_nova_rabbitmq_password) 2026-04-17 08:23:16.258421 | instance | ok: [localhost] => (item=openstack_helm_endpoints_ironic_keystone_password) 2026-04-17 08:23:16.258635 | instance | ok: [localhost] => (item=openstack_helm_endpoints_ironic_mariadb_password) 2026-04-17 08:23:16.258822 | instance | ok: [localhost] => (item=openstack_helm_endpoints_ironic_rabbitmq_password) 2026-04-17 08:23:16.259004 | instance | ok: [localhost] => (item=openstack_helm_endpoints_designate_keystone_password) 2026-04-17 08:23:16.259189 | instance | ok: [localhost] => (item=openstack_helm_endpoints_designate_mariadb_password) 2026-04-17 08:23:16.259371 | instance | ok: [localhost] => (item=openstack_helm_endpoints_designate_rabbitmq_password) 2026-04-17 08:23:16.259553 | instance | ok: [localhost] => (item=openstack_helm_endpoints_octavia_keystone_password) 2026-04-17 08:23:16.259735 | instance | ok: [localhost] => (item=openstack_helm_endpoints_octavia_mariadb_password) 2026-04-17 08:23:16.259916 | instance | ok: [localhost] => (item=openstack_helm_endpoints_octavia_rabbitmq_password) 2026-04-17 08:23:16.260106 | instance | ok: [localhost] => (item=openstack_helm_endpoints_magnum_keystone_password) 2026-04-17 08:23:16.260289 | instance | ok: [localhost] => (item=openstack_helm_endpoints_magnum_mariadb_password) 2026-04-17 08:23:16.260471 | instance | ok: [localhost] => (item=openstack_helm_endpoints_magnum_rabbitmq_password) 2026-04-17 08:23:16.260654 | instance | ok: [localhost] => (item=openstack_helm_endpoints_heat_keystone_password) 2026-04-17 08:23:16.260847 | instance | ok: [localhost] => (item=openstack_helm_endpoints_heat_trustee_keystone_password) 2026-04-17 08:23:16.261044 | instance | ok: [localhost] => (item=openstack_helm_endpoints_heat_stack_user_keystone_password) 2026-04-17 08:23:16.261227 | instance | ok: [localhost] => (item=openstack_helm_endpoints_heat_mariadb_password) 2026-04-17 08:23:16.261408 | instance | ok: [localhost] => (item=openstack_helm_endpoints_heat_rabbitmq_password) 2026-04-17 08:23:16.261589 | instance | ok: [localhost] => (item=openstack_helm_endpoints_horizon_mariadb_password) 2026-04-17 08:23:16.261816 | instance | ok: [localhost] => (item=openstack_helm_endpoints_tempest_keystone_password) 2026-04-17 08:23:16.262015 | instance | ok: [localhost] => (item=openstack_helm_endpoints_openstack_exporter_keystone_password) 2026-04-17 08:23:16.262198 | instance | ok: [localhost] => (item=openstack_helm_endpoints_rgw_keystone_password) 2026-04-17 08:23:16.262475 | instance | ok: [localhost] => (item=openstack_helm_endpoints_manila_keystone_password) 2026-04-17 08:23:16.262741 | instance | ok: [localhost] => (item=openstack_helm_endpoints_manila_mariadb_password) 2026-04-17 08:23:16.262936 | instance | ok: [localhost] => (item=openstack_helm_endpoints_staffeln_mariadb_password) 2026-04-17 08:23:16.263109 | instance | 2026-04-17 08:23:16.263292 | instance | TASK [Generate base64 encoded secrets] ***************************************** 2026-04-17 08:23:16.263477 | instance | Friday 17 April 2026 08:23:16 +0000 (0:00:00.417) 0:00:12.135 ********** 2026-04-17 08:23:16.313425 | instance | ok: [localhost] => (item=barbican_kek) 2026-04-17 08:23:16.313613 | instance | 2026-04-17 08:23:16.313844 | instance | TASK [Generate temporary files for generating keys for missing variables] ****** 2026-04-17 08:23:16.314034 | instance | Friday 17 April 2026 08:23:16 +0000 (0:00:00.060) 0:00:12.196 ********** 2026-04-17 08:23:16.727748 | instance | changed: [localhost] => (item=manila_ssh_key) 2026-04-17 08:23:16.728215 | instance | changed: [localhost] => (item=nova_ssh_key) 2026-04-17 08:23:16.728605 | instance | 2026-04-17 08:23:16.728975 | instance | TASK [Generate SSH keys for missing variables] ********************************* 2026-04-17 08:23:16.729354 | instance | Friday 17 April 2026 08:23:16 +0000 (0:00:00.413) 0:00:12.609 ********** 2026-04-17 08:23:19.869531 | instance | changed: [localhost] => (item=manila_ssh_key) 2026-04-17 08:23:19.869608 | instance | changed: [localhost] => (item=nova_ssh_key) 2026-04-17 08:23:19.869620 | instance | 2026-04-17 08:23:19.869639 | instance | TASK [Set values for SSH keys] ************************************************* 2026-04-17 08:23:19.869650 | instance | Friday 17 April 2026 08:23:19 +0000 (0:00:03.140) 0:00:15.750 ********** 2026-04-17 08:23:19.926688 | instance | ok: [localhost] => (item=manila_ssh_key) 2026-04-17 08:23:19.926846 | instance | ok: [localhost] => (item=nova_ssh_key) 2026-04-17 08:23:19.926858 | instance | 2026-04-17 08:23:19.926877 | instance | TASK [Delete the temporary files generated for SSH keys] *********************** 2026-04-17 08:23:19.927022 | instance | Friday 17 April 2026 08:23:19 +0000 (0:00:00.058) 0:00:15.808 ********** 2026-04-17 08:23:20.270107 | instance | changed: [localhost] => (item=manila_ssh_key) 2026-04-17 08:23:20.270204 | instance | changed: [localhost] => (item=nova_ssh_key) 2026-04-17 08:23:20.270266 | instance | 2026-04-17 08:23:20.270391 | instance | TASK [Write new secrets file to disk] ****************************************** 2026-04-17 08:23:20.270518 | instance | Friday 17 April 2026 08:23:20 +0000 (0:00:00.343) 0:00:16.152 ********** 2026-04-17 08:23:20.653167 | instance | changed: [localhost] 2026-04-17 08:23:20.653289 | instance | 2026-04-17 08:23:20.653304 | instance | TASK [Encrypt secrets file with Vault password] ******************************** 2026-04-17 08:23:20.653449 | instance | Friday 17 April 2026 08:23:20 +0000 (0:00:00.382) 0:00:16.535 ********** 2026-04-17 08:23:20.694200 | instance | skipping: [localhost] 2026-04-17 08:23:20.694482 | instance | 2026-04-17 08:23:20.694913 | instance | PLAY [Setup networking] ******************************************************** 2026-04-17 08:23:20.695170 | instance | 2026-04-17 08:23:20.695444 | instance | TASK [Gathering Facts] ********************************************************* 2026-04-17 08:23:20.695745 | instance | Friday 17 April 2026 08:23:20 +0000 (0:00:00.040) 0:00:16.576 ********** 2026-04-17 08:23:21.418289 | instance | ok: [instance] 2026-04-17 08:23:21.418347 | instance | 2026-04-17 08:23:21.418361 | instance | TASK [Create bridge for management network] ************************************ 2026-04-17 08:23:21.418379 | instance | Friday 17 April 2026 08:23:21 +0000 (0:00:00.723) 0:00:17.299 ********** 2026-04-17 08:23:21.768304 | instance | ok: [instance] 2026-04-17 08:23:21.768547 | instance | 2026-04-17 08:23:21.769260 | instance | TASK [Create fake interface for management bridge] ***************************** 2026-04-17 08:23:21.769421 | instance | Friday 17 April 2026 08:23:21 +0000 (0:00:00.350) 0:00:17.650 ********** 2026-04-17 08:23:21.992038 | instance | ok: [instance] 2026-04-17 08:23:21.992334 | instance | 2026-04-17 08:23:21.992613 | instance | TASK [Assign dummy interface to management bridge] ***************************** 2026-04-17 08:23:21.992855 | instance | Friday 17 April 2026 08:23:21 +0000 (0:00:00.224) 0:00:17.874 ********** 2026-04-17 08:23:22.191147 | instance | ok: [instance] 2026-04-17 08:23:22.191255 | instance | 2026-04-17 08:23:22.191494 | instance | TASK [Assign IP address for management bridge] ********************************* 2026-04-17 08:23:22.191779 | instance | Friday 17 April 2026 08:23:22 +0000 (0:00:00.199) 0:00:18.073 ********** 2026-04-17 08:23:22.403245 | instance | ok: [instance] 2026-04-17 08:23:22.403448 | instance | 2026-04-17 08:23:22.403622 | instance | TASK [Bring up interfaces] ***************************************************** 2026-04-17 08:23:22.403780 | instance | Friday 17 April 2026 08:23:22 +0000 (0:00:00.209) 0:00:18.282 ********** 2026-04-17 08:23:22.786712 | instance | ok: [instance] => (item=br-mgmt) 2026-04-17 08:23:22.787025 | instance | ok: [instance] => (item=dummy0) 2026-04-17 08:23:22.787038 | instance | 2026-04-17 08:23:22.787056 | instance | PLAY [Create devices for Ceph] ************************************************* 2026-04-17 08:23:22.787067 | instance | 2026-04-17 08:23:22.787198 | instance | TASK [Gathering Facts] ********************************************************* 2026-04-17 08:23:22.787500 | instance | Friday 17 April 2026 08:23:22 +0000 (0:00:00.385) 0:00:18.668 ********** 2026-04-17 08:23:23.547929 | instance | ok: [instance] 2026-04-17 08:23:23.548141 | instance | 2026-04-17 08:23:23.548447 | instance | TASK [Install depedencies] ***************************************************** 2026-04-17 08:23:23.548702 | instance | Friday 17 April 2026 08:23:23 +0000 (0:00:00.761) 0:00:19.429 ********** 2026-04-17 08:24:14.065146 | instance | changed: [instance] 2026-04-17 08:24:14.065245 | instance | 2026-04-17 08:24:14.065258 | instance | TASK [Start up service] ******************************************************** 2026-04-17 08:24:14.065268 | instance | Friday 17 April 2026 08:24:14 +0000 (0:00:50.516) 0:01:09.946 ********** 2026-04-17 08:24:14.621190 | instance | ok: [instance] 2026-04-17 08:24:14.621278 | instance | 2026-04-17 08:24:14.621597 | instance | TASK [Generate lvm.conf] ******************************************************* 2026-04-17 08:24:14.621653 | instance | Friday 17 April 2026 08:24:14 +0000 (0:00:00.557) 0:01:10.503 ********** 2026-04-17 08:24:14.853139 | instance | ok: [instance] 2026-04-17 08:24:14.853225 | instance | 2026-04-17 08:24:14.853612 | instance | TASK [Write /etc/lvm/lvm.conf] ************************************************* 2026-04-17 08:24:14.853680 | instance | Friday 17 April 2026 08:24:14 +0000 (0:00:00.231) 0:01:10.735 ********** 2026-04-17 08:24:15.313813 | instance | changed: [instance] 2026-04-17 08:24:15.313914 | instance | 2026-04-17 08:24:15.314273 | instance | TASK [Get list of all loopback devices] **************************************** 2026-04-17 08:24:15.314508 | instance | Friday 17 April 2026 08:24:15 +0000 (0:00:00.460) 0:01:11.196 ********** 2026-04-17 08:24:15.515143 | instance | ok: [instance] 2026-04-17 08:24:15.515230 | instance | 2026-04-17 08:24:15.515638 | instance | TASK [Fail if there is any existing loopback devices] ************************** 2026-04-17 08:24:15.515688 | instance | Friday 17 April 2026 08:24:15 +0000 (0:00:00.201) 0:01:11.397 ********** 2026-04-17 08:24:15.546125 | instance | skipping: [instance] 2026-04-17 08:24:15.546340 | instance | 2026-04-17 08:24:15.546531 | instance | TASK [Create devices for Ceph] ************************************************* 2026-04-17 08:24:15.546716 | instance | Friday 17 April 2026 08:24:15 +0000 (0:00:00.030) 0:01:11.427 ********** 2026-04-17 08:24:16.103058 | instance | changed: [instance] => (item=osd0) 2026-04-17 08:24:16.103339 | instance | changed: [instance] => (item=osd1) 2026-04-17 08:24:16.103573 | instance | changed: [instance] => (item=osd2) 2026-04-17 08:24:16.103778 | instance | 2026-04-17 08:24:16.103992 | instance | TASK [Set permissions on loopback devices] ************************************* 2026-04-17 08:24:16.104205 | instance | Friday 17 April 2026 08:24:16 +0000 (0:00:00.555) 0:01:11.982 ********** 2026-04-17 08:24:16.648749 | instance | changed: [instance] => (item=osd0) 2026-04-17 08:24:16.648827 | instance | changed: [instance] => (item=osd1) 2026-04-17 08:24:16.648838 | instance | changed: [instance] => (item=osd2) 2026-04-17 08:24:16.648848 | instance | 2026-04-17 08:24:16.648858 | instance | TASK [Start loop devices] ****************************************************** 2026-04-17 08:24:16.648941 | instance | Friday 17 April 2026 08:24:16 +0000 (0:00:00.546) 0:01:12.529 ********** 2026-04-17 08:24:17.418854 | instance | changed: [instance] => (item=osd0) 2026-04-17 08:24:17.418948 | instance | changed: [instance] => (item=osd1) 2026-04-17 08:24:17.418963 | instance | changed: [instance] => (item=osd2) 2026-04-17 08:24:17.418977 | instance | 2026-04-17 08:24:17.418991 | instance | TASK [Create a volume group for each loop device] ****************************** 2026-04-17 08:24:17.419005 | instance | Friday 17 April 2026 08:24:17 +0000 (0:00:00.769) 0:01:13.298 ********** 2026-04-17 08:24:20.566525 | instance | changed: [instance] => (item=osd0) 2026-04-17 08:24:20.567046 | instance | changed: [instance] => (item=osd1) 2026-04-17 08:24:20.567398 | instance | changed: [instance] => (item=osd2) 2026-04-17 08:24:20.567718 | instance | 2026-04-17 08:24:20.568118 | instance | TASK [Create a logical volume for each loop device] **************************** 2026-04-17 08:24:20.568469 | instance | Friday 17 April 2026 08:24:20 +0000 (0:00:03.148) 0:01:16.447 ********** 2026-04-17 08:24:22.455550 | instance | changed: [instance] => (item=ceph-instance-osd0) 2026-04-17 08:24:22.456362 | instance | changed: [instance] => (item=ceph-instance-osd1) 2026-04-17 08:24:22.456722 | instance | changed: [instance] => (item=ceph-instance-osd2) 2026-04-17 08:24:22.457080 | instance | 2026-04-17 08:24:22.457426 | instance | PLAY [controllers] ************************************************************* 2026-04-17 08:24:22.457745 | instance | 2026-04-17 08:24:22.458102 | instance | TASK [Gathering Facts] ********************************************************* 2026-04-17 08:24:22.458445 | instance | Friday 17 April 2026 08:24:22 +0000 (0:00:01.889) 0:01:18.336 ********** 2026-04-17 08:24:23.376392 | instance | ok: [instance] 2026-04-17 08:24:23.376479 | instance | 2026-04-17 08:24:23.376601 | instance | TASK [Set masquerade rule] ***************************************************** 2026-04-17 08:24:23.376826 | instance | Friday 17 April 2026 08:24:23 +0000 (0:00:00.920) 0:01:19.257 ********** 2026-04-17 08:24:23.883219 | instance | changed: [instance] 2026-04-17 08:24:23.883422 | instance | 2026-04-17 08:24:23.883579 | instance | PLAY RECAP ********************************************************************* 2026-04-17 08:24:23.883745 | instance | instance : ok=24 changed=10 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 2026-04-17 08:24:23.883902 | instance | localhost : ok=40 changed=21 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 2026-04-17 08:24:23.884041 | instance | 2026-04-17 08:24:23.884211 | instance | Friday 17 April 2026 08:24:23 +0000 (0:00:00.506) 0:01:19.763 ********** 2026-04-17 08:24:23.884359 | instance | =============================================================================== 2026-04-17 08:24:23.884504 | instance | Install depedencies ---------------------------------------------------- 50.52s 2026-04-17 08:24:23.884651 | instance | Create a volume group for each loop device ------------------------------ 3.15s 2026-04-17 08:24:23.884797 | instance | Generate SSH keys for missing variables --------------------------------- 3.14s 2026-04-17 08:24:23.884942 | instance | Create a logical volume for each loop device ---------------------------- 1.89s 2026-04-17 08:24:23.885090 | instance | Install "dirmngr" for GPG keyserver operations -------------------------- 1.25s 2026-04-17 08:24:23.885238 | instance | Gathering Facts --------------------------------------------------------- 1.19s 2026-04-17 08:24:23.885395 | instance | Create folders for workspace -------------------------------------------- 1.08s 2026-04-17 08:24:23.885539 | instance | Gathering Facts --------------------------------------------------------- 0.92s 2026-04-17 08:24:23.885683 | instance | Generate endpoint skeleton for missing variables ------------------------ 0.79s 2026-04-17 08:24:23.885827 | instance | Purge "snapd" package --------------------------------------------------- 0.78s 2026-04-17 08:24:23.885971 | instance | Start loop devices ------------------------------------------------------ 0.77s 2026-04-17 08:24:23.886116 | instance | Gathering Facts --------------------------------------------------------- 0.76s 2026-04-17 08:24:23.886262 | instance | Gathering Facts --------------------------------------------------------- 0.72s 2026-04-17 08:24:23.886421 | instance | Gathering Facts --------------------------------------------------------- 0.68s 2026-04-17 08:24:23.886586 | instance | Configure short hostname ------------------------------------------------ 0.66s 2026-04-17 08:24:23.886735 | instance | Start up service -------------------------------------------------------- 0.56s 2026-04-17 08:24:23.886882 | instance | Create devices for Ceph ------------------------------------------------- 0.56s 2026-04-17 08:24:23.887031 | instance | Write new Ceph control plane configuration file to disk ----------------- 0.55s 2026-04-17 08:24:23.887177 | instance | Set permissions on loopback devices ------------------------------------- 0.55s 2026-04-17 08:24:23.887321 | instance | Set masquerade rule ----------------------------------------------------- 0.51s 2026-04-17 08:24:23.986963 | instance | INFO [aio > prepare] Executed: Successful 2026-04-17 08:24:23.987647 | instance | INFO Molecule executed 1 scenario (1 successful) 2026-04-17 08:24:24.223646 | instance | ok: Runtime: 0:02:12.850737 2026-04-17 08:24:24.229860 | 2026-04-17 08:24:24.230322 | PLAY RECAP 2026-04-17 08:24:24.230383 | instance | ok: 12 changed: 9 unreachable: 0 failed: 0 skipped: 0 rescued: 0 ignored: 0 2026-04-17 08:24:24.230415 | 2026-04-17 08:24:24.361170 | PRE-RUN END RESULT_NORMAL: [untrusted : github.com/vexxhost/atmosphere/test-playbooks/molecule/pre.yml@main] 2026-04-17 08:24:24.372929 | RUN START: [untrusted : github.com/vexxhost/zuul-jobs/playbooks/molecule/run.yaml@main] 2026-04-17 08:24:24.966667 | 2026-04-17 08:24:24.966788 | PLAY [all] 2026-04-17 08:24:24.977664 | 2026-04-17 08:24:24.977736 | TASK [Copy inventory file for Zuul] 2026-04-17 08:24:25.861334 | instance | changed 2026-04-17 08:24:25.868527 | 2026-04-17 08:24:25.868650 | TASK [Switch "ansible_host" to private IP] 2026-04-17 08:24:26.183762 | instance | changed: 1 replacements made 2026-04-17 08:24:26.191108 | 2026-04-17 08:24:26.191213 | TASK [Run Molecule scenario] 2026-04-17 08:24:26.859482 | instance | WARNING Molecule scenarios should migrate to 'extensions/molecule' 2026-04-17 08:24:27.311018 | instance | INFO [aio > discovery] scenario test matrix: dependency, cleanup, destroy, syntax, create, prepare, converge, idempotence, side_effect, verify, cleanup, destroy 2026-04-17 08:24:27.311132 | instance | INFO [aio > prerun] Performing prerun with role_name_check=0... 2026-04-17 08:24:40.248814 | instance | INFO [aio > dependency] Executing 2026-04-17 08:24:40.248990 | instance | WARNING [aio > dependency] Missing roles requirements file: requirements.yml 2026-04-17 08:24:40.249252 | instance | WARNING [aio > dependency] Missing collections requirements file: collections.yml 2026-04-17 08:24:40.249338 | instance | WARNING [aio > dependency] Executed: 2 missing (Remove from test_sequence to suppress) 2026-04-17 08:24:40.257251 | instance | INFO [aio > cleanup] Executing 2026-04-17 08:24:40.257581 | instance | WARNING [aio > cleanup] Executed: Missing playbook (Remove from test_sequence to suppress) 2026-04-17 08:24:40.265483 | instance | INFO [aio > destroy] Executing 2026-04-17 08:24:40.265516 | instance | WARNING [aio > destroy] Skipping, '--destroy=never' requested. 2026-04-17 08:24:40.265616 | instance | INFO [aio > destroy] Executed: Successful 2026-04-17 08:24:40.273116 | instance | INFO [aio > syntax] Executing 2026-04-17 08:24:41.049618 | instance | 2026-04-17 08:24:41.049932 | instance | playbook: /home/zuul/src/github.com/vexxhost/atmosphere/molecule/aio/converge.yml 2026-04-17 08:24:41.100710 | instance | INFO [aio > syntax] Executed: Successful 2026-04-17 08:24:41.114386 | instance | INFO [aio > create] Executing 2026-04-17 08:24:41.116729 | instance | WARNING [aio > create] Executed: Missing playbook (Remove from test_sequence to suppress) 2026-04-17 08:24:41.125085 | instance | INFO [aio > prepare] Executing 2026-04-17 08:24:41.125153 | instance | WARNING [aio > prepare] Skipping, instances already prepared. 2026-04-17 08:24:41.125233 | instance | INFO [aio > prepare] Executed: Successful 2026-04-17 08:24:41.132817 | instance | INFO [aio > converge] Executing 2026-04-17 08:24:41.864434 | instance | 2026-04-17 08:24:41.864671 | instance | PLAY [all] ********************************************************************* 2026-04-17 08:24:41.864914 | instance | 2026-04-17 08:24:41.865195 | instance | TASK [Build atmosphere binary] ************************************************* 2026-04-17 08:24:41.865470 | instance | Friday 17 April 2026 08:24:41 +0000 (0:00:00.010) 0:00:00.010 ********** 2026-04-17 08:24:49.266355 | instance | [WARNING]: Platform linux on host instance is using the discovered Python 2026-04-17 08:24:49.266669 | instance | interpreter at /usr/bin/python3.10, but future installation of another Python 2026-04-17 08:24:49.266914 | instance | interpreter could change the meaning of that path. See 2026-04-17 08:24:49.267189 | instance | https://docs.ansible.com/ansible- 2026-04-17 08:24:49.267516 | instance | core/2.17/reference_appendices/interpreter_discovery.html for more information. 2026-04-17 08:24:49.271001 | instance | ok: [instance] 2026-04-17 08:24:49.271252 | instance | 2026-04-17 08:24:49.271640 | instance | TASK [Deploy with parallel orchestrator] *************************************** 2026-04-17 08:24:49.271833 | instance | Friday 17 April 2026 08:24:49 +0000 (0:00:07.406) 0:00:07.416 ********** 2026-04-17 08:25:51.336590 | instance | fatal: [instance]: FAILED! => {"changed": false, "cmd": ". .venv/bin/activate && ./bin/atmosphere deploy --inventory ./inventory.yaml \n", "delta": "0:01:01.872261", "end": "2026-04-17 08:25:51.301895", "msg": "non-zero return code", "rc": 1, "start": "2026-04-17 08:24:49.429634", "stderr": "Error: component ceph failed: ansible-playbook failed for ceph/ceph: exit status 2\nUsage:\n atmosphere deploy [flags]\n\nFlags:\n --concurrency int Max concurrent deployments per wave (0 = unlimited)\n -h, --help help for deploy\n -i, --inventory string Path to Ansible inventory file (required)\n -t, --tags string Comma-separated list of component tags to deploy\n\ncomponent ceph failed: ansible-playbook failed for ceph/ceph: exit status 2", "stderr_lines": ["Error: component ceph failed: ansible-playbook failed for ceph/ceph: exit status 2", "Usage:", " atmosphere deploy [flags]", "", "Flags:", " --concurrency int Max concurrent deployments per wave (0 = unlimited)", " -h, --help help for deploy", " -i, --inventory string Path to Ansible inventory file (required)", " -t, --tags string Comma-separated list of component tags to deploy", "", "component ceph failed: ansible-playbook failed for ceph/ceph: exit status 2"], "stdout": "==> Running preflight checks\n[preflight] \n[preflight] PLAY [Preflight checks] ********************************************************\n[preflight] \n[preflight] TASK [Fail if atmosphere_ceph_enabled is set] **********************************\n[preflight] skipping: [instance]\n[preflight] \n[preflight] PLAY RECAP *********************************************************************\n[preflight] instance : ok=0 changed=0 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 \n[preflight] \n==> Preflight checks passed\n==> Starting parallel deployment\n==> [ceph] Starting deployment\n==> [udev] Starting deployment\n==> [lpfc] Starting deployment\n==> [iscsi] Starting deployment\n==> [multipathd] Starting deployment\n==> [kubernetes] Starting deployment\n[multipathd/multipathd] \n[multipathd/multipathd] PLAY [controllers:computes] ****************************************************\n[multipathd/multipathd] \n[multipathd/multipathd] TASK [Gathering Facts] *********************************************************\n[iscsi/iscsi] \n[iscsi/iscsi] PLAY [controllers:computes] ****************************************************\n[iscsi/iscsi] \n[iscsi/iscsi] TASK [Gathering Facts] *********************************************************\n[lpfc/lpfc] \n[lpfc/lpfc] PLAY [controllers:computes] ****************************************************\n[lpfc/lpfc] \n[lpfc/lpfc] TASK [Gathering Facts] *********************************************************\n[udev/udev] \n[udev/udev] PLAY [controllers:computes] ****************************************************\n[udev/udev] \n[udev/udev] TASK [Gathering Facts] *********************************************************\n[ceph/ceph] \n[ceph/ceph] PLAY [all] *********************************************************************\n[ceph/ceph] \n[ceph/ceph] TASK [Gathering Facts] *********************************************************\n[multipathd/multipathd] [WARNING]: Platform linux on host instance is using the discovered Python\n[multipathd/multipathd] interpreter at /usr/bin/python3.10, but future installation of another Python\n[multipathd/multipathd] interpreter could change the meaning of that path. See\n[multipathd/multipathd] https://docs.ansible.com/ansible-\n[multipathd/multipathd] core/2.17/reference_appendices/interpreter_discovery.html for more information.\n[multipathd/multipathd] ok: [instance]\n[multipathd/multipathd] \n[multipathd/multipathd] TASK [vexxhost.atmosphere.multipathd : Add backports PPA] **********************\n[udev/udev] [WARNING]: Platform linux on host instance is using the discovered Python\n[udev/udev] interpreter at /usr/bin/python3.10, but future installation of another Python\n[udev/udev] interpreter could change the meaning of that path. See\n[udev/udev] https://docs.ansible.com/ansible-\n[udev/udev] core/2.17/reference_appendices/interpreter_discovery.html for more information.\n[iscsi/iscsi] [WARNING]: Platform linux on host instance is using the discovered Python\n[iscsi/iscsi] interpreter at /usr/bin/python3.10, but future installation of another Python\n[iscsi/iscsi] interpreter could change the meaning of that path. See\n[iscsi/iscsi] https://docs.ansible.com/ansible-\n[iscsi/iscsi] core/2.17/reference_appendices/interpreter_discovery.html for more information.\n[iscsi/iscsi] ok: [instance]\n[iscsi/iscsi] \n[iscsi/iscsi] TASK [vexxhost.atmosphere.iscsi : Install iscsi package] ***********************\n[lpfc/lpfc] [WARNING]: Platform linux on host instance is using the discovered Python\n[lpfc/lpfc] interpreter at /usr/bin/python3.10, but future installation of another Python\n[lpfc/lpfc] interpreter could change the meaning of that path. See\n[lpfc/lpfc] https://docs.ansible.com/ansible-\n[lpfc/lpfc] core/2.17/reference_appendices/interpreter_discovery.html for more information.\n[lpfc/lpfc] ok: [instance]\n[lpfc/lpfc] \n[lpfc/lpfc] TASK [vexxhost.atmosphere.lpfc : Detect if the \"lpfc\" module is loaded] ********\n[udev/udev] ok: [instance]\n[udev/udev] \n[udev/udev] TASK [vexxhost.atmosphere.udev : Add udev rules for Pure Storage FlashArray] ***\n[ceph/ceph] [WARNING]: Platform linux on host instance is using the discovered Python\n[ceph/ceph] interpreter at /usr/bin/python3.10, but future installation of another Python\n[ceph/ceph] interpreter could change the meaning of that path. See\n[ceph/ceph] https://docs.ansible.com/ansible-\n[ceph/ceph] core/2.17/reference_appendices/interpreter_discovery.html for more information.\n[ceph/ceph] ok: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [Fail if atmosphere_ceph_enabled is set] **********************************\n[ceph/ceph] skipping: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [Set a fact with the \"atmosphere_images\" for other plays] *****************\n[ceph/ceph] ok: [instance]\n[ceph/ceph] \n[ceph/ceph] PLAY [Deploy Ceph monitors & managers] *****************************************\n[ceph/ceph] \n[ceph/ceph] TASK [Gathering Facts] *********************************************************\n[lpfc/lpfc] ok: [instance]\n[lpfc/lpfc] \n[lpfc/lpfc] TASK [vexxhost.atmosphere.lpfc : Install the configuration file] ***************\n[lpfc/lpfc] skipping: [instance]\n[lpfc/lpfc] \n[lpfc/lpfc] TASK [vexxhost.atmosphere.lpfc : Get the values for the module parameters] *****\n[lpfc/lpfc] skipping: [instance] => (item=lpfc_lun_queue_depth) \n[lpfc/lpfc] skipping: [instance] => (item=lpfc_sg_seg_cnt) \n[lpfc/lpfc] skipping: [instance] => (item=lpfc_max_luns) \n[lpfc/lpfc] skipping: [instance] => (item=lpfc_enable_fc4_type) \n[lpfc/lpfc] skipping: [instance]\n[lpfc/lpfc] \n[lpfc/lpfc] TASK [vexxhost.atmosphere.lpfc : Detect if the run-time module parameters are set correctly] ***\n[lpfc/lpfc] skipping: [instance] => (item=lpfc_lun_queue_depth) \n[lpfc/lpfc] skipping: [instance] => (item=lpfc_sg_seg_cnt) \n[lpfc/lpfc] skipping: [instance] => (item=lpfc_max_luns) \n[lpfc/lpfc] skipping: [instance] => (item=lpfc_enable_fc4_type) \n[lpfc/lpfc] skipping: [instance]\n[lpfc/lpfc] \n[lpfc/lpfc] TASK [vexxhost.atmosphere.lpfc : Update \"initramfs\" if the configuration file has changed] ***\n[lpfc/lpfc] skipping: [instance]\n[lpfc/lpfc] \n[lpfc/lpfc] TASK [Reboot the system if the configuration file has changed] *****************\n[lpfc/lpfc] skipping: [instance]\n[lpfc/lpfc] \n[lpfc/lpfc] PLAY RECAP *********************************************************************\n[lpfc/lpfc] instance : ok=2 changed=0 unreachable=0 failed=0 skipped=5 rescued=0 ignored=0 \n[lpfc/lpfc] \n==> [lpfc] Deployment complete\n[udev/udev] changed: [instance]\n[udev/udev] \n[udev/udev] TASK [vexxhost.atmosphere.udev : Add udev rules for SCSI Unit Attention] *******\n[udev/udev] changed: [instance]\n[udev/udev] \n[udev/udev] RUNNING HANDLER [vexxhost.atmosphere.udev : Reload udev] ***********************\n[ceph/ceph] ok: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.forget_package : Forget package] *********************\n[iscsi/iscsi] ok: [instance]\n[iscsi/iscsi] \n[iscsi/iscsi] TASK [vexxhost.atmosphere.iscsi : Ensure iscsid is started] ********************\n[udev/udev] ok: [instance]\n[udev/udev] \n[udev/udev] PLAY RECAP *********************************************************************\n[udev/udev] instance : ok=4 changed=2 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 \n[udev/udev] \n[ceph/ceph] ok: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.package : Update state for tar] **********************\n[ceph/ceph] skipping: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.directory : Create directory (/var/lib/downloads)] ***\n==> [udev] Deployment complete\n[iscsi/iscsi] changed: [instance]\n[iscsi/iscsi] \n[iscsi/iscsi] PLAY RECAP *********************************************************************\n[iscsi/iscsi] instance : ok=3 changed=1 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 \n[iscsi/iscsi] \n==> [iscsi] Deployment complete\n[ceph/ceph] changed: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.download_artifact : Starting download of file] *******\n[ceph/ceph] ok: [instance] => {\n[ceph/ceph] \"msg\": \"https://github.com/opencontainers/runc/releases/download/v1.4.2/runc.amd64\"\n[ceph/ceph] }\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.download_artifact : Download item] *******************\n[ceph/ceph] changed: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.download_artifact : Extract archive] *****************\n[ceph/ceph] skipping: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.package : Update state for tar] **********************\n[ceph/ceph] skipping: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.forget_package : Forget package] *********************\n[ceph/ceph] ok: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.package : Update state for tar] **********************\n[ceph/ceph] ok: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.download_artifact : Starting download of file] *******\n[ceph/ceph] ok: [instance] => {\n[ceph/ceph] \"msg\": \"https://github.com/containerd/containerd/releases/download/v2.2.2/containerd-2.2.2-linux-amd64.tar.gz\"\n[ceph/ceph] }\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.download_artifact : Download item] *******************\n[multipathd/multipathd] changed: [instance]\n[multipathd/multipathd] \n[multipathd/multipathd] TASK [vexxhost.atmosphere.multipathd : Install the multipathd package] *********\n[ceph/ceph] changed: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.download_artifact : Extract archive] *****************\n[ceph/ceph] changed: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.containerd : Install SELinux packages] ***************\n[ceph/ceph] skipping: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.containerd : Set SELinux to permissive at runtime] ***\n[ceph/ceph] skipping: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.containerd : Persist SELinux permissive mode] ********\n[ceph/ceph] skipping: [instance]\n[ceph/ceph] \n[ceph/ceph] TASK [vexxhost.containers.containerd : Install AppArmor packages] **************\n[ceph/ceph] fatal: [instance]: FAILED! => {\"cache_update_time\": 1776414298, \"cache_updated\": false, \"changed\": false, \"msg\": \"'/usr/bin/apt-get -y -o \\\"Dpkg::Options::=--force-confdef\\\" -o \\\"Dpkg::Options::=--force-confold\\\" install 'apparmor=3.0.4-2ubuntu2.5' 'apparmor-utils=3.0.4-2ubuntu2.5'' failed: E: Could not get lock /var/lib/dpkg/lock-frontend. It is held by process 12331 (apt-get)\\nE: Unable to acquire the dpkg frontend lock (/var/lib/dpkg/lock-frontend), is another process using it?\\n\", \"rc\": 100, \"stderr\": \"E: Could not get lock /var/lib/dpkg/lock-frontend. It is held by process 12331 (apt-get)\\nE: Unable to acquire the dpkg frontend lock (/var/lib/dpkg/lock-frontend), is another process using it?\\n\", \"stderr_lines\": [\"E: Could not get lock /var/lib/dpkg/lock-frontend. It is held by process 12331 (apt-get)\", \"E: Unable to acquire the dpkg frontend lock (/var/lib/dpkg/lock-frontend), is another process using it?\"], \"stdout\": \"\", \"stdout_lines\": []}\n[ceph/ceph] \n[ceph/ceph] PLAY RECAP *********************************************************************\n[ceph/ceph] instance : ok=12 changed=4 unreachable=0 failed=1 skipped=7 rescued=0 ignored=0 \n[ceph/ceph] ", "stdout_lines": ["==> Running preflight checks", "[preflight] ", "[preflight] PLAY [Preflight checks] ********************************************************", "[preflight] ", "[preflight] TASK [Fail if atmosphere_ceph_enabled is set] **********************************", "[preflight] skipping: [instance]", "[preflight] ", "[preflight] PLAY RECAP *********************************************************************", "[preflight] instance : ok=0 changed=0 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 ", "[preflight] ", "==> Preflight checks passed", "==> Starting parallel deployment", "==> [ceph] Starting deployment", "==> [udev] Starting deployment", "==> [lpfc] Starting deployment", "==> [iscsi] Starting deployment", "==> [multipathd] Starting deployment", "==> [kubernetes] Starting deployment", "[multipathd/multipathd] ", "[multipathd/multipathd] PLAY [controllers:computes] ****************************************************", "[multipathd/multipathd] ", "[multipathd/multipathd] TASK [Gathering Facts] *********************************************************", "[iscsi/iscsi] ", "[iscsi/iscsi] PLAY [controllers:computes] ****************************************************", "[iscsi/iscsi] ", "[iscsi/iscsi] TASK [Gathering Facts] *********************************************************", "[lpfc/lpfc] ", "[lpfc/lpfc] PLAY [controllers:computes] ****************************************************", "[lpfc/lpfc] ", "[lpfc/lpfc] TASK [Gathering Facts] *********************************************************", "[udev/udev] ", "[udev/udev] PLAY [controllers:computes] ****************************************************", "[udev/udev] ", "[udev/udev] TASK [Gathering Facts] *********************************************************", "[ceph/ceph] ", "[ceph/ceph] PLAY [all] *********************************************************************", "[ceph/ceph] ", "[ceph/ceph] TASK [Gathering Facts] *********************************************************", "[multipathd/multipathd] [WARNING]: Platform linux on host instance is using the discovered Python", "[multipathd/multipathd] interpreter at /usr/bin/python3.10, but future installation of another Python", "[multipathd/multipathd] interpreter could change the meaning of that path. See", "[multipathd/multipathd] https://docs.ansible.com/ansible-", "[multipathd/multipathd] core/2.17/reference_appendices/interpreter_discovery.html for more information.", "[multipathd/multipathd] ok: [instance]", "[multipathd/multipathd] ", "[multipathd/multipathd] TASK [vexxhost.atmosphere.multipathd : Add backports PPA] **********************", "[udev/udev] [WARNING]: Platform linux on host instance is using the discovered Python", "[udev/udev] interpreter at /usr/bin/python3.10, but future installation of another Python", "[udev/udev] interpreter could change the meaning of that path. See", "[udev/udev] https://docs.ansible.com/ansible-", "[udev/udev] core/2.17/reference_appendices/interpreter_discovery.html for more information.", "[iscsi/iscsi] [WARNING]: Platform linux on host instance is using the discovered Python", "[iscsi/iscsi] interpreter at /usr/bin/python3.10, but future installation of another Python", "[iscsi/iscsi] interpreter could change the meaning of that path. See", "[iscsi/iscsi] https://docs.ansible.com/ansible-", "[iscsi/iscsi] core/2.17/reference_appendices/interpreter_discovery.html for more information.", "[iscsi/iscsi] ok: [instance]", "[iscsi/iscsi] ", "[iscsi/iscsi] TASK [vexxhost.atmosphere.iscsi : Install iscsi package] ***********************", "[lpfc/lpfc] [WARNING]: Platform linux on host instance is using the discovered Python", "[lpfc/lpfc] interpreter at /usr/bin/python3.10, but future installation of another Python", "[lpfc/lpfc] interpreter could change the meaning of that path. See", "[lpfc/lpfc] https://docs.ansible.com/ansible-", "[lpfc/lpfc] core/2.17/reference_appendices/interpreter_discovery.html for more information.", "[lpfc/lpfc] ok: [instance]", "[lpfc/lpfc] ", "[lpfc/lpfc] TASK [vexxhost.atmosphere.lpfc : Detect if the \"lpfc\" module is loaded] ********", "[udev/udev] ok: [instance]", "[udev/udev] ", "[udev/udev] TASK [vexxhost.atmosphere.udev : Add udev rules for Pure Storage FlashArray] ***", "[ceph/ceph] [WARNING]: Platform linux on host instance is using the discovered Python", "[ceph/ceph] interpreter at /usr/bin/python3.10, but future installation of another Python", "[ceph/ceph] interpreter could change the meaning of that path. See", "[ceph/ceph] https://docs.ansible.com/ansible-", "[ceph/ceph] core/2.17/reference_appendices/interpreter_discovery.html for more information.", "[ceph/ceph] ok: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [Fail if atmosphere_ceph_enabled is set] **********************************", "[ceph/ceph] skipping: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [Set a fact with the \"atmosphere_images\" for other plays] *****************", "[ceph/ceph] ok: [instance]", "[ceph/ceph] ", "[ceph/ceph] PLAY [Deploy Ceph monitors & managers] *****************************************", "[ceph/ceph] ", "[ceph/ceph] TASK [Gathering Facts] *********************************************************", "[lpfc/lpfc] ok: [instance]", "[lpfc/lpfc] ", "[lpfc/lpfc] TASK [vexxhost.atmosphere.lpfc : Install the configuration file] ***************", "[lpfc/lpfc] skipping: [instance]", "[lpfc/lpfc] ", "[lpfc/lpfc] TASK [vexxhost.atmosphere.lpfc : Get the values for the module parameters] *****", "[lpfc/lpfc] skipping: [instance] => (item=lpfc_lun_queue_depth) ", "[lpfc/lpfc] skipping: [instance] => (item=lpfc_sg_seg_cnt) ", "[lpfc/lpfc] skipping: [instance] => (item=lpfc_max_luns) ", "[lpfc/lpfc] skipping: [instance] => (item=lpfc_enable_fc4_type) ", "[lpfc/lpfc] skipping: [instance]", "[lpfc/lpfc] ", "[lpfc/lpfc] TASK [vexxhost.atmosphere.lpfc : Detect if the run-time module parameters are set correctly] ***", "[lpfc/lpfc] skipping: [instance] => (item=lpfc_lun_queue_depth) ", "[lpfc/lpfc] skipping: [instance] => (item=lpfc_sg_seg_cnt) ", "[lpfc/lpfc] skipping: [instance] => (item=lpfc_max_luns) ", "[lpfc/lpfc] skipping: [instance] => (item=lpfc_enable_fc4_type) ", "[lpfc/lpfc] skipping: [instance]", "[lpfc/lpfc] ", "[lpfc/lpfc] TASK [vexxhost.atmosphere.lpfc : Update \"initramfs\" if the configuration file has changed] ***", "[lpfc/lpfc] skipping: [instance]", "[lpfc/lpfc] ", "[lpfc/lpfc] TASK [Reboot the system if the configuration file has changed] *****************", "[lpfc/lpfc] skipping: [instance]", "[lpfc/lpfc] ", "[lpfc/lpfc] PLAY RECAP *********************************************************************", "[lpfc/lpfc] instance : ok=2 changed=0 unreachable=0 failed=0 skipped=5 rescued=0 ignored=0 ", "[lpfc/lpfc] ", "==> [lpfc] Deployment complete", "[udev/udev] changed: [instance]", "[udev/udev] ", "[udev/udev] TASK [vexxhost.atmosphere.udev : Add udev rules for SCSI Unit Attention] *******", "[udev/udev] changed: [instance]", "[udev/udev] ", "[udev/udev] RUNNING HANDLER [vexxhost.atmosphere.udev : Reload udev] ***********************", "[ceph/ceph] ok: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.forget_package : Forget package] *********************", "[iscsi/iscsi] ok: [instance]", "[iscsi/iscsi] ", "[iscsi/iscsi] TASK [vexxhost.atmosphere.iscsi : Ensure iscsid is started] ********************", "[udev/udev] ok: [instance]", "[udev/udev] ", "[udev/udev] PLAY RECAP *********************************************************************", "[udev/udev] instance : ok=4 changed=2 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 ", "[udev/udev] ", "[ceph/ceph] ok: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.package : Update state for tar] **********************", "[ceph/ceph] skipping: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.directory : Create directory (/var/lib/downloads)] ***", "==> [udev] Deployment complete", "[iscsi/iscsi] changed: [instance]", "[iscsi/iscsi] ", "[iscsi/iscsi] PLAY RECAP *********************************************************************", "[iscsi/iscsi] instance : ok=3 changed=1 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 ", "[iscsi/iscsi] ", "==> [iscsi] Deployment complete", "[ceph/ceph] changed: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.download_artifact : Starting download of file] *******", "[ceph/ceph] ok: [instance] => {", "[ceph/ceph] \"msg\": \"https://github.com/opencontainers/runc/releases/download/v1.4.2/runc.amd64\"", "[ceph/ceph] }", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.download_artifact : Download item] *******************", "[ceph/ceph] changed: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.download_artifact : Extract archive] *****************", "[ceph/ceph] skipping: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.package : Update state for tar] **********************", "[ceph/ceph] skipping: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.forget_package : Forget package] *********************", "[ceph/ceph] ok: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.package : Update state for tar] **********************", "[ceph/ceph] ok: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.download_artifact : Starting download of file] *******", "[ceph/ceph] ok: [instance] => {", "[ceph/ceph] \"msg\": \"https://github.com/containerd/containerd/releases/download/v2.2.2/containerd-2.2.2-linux-amd64.tar.gz\"", "[ceph/ceph] }", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.download_artifact : Download item] *******************", "[multipathd/multipathd] changed: [instance]", "[multipathd/multipathd] ", "[multipathd/multipathd] TASK [vexxhost.atmosphere.multipathd : Install the multipathd package] *********", "[ceph/ceph] changed: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.download_artifact : Extract archive] *****************", "[ceph/ceph] changed: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.containerd : Install SELinux packages] ***************", "[ceph/ceph] skipping: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.containerd : Set SELinux to permissive at runtime] ***", "[ceph/ceph] skipping: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.containerd : Persist SELinux permissive mode] ********", "[ceph/ceph] skipping: [instance]", "[ceph/ceph] ", "[ceph/ceph] TASK [vexxhost.containers.containerd : Install AppArmor packages] **************", "[ceph/ceph] fatal: [instance]: FAILED! => {\"cache_update_time\": 1776414298, \"cache_updated\": false, \"changed\": false, \"msg\": \"'/usr/bin/apt-get -y -o \\\"Dpkg::Options::=--force-confdef\\\" -o \\\"Dpkg::Options::=--force-confold\\\" install 'apparmor=3.0.4-2ubuntu2.5' 'apparmor-utils=3.0.4-2ubuntu2.5'' failed: E: Could not get lock /var/lib/dpkg/lock-frontend. It is held by process 12331 (apt-get)\\nE: Unable to acquire the dpkg frontend lock (/var/lib/dpkg/lock-frontend), is another process using it?\\n\", \"rc\": 100, \"stderr\": \"E: Could not get lock /var/lib/dpkg/lock-frontend. It is held by process 12331 (apt-get)\\nE: Unable to acquire the dpkg frontend lock (/var/lib/dpkg/lock-frontend), is another process using it?\\n\", \"stderr_lines\": [\"E: Could not get lock /var/lib/dpkg/lock-frontend. It is held by process 12331 (apt-get)\", \"E: Unable to acquire the dpkg frontend lock (/var/lib/dpkg/lock-frontend), is another process using it?\"], \"stdout\": \"\", \"stdout_lines\": []}", "[ceph/ceph] ", "[ceph/ceph] PLAY RECAP *********************************************************************", "[ceph/ceph] instance : ok=12 changed=4 unreachable=0 failed=1 skipped=7 rescued=0 ignored=0 ", "[ceph/ceph] "]} 2026-04-17 08:25:51.336948 | instance | 2026-04-17 08:25:51.337124 | instance | PLAY RECAP ********************************************************************* 2026-04-17 08:25:51.337408 | instance | instance : ok=1 changed=0 unreachable=0 failed=1 skipped=0 rescued=0 ignored=0 2026-04-17 08:25:51.337655 | instance | 2026-04-17 08:25:51.337934 | instance | Friday 17 April 2026 08:25:51 +0000 (0:01:02.066) 0:01:09.483 ********** 2026-04-17 08:25:51.338186 | instance | =============================================================================== 2026-04-17 08:25:51.338447 | instance | Deploy with parallel orchestrator -------------------------------------- 62.07s 2026-04-17 08:25:51.338774 | instance | Build atmosphere binary ------------------------------------------------- 7.41s 2026-04-17 08:25:51.429429 | instance | CRITICAL Ansible return code was 2, command was: ansible-playbook --inventory /home/zuul/.ansible/tmp/molecule.v9Wo.aio/inventory --skip-tags molecule-notest,notest --inventory=/home/zuul/src/github.com/vexxhost/atmosphere/inventory.yaml /home/zuul/src/github.com/vexxhost/atmosphere/molecule/aio/converge.yml 2026-04-17 08:25:51.429609 | instance | ERROR [aio > converge] Executed: Failed 2026-04-17 08:25:51.429744 | instance | ERROR Ansible return code was 2, command was: ansible-playbook --inventory /home/zuul/.ansible/tmp/molecule.v9Wo.aio/inventory --skip-tags molecule-notest,notest --inventory=/home/zuul/src/github.com/vexxhost/atmosphere/inventory.yaml /home/zuul/src/github.com/vexxhost/atmosphere/molecule/aio/converge.yml 2026-04-17 08:25:51.956304 | instance | ERROR 2026-04-17 08:25:51.956508 | instance | { 2026-04-17 08:25:51.956539 | instance | "delta": "0:01:24.962428", 2026-04-17 08:25:51.956560 | instance | "end": "2026-04-17 08:25:51.486054", 2026-04-17 08:25:51.956579 | instance | "msg": "non-zero return code", 2026-04-17 08:25:51.956597 | instance | "rc": 2, 2026-04-17 08:25:51.956619 | instance | "start": "2026-04-17 08:24:26.523626" 2026-04-17 08:25:51.956639 | instance | } failure 2026-04-17 08:25:51.959229 | 2026-04-17 08:25:51.959288 | PLAY RECAP 2026-04-17 08:25:51.959336 | instance | ok: 2 changed: 2 unreachable: 0 failed: 1 skipped: 0 rescued: 0 ignored: 0 2026-04-17 08:25:51.959358 | 2026-04-17 08:25:52.175709 | RUN END RESULT_NORMAL: [untrusted : github.com/vexxhost/zuul-jobs/playbooks/molecule/run.yaml@main] 2026-04-17 08:25:52.186272 | POST-RUN START: [untrusted : github.com/vexxhost/atmosphere/test-playbooks/molecule/post.yml@main] 2026-04-17 08:25:52.822659 | 2026-04-17 08:25:52.822829 | PLAY [all] 2026-04-17 08:25:52.838732 | 2026-04-17 08:25:52.838897 | TASK [gather-host-logs : creating directory for system status] 2026-04-17 08:25:53.183419 | instance | changed 2026-04-17 08:25:53.190714 | 2026-04-17 08:25:53.190850 | TASK [gather-host-logs : Get logs for each host] 2026-04-17 08:25:53.546854 | instance | + systemd-cgls --full --all --no-pager 2026-04-17 08:25:53.561084 | instance | + ip addr 2026-04-17 08:25:53.562622 | instance | + ip route 2026-04-17 08:25:53.563878 | instance | + lsblk 2026-04-17 08:25:53.566679 | instance | + mount 2026-04-17 08:25:53.568015 | instance | + docker images 2026-04-17 08:25:53.568378 | instance | /bin/bash: line 7: docker: command not found 2026-04-17 08:25:53.568599 | instance | + brctl show 2026-04-17 08:25:53.568975 | instance | /bin/bash: line 8: brctl: command not found 2026-04-17 08:25:53.569183 | instance | + ps aux --sort=-%mem 2026-04-17 08:25:53.581096 | instance | + dpkg -l 2026-04-17 08:25:53.593248 | instance | + CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul)) 2026-04-17 08:25:53.594013 | instance | ++ docker ps -a --format '{{ .Names }}' --filter label=zuul 2026-04-17 08:25:53.594139 | instance | /bin/bash: line 11: docker: command not found 2026-04-17 08:25:53.594563 | instance | + '[' '!' -z '' ']' 2026-04-17 08:25:53.733454 | instance | ok: Runtime: 0:00:00.053846 2026-04-17 08:25:53.745929 | 2026-04-17 08:25:53.746098 | TASK [gather-host-logs : Downloads logs to executor] 2026-04-17 08:25:54.376950 | instance | changed: 2026-04-17 08:25:54.377205 | instance | created directory /var/lib/zuul/builds/b48b679401b84230b69e09a70b891bfd/work/logs/instance 2026-04-17 08:25:54.377249 | instance | cd+++++++++ system/ 2026-04-17 08:25:54.377282 | instance | >f+++++++++ system/brctl-show.txt 2026-04-17 08:25:54.377312 | instance | >f+++++++++ system/docker-images.txt 2026-04-17 08:25:54.377341 | instance | >f+++++++++ system/ip-addr.txt 2026-04-17 08:25:54.377373 | instance | >f+++++++++ system/ip-route.txt 2026-04-17 08:25:54.377403 | instance | >f+++++++++ system/lsblk.txt 2026-04-17 08:25:54.377433 | instance | >f+++++++++ system/mount.txt 2026-04-17 08:25:54.377464 | instance | >f+++++++++ system/packages.txt 2026-04-17 08:25:54.377495 | instance | >f+++++++++ system/ps.txt 2026-04-17 08:25:54.377524 | instance | >f+++++++++ system/systemd-cgls.txt 2026-04-17 08:25:54.389585 | 2026-04-17 08:25:54.389653 | LOOP [helm-release-status : creating directory for helm release status] 2026-04-17 08:25:54.589562 | instance | changed: "values" 2026-04-17 08:25:54.767326 | instance | changed: "releases" 2026-04-17 08:25:54.787493 | 2026-04-17 08:25:54.787663 | TASK [helm-release-status : Gather get release status for helm charts] 2026-04-17 08:25:54.996998 | instance | /bin/bash: line 3: kubectl: command not found 2026-04-17 08:25:55.325264 | instance | ok: Runtime: 0:00:00.006553 2026-04-17 08:25:55.330564 | 2026-04-17 08:25:55.330636 | TASK [helm-release-status : Downloads logs to executor] 2026-04-17 08:25:55.801133 | instance | changed: 2026-04-17 08:25:55.801338 | instance | cd+++++++++ helm/ 2026-04-17 08:25:55.801377 | instance | cd+++++++++ helm/releases/ 2026-04-17 08:25:55.801407 | instance | cd+++++++++ helm/values/ 2026-04-17 08:25:55.814120 | 2026-04-17 08:25:55.814209 | TASK [describe-kubernetes-objects : creating directory for cluster scoped objects] 2026-04-17 08:25:56.021317 | instance | changed 2026-04-17 08:25:56.058374 | 2026-04-17 08:25:56.058516 | TASK [describe-kubernetes-objects : Gathering descriptions for cluster scoped objects] 2026-04-17 08:25:56.265679 | instance | xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value 2026-04-17 08:25:56.266045 | instance | xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value 2026-04-17 08:25:56.273171 | instance | environment: line 1: kubectl: command not found 2026-04-17 08:25:56.274960 | instance | xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value 2026-04-17 08:25:56.276008 | instance | environment: line 1: kubectl: command not found 2026-04-17 08:25:56.277737 | instance | xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value 2026-04-17 08:25:56.278505 | instance | environment: line 1: kubectl: command not found 2026-04-17 08:25:56.280320 | instance | xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value 2026-04-17 08:25:56.281143 | instance | environment: line 1: kubectl: command not found 2026-04-17 08:25:56.283071 | instance | xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value 2026-04-17 08:25:56.283690 | instance | environment: line 1: kubectl: command not found 2026-04-17 08:25:56.285269 | instance | xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value 2026-04-17 08:25:56.595837 | instance | ok: Runtime: 0:00:00.027331 2026-04-17 08:25:56.601669 | 2026-04-17 08:25:56.601770 | TASK [describe-kubernetes-objects : creating directory for namespace scoped objects] 2026-04-17 08:25:56.807863 | instance | changed 2026-04-17 08:25:56.815068 | 2026-04-17 08:25:56.815143 | TASK [describe-kubernetes-objects : Gathering descriptions for namespace scoped objects] 2026-04-17 08:25:57.027452 | instance | environment: line 5: kubectl: command not found 2026-04-17 08:25:57.028236 | instance | xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value 2026-04-17 08:25:57.028406 | instance | xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value 2026-04-17 08:25:57.028856 | instance | xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value 2026-04-17 08:25:57.353308 | instance | ok: Runtime: 0:00:00.010051 2026-04-17 08:25:57.360890 | 2026-04-17 08:25:57.361009 | TASK [describe-kubernetes-objects : Downloads logs to executor] 2026-04-17 08:25:57.841277 | instance | changed: 2026-04-17 08:25:57.841683 | instance | cd+++++++++ objects/ 2026-04-17 08:25:57.841781 | instance | cd+++++++++ objects/cluster/ 2026-04-17 08:25:57.841841 | instance | cd+++++++++ objects/namespaced/ 2026-04-17 08:25:57.851242 | 2026-04-17 08:25:57.851312 | TASK [gather-pod-logs : creating directory for pod logs] 2026-04-17 08:25:58.054193 | instance | changed 2026-04-17 08:25:58.059335 | 2026-04-17 08:25:58.059400 | TASK [gather-pod-logs : creating directory for failed pod logs] 2026-04-17 08:25:58.260866 | instance | changed 2026-04-17 08:25:58.269860 | 2026-04-17 08:25:58.269970 | TASK [gather-pod-logs : retrieve all kubernetes logs, current and previous (if they exist)] 2026-04-17 08:25:58.495716 | instance | environment: line 3: kubectl: command not found 2026-04-17 08:25:58.818273 | instance | ok: Runtime: 0:00:00.008472 2026-04-17 08:25:58.825332 | 2026-04-17 08:25:58.825428 | TASK [gather-pod-logs : Downloads pod logs to executor] 2026-04-17 08:25:59.311418 | instance | changed: 2026-04-17 08:25:59.311665 | instance | cd+++++++++ pod-logs/ 2026-04-17 08:25:59.311715 | instance | cd+++++++++ pod-logs/failed-pods/ 2026-04-17 08:25:59.324591 | 2026-04-17 08:25:59.324659 | TASK [gather-prom-metrics : creating directory for helm release descriptions] 2026-04-17 08:25:59.529507 | instance | changed 2026-04-17 08:25:59.535369 | 2026-04-17 08:25:59.535471 | TASK [gather-prom-metrics : Get metrics from exporter services in all namespaces] 2026-04-17 08:25:59.744506 | instance | /bin/bash: line 2: kubectl: command not found 2026-04-17 08:26:00.228043 | instance | ok: Runtime: 0:00:00.038824 2026-04-17 08:26:00.232807 | 2026-04-17 08:26:00.232872 | TASK [gather-prom-metrics : Get ceph metrics from ceph-mgr] 2026-04-17 08:26:00.438397 | instance | /bin/bash: line 2: kubectl: command not found 2026-04-17 08:26:00.467392 | instance | ceph-mgr endpoints: 2026-04-17 08:26:00.767615 | instance | ok: Runtime: 0:00:00.034272 2026-04-17 08:26:00.774512 | 2026-04-17 08:26:00.774618 | TASK [gather-prom-metrics : Get metrics from fluentd pods] 2026-04-17 08:26:00.982020 | instance | /bin/bash: line 4: kubectl: command not found 2026-04-17 08:26:01.313957 | instance | ok: Runtime: 0:00:00.036260 2026-04-17 08:26:01.323885 | 2026-04-17 08:26:01.324022 | TASK [gather-prom-metrics : Downloads logs to executor] 2026-04-17 08:26:01.821268 | instance | changed: cd+++++++++ prometheus/ 2026-04-17 08:26:01.832624 | 2026-04-17 08:26:01.832719 | TASK [gather-selenium-data : creating directory for helm release descriptions] 2026-04-17 08:26:02.094130 | instance | changed 2026-04-17 08:26:02.099484 | 2026-04-17 08:26:02.099555 | TASK [gather-selenium-data : Get selenium data] 2026-04-17 08:26:02.307524 | instance | + cp '/tmp/artifacts/*' /tmp/logs/selenium/. 2026-04-17 08:26:02.309063 | instance | cp: cannot stat '/tmp/artifacts/*': No such file or directory 2026-04-17 08:26:02.635767 | instance | ERROR 2026-04-17 08:26:02.636021 | instance | { 2026-04-17 08:26:02.636056 | instance | "delta": "0:00:00.007059", 2026-04-17 08:26:02.636078 | instance | "end": "2026-04-17 08:26:02.309471", 2026-04-17 08:26:02.636097 | instance | "msg": "non-zero return code", 2026-04-17 08:26:02.636115 | instance | "rc": 1, 2026-04-17 08:26:02.636133 | instance | "start": "2026-04-17 08:26:02.302412" 2026-04-17 08:26:02.636151 | instance | } 2026-04-17 08:26:02.636175 | instance | ERROR: Ignoring Errors 2026-04-17 08:26:02.642308 | 2026-04-17 08:26:02.642396 | TASK [gather-selenium-data : Downloads logs to executor] 2026-04-17 08:26:03.136177 | instance | changed: cd+++++++++ selenium/ 2026-04-17 08:26:03.142962 | 2026-04-17 08:26:03.143016 | PLAY RECAP 2026-04-17 08:26:03.143063 | instance | ok: 23 changed: 23 unreachable: 0 failed: 0 skipped: 0 rescued: 0 ignored: 1 2026-04-17 08:26:03.143086 | 2026-04-17 08:26:03.265923 | POST-RUN END RESULT_NORMAL: [untrusted : github.com/vexxhost/atmosphere/test-playbooks/molecule/post.yml@main] 2026-04-17 08:26:03.279930 | POST-RUN START: [trusted : github.com/vexxhost/zuul-config/playbooks/base/post.yaml@main] 2026-04-17 08:26:03.905364 | 2026-04-17 08:26:03.905543 | PLAY [all] 2026-04-17 08:26:03.917383 | 2026-04-17 08:26:03.917475 | TASK [fetch-output : Set log path for multiple nodes] 2026-04-17 08:26:03.962331 | instance | skipping: Conditional result was False 2026-04-17 08:26:03.973776 | 2026-04-17 08:26:03.973905 | TASK [fetch-output : Set log path for single node] 2026-04-17 08:26:04.018838 | instance | ok 2026-04-17 08:26:04.026593 | 2026-04-17 08:26:04.026691 | LOOP [fetch-output : Ensure local output dirs] 2026-04-17 08:26:04.395030 | instance -> localhost | ok: "/var/lib/zuul/builds/b48b679401b84230b69e09a70b891bfd/work/logs" 2026-04-17 08:26:04.600707 | instance -> localhost | changed: "/var/lib/zuul/builds/b48b679401b84230b69e09a70b891bfd/work/artifacts" 2026-04-17 08:26:04.815367 | instance -> localhost | changed: "/var/lib/zuul/builds/b48b679401b84230b69e09a70b891bfd/work/docs" 2026-04-17 08:26:04.831273 | 2026-04-17 08:26:04.831444 | LOOP [fetch-output : Collect logs, artifacts and docs] 2026-04-17 08:26:05.457975 | instance | changed: .d..t...... ./ 2026-04-17 08:26:05.458208 | instance | changed: All items complete 2026-04-17 08:26:05.458246 | 2026-04-17 08:26:05.905738 | instance | changed: .d..t...... ./ 2026-04-17 08:26:06.352609 | instance | changed: .d..t...... ./ 2026-04-17 08:26:06.376693 | 2026-04-17 08:26:06.376818 | LOOP [merge-output-to-logs : Move artifacts and docs to logs dir] 2026-04-17 08:26:06.817888 | instance -> localhost | ok: Item: artifacts Runtime: 0:00:00.007539 2026-04-17 08:26:07.030340 | instance -> localhost | ok: Item: docs Runtime: 0:00:00.007042 2026-04-17 08:26:07.040280 | 2026-04-17 08:26:07.040444 | PLAY [all] 2026-04-17 08:26:07.050586 | 2026-04-17 08:26:07.050686 | TASK [remove-build-sshkey : Remove the build SSH key from all nodes] 2026-04-17 08:26:07.493190 | instance | changed 2026-04-17 08:26:07.499216 | 2026-04-17 08:26:07.499297 | PLAY RECAP 2026-04-17 08:26:07.499343 | instance | ok: 5 changed: 4 unreachable: 0 failed: 0 skipped: 1 rescued: 0 ignored: 0 2026-04-17 08:26:07.499368 | 2026-04-17 08:26:07.633880 | POST-RUN END RESULT_NORMAL: [trusted : github.com/vexxhost/zuul-config/playbooks/base/post.yaml@main] 2026-04-17 08:26:07.641388 | POST-RUN START: [trusted : github.com/vexxhost/zuul-config/playbooks/base/post-logs.yaml@main] 2026-04-17 08:26:08.235898 | 2026-04-17 08:26:08.236054 | PLAY [localhost] 2026-04-17 08:26:08.247836 | 2026-04-17 08:26:08.247935 | TASK [Generate Zuul manifest] 2026-04-17 08:26:08.269549 | localhost | ok 2026-04-17 08:26:08.285191 | 2026-04-17 08:26:08.285309 | TASK [generate-zuul-manifest : Generate Zuul manifest] 2026-04-17 08:26:08.634576 | localhost | changed 2026-04-17 08:26:08.644322 | 2026-04-17 08:26:08.644407 | TASK [generate-zuul-manifest : Return Zuul manifest URL to Zuul] 2026-04-17 08:26:08.671384 | localhost | ok 2026-04-17 08:26:08.677782 | 2026-04-17 08:26:08.677845 | TASK [Upload logs] 2026-04-17 08:26:08.698639 | localhost | ok 2026-04-17 08:26:08.800364 | 2026-04-17 08:26:08.800501 | TASK [Set zuul-log-path fact] 2026-04-17 08:26:08.820429 | localhost | ok 2026-04-17 08:26:08.834059 | 2026-04-17 08:26:08.834132 | TASK [set-zuul-log-path-fact : Set log path for a build] 2026-04-17 08:26:08.865587 | localhost | ok 2026-04-17 08:26:08.873472 | 2026-04-17 08:26:08.873543 | TASK [upload-logs : Create log directories] 2026-04-17 08:26:09.230212 | localhost | changed 2026-04-17 08:26:09.236505 | 2026-04-17 08:26:09.236582 | TASK [upload-logs : Ensure logs are readable before uploading] 2026-04-17 08:26:09.603393 | localhost -> localhost | ok: Runtime: 0:00:00.006646 2026-04-17 08:26:09.610546 | 2026-04-17 08:26:09.610643 | TASK [upload-logs : Upload logs to log server] 2026-04-17 08:26:10.087843 | localhost | Output suppressed because no_log was given 2026-04-17 08:26:10.092685 | 2026-04-17 08:26:10.092774 | LOOP [upload-logs : Compress console log and json output] 2026-04-17 08:26:10.143712 | localhost | skipping: Conditional result was False 2026-04-17 08:26:10.151051 | localhost | skipping: Conditional result was False 2026-04-17 08:26:10.164463 | 2026-04-17 08:26:10.164600 | LOOP [upload-logs : Upload compressed console log and json output] 2026-04-17 08:26:10.208124 | localhost | skipping: Conditional result was False 2026-04-17 08:26:10.208613 | 2026-04-17 08:26:10.212004 | localhost | skipping: Conditional result was False 2026-04-17 08:26:10.221394 | 2026-04-17 08:26:10.221587 | LOOP [upload-logs : Upload console log and json output]