[
{
    "branch": "main",
    "index": "0",
    "phase": "pre",
    "playbook": "github.com/vexxhost/zuul-config/playbooks/base/pre.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-04-04T10:06:12.282443Z",
                    "start": "2026-04-04T10:06:08.523503Z"
                },
                "id": "0242ac17-0011-9f12-8fac-000000000002",
                "name": "localhost"
            },
            "tasks": [
                {
                    "hosts": {
                        "localhost": {
                            "action": "gather_facts",
                            "ansible_facts": {
                                "ansible_apparmor": {
                                    "status": "disabled"
                                },
                                "ansible_architecture": "x86_64",
                                "ansible_bios_date": "NA",
                                "ansible_bios_vendor": "NA",
                                "ansible_bios_version": "NA",
                                "ansible_board_asset_tag": "NA",
                                "ansible_board_name": "NA",
                                "ansible_board_serial": "NA",
                                "ansible_board_vendor": "NA",
                                "ansible_board_version": "NA",
                                "ansible_chassis_asset_tag": "NA",
                                "ansible_chassis_serial": "NA",
                                "ansible_chassis_vendor": "NA",
                                "ansible_chassis_version": "NA",
                                "ansible_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-5.15.0-130-generic",
                                    "console": "ttyS0",
                                    "ro": true,
                                    "root": "UUID=5a569d86-b935-46dd-ae79-7a72a25b6a4c"
                                },
                                "ansible_date_time": {
                                    "date": "2026-04-04",
                                    "day": "04",
                                    "epoch": "1775297169",
                                    "epoch_int": "1775297169",
                                    "hour": "10",
                                    "iso8601": "2026-04-04T10:06:09Z",
                                    "iso8601_basic": "20260404T100609457711",
                                    "iso8601_basic_short": "20260404T100609",
                                    "iso8601_micro": "2026-04-04T10:06:09.457711Z",
                                    "minute": "06",
                                    "month": "04",
                                    "second": "09",
                                    "time": "10:06:09",
                                    "tz": "UTC",
                                    "tz_dst": "UTC",
                                    "tz_offset": "+0000",
                                    "weekday": "Saturday",
                                    "weekday_number": "6",
                                    "weeknumber": "13",
                                    "year": "2026"
                                },
                                "ansible_devices": {},
                                "ansible_distribution": "Debian",
                                "ansible_distribution_major_version": "12",
                                "ansible_distribution_release": "bookworm",
                                "ansible_distribution_version": "12",
                                "ansible_dns": {
                                    "nameservers": [
                                        "127.0.0.11"
                                    ],
                                    "options": {
                                        "edns0": true,
                                        "ndots": "0",
                                        "trust-ad": true
                                    },
                                    "search": [
                                        "openstacklocal"
                                    ]
                                },
                                "ansible_domain": "",
                                "ansible_effective_group_id": 0,
                                "ansible_effective_user_id": 0,
                                "ansible_env": {
                                    "ANSIBLE_CONFIG": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/ansible/pre_playbook_0/ansible.cfg",
                                    "DEBIAN_FRONTEND": "noninteractive",
                                    "GPG_KEY": "A035C8C19219BA821ECEA86B64E628F8D684696D",
                                    "HOME": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work",
                                    "HOSTNAME": "0a8996d2b663",
                                    "LANG": "C.UTF-8",
                                    "PATH": "/usr/local/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",
                                    "PWD": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_0/github.com/vexxhost/zuul-config/playbooks/base",
                                    "PYTHONPATH": "/var/lib/zuul/ansible/9",
                                    "PYTHON_SHA256": "8d3ed8ec5c88c1c95f5e558612a725450d2452813ddad5e58fdb1a53b1209b78",
                                    "PYTHON_VERSION": "3.11.14",
                                    "SSH_AGENT_PID": "1133740",
                                    "SSH_AUTH_SOCK": "/tmp/ssh-dGmuYGbOiaJB/agent.1133739",
                                    "TMP": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/tmp",
                                    "ZUUL_ANSIBLE_SPLIT_STREAMS": "False",
                                    "ZUUL_JOBDIR": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078",
                                    "ZUUL_JOB_FAILURE_OUTPUT": "[]",
                                    "ZUUL_JOB_LOG_CONFIG": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/ansible/logging.json",
                                    "ZUUL_OUTPUT_MAX_BYTES": "1073741824"
                                },
                                "ansible_fibre_channel_wwn": [],
                                "ansible_fips": false,
                                "ansible_form_factor": "NA",
                                "ansible_fqdn": "0a8996d2b663",
                                "ansible_hostname": "0a8996d2b663",
                                "ansible_hostnqn": "",
                                "ansible_is_chroot": false,
                                "ansible_iscsi_iqn": "",
                                "ansible_kernel": "5.15.0-130-generic",
                                "ansible_kernel_version": "#140-Ubuntu SMP Wed Dec 18 17:59:53 UTC 2024",
                                "ansible_loadavg": {
                                    "15m": 2.87841796875,
                                    "1m": 9.56982421875,
                                    "5m": 4.6240234375
                                },
                                "ansible_local": {},
                                "ansible_lsb": {},
                                "ansible_lvm": "N/A",
                                "ansible_machine": "x86_64",
                                "ansible_memfree_mb": 3699,
                                "ansible_memory_mb": {
                                    "nocache": {
                                        "free": 16769,
                                        "used": 15321
                                    },
                                    "real": {
                                        "free": 3699,
                                        "total": 32090,
                                        "used": 28391
                                    },
                                    "swap": {
                                        "cached": 0,
                                        "free": 0,
                                        "total": 0,
                                        "used": 0
                                    }
                                },
                                "ansible_memtotal_mb": 32090,
                                "ansible_mounts": [
                                    {
                                        "block_available": 8836399,
                                        "block_size": 4096,
                                        "block_total": 40601580,
                                        "block_used": 31765181,
                                        "device": "/dev/vda1",
                                        "fstype": "ext4",
                                        "inode_available": 16499355,
                                        "inode_total": 20643840,
                                        "inode_used": 4144485,
                                        "mount": "/etc/resolv.conf",
                                        "options": "ro,nosuid,nodev,relatime,discard,errors=remount-ro,bind",
                                        "size_available": 36193890304,
                                        "size_total": 166304071680,
                                        "uuid": "N/A"
                                    },
                                    {
                                        "block_available": 8836399,
                                        "block_size": 4096,
                                        "block_total": 40601580,
                                        "block_used": 31765181,
                                        "device": "/dev/vda1",
                                        "fstype": "ext4",
                                        "inode_available": 16499355,
                                        "inode_total": 20643840,
                                        "inode_used": 4144485,
                                        "mount": "/etc/hosts",
                                        "options": "ro,nosuid,nodev,relatime,discard,errors=remount-ro,bind",
                                        "size_available": 36193890304,
                                        "size_total": 166304071680,
                                        "uuid": "N/A"
                                    },
                                    {
                                        "block_available": 8836367,
                                        "block_size": 4096,
                                        "block_total": 40601580,
                                        "block_used": 31765213,
                                        "device": "/dev/vda1",
                                        "fstype": "ext4",
                                        "inode_available": 16499355,
                                        "inode_total": 20643840,
                                        "inode_used": 4144485,
                                        "mount": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work",
                                        "options": "rw,nosuid,nodev,relatime,discard,errors=remount-ro,bind",
                                        "size_available": 36193759232,
                                        "size_total": 166304071680,
                                        "uuid": "N/A"
                                    },
                                    {
                                        "block_available": 8836367,
                                        "block_size": 4096,
                                        "block_total": 40601580,
                                        "block_used": 31765213,
                                        "device": "/dev/vda1",
                                        "fstype": "ext4",
                                        "inode_available": 16499355,
                                        "inode_total": 20643840,
                                        "inode_used": 4144485,
                                        "mount": "/var/lib/zuul/ansible/9",
                                        "options": "ro,nosuid,nodev,relatime,discard,errors=remount-ro,bind",
                                        "size_available": 36193759232,
                                        "size_total": 166304071680,
                                        "uuid": "N/A"
                                    },
                                    {
                                        "block_available": 8836367,
                                        "block_size": 4096,
                                        "block_total": 40601580,
                                        "block_used": 31765213,
                                        "device": "/dev/vda1",
                                        "fstype": "ext4",
                                        "inode_available": 16499355,
                                        "inode_total": 20643840,
                                        "inode_used": 4144485,
                                        "mount": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/ansible",
                                        "options": "ro,nosuid,nodev,relatime,discard,errors=remount-ro,bind",
                                        "size_available": 36193759232,
                                        "size_total": 166304071680,
                                        "uuid": "N/A"
                                    },
                                    {
                                        "block_available": 8836367,
                                        "block_size": 4096,
                                        "block_total": 40601580,
                                        "block_used": 31765213,
                                        "device": "/dev/vda1",
                                        "fstype": "ext4",
                                        "inode_available": 16499355,
                                        "inode_total": 20643840,
                                        "inode_used": 4144485,
                                        "mount": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted",
                                        "options": "ro,nosuid,nodev,relatime,discard,errors=remount-ro,bind",
                                        "size_available": 36193759232,
                                        "size_total": 166304071680,
                                        "uuid": "N/A"
                                    },
                                    {
                                        "block_available": 8836338,
                                        "block_size": 4096,
                                        "block_total": 40601580,
                                        "block_used": 31765242,
                                        "device": "/dev/vda1",
                                        "fstype": "ext4",
                                        "inode_available": 16499355,
                                        "inode_total": 20643840,
                                        "inode_used": 4144485,
                                        "mount": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted",
                                        "options": "ro,nosuid,nodev,relatime,discard,errors=remount-ro,bind",
                                        "size_available": 36193640448,
                                        "size_total": 166304071680,
                                        "uuid": "N/A"
                                    },
                                    {
                                        "block_available": 8836335,
                                        "block_size": 4096,
                                        "block_total": 40601580,
                                        "block_used": 31765245,
                                        "device": "/dev/vda1",
                                        "fstype": "ext4",
                                        "inode_available": 16499355,
                                        "inode_total": 20643840,
                                        "inode_used": 4144485,
                                        "mount": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/ansible/pre_playbook_0",
                                        "options": "ro,nosuid,nodev,relatime,discard,errors=remount-ro,bind",
                                        "size_available": 36193628160,
                                        "size_total": 166304071680,
                                        "uuid": "N/A"
                                    },
                                    {
                                        "block_available": 8836335,
                                        "block_size": 4096,
                                        "block_total": 40601580,
                                        "block_used": 31765245,
                                        "device": "/dev/vda1",
                                        "fstype": "ext4",
                                        "inode_available": 16499355,
                                        "inode_total": 20643840,
                                        "inode_used": 4144485,
                                        "mount": "/etc/zuul/site-variables.yaml",
                                        "options": "ro,nosuid,nodev,relatime,discard,errors=remount-ro,bind",
                                        "size_available": 36193628160,
                                        "size_total": 166304071680,
                                        "uuid": "N/A"
                                    },
                                    {
                                        "block_available": 203793654,
                                        "block_size": 4096,
                                        "block_total": 263940717,
                                        "block_used": 60147063,
                                        "device": "/dev/vdb",
                                        "fstype": "ext4",
                                        "inode_available": 61013274,
                                        "inode_total": 67108864,
                                        "inode_used": 6095590,
                                        "mount": "/srv/static/logs",
                                        "options": "rw,nosuid,nodev,relatime,discard",
                                        "size_available": 834738806784,
                                        "size_total": 1081101176832,
                                        "uuid": "N/A"
                                    },
                                    {
                                        "block_available": 8836303,
                                        "block_size": 4096,
                                        "block_total": 40601580,
                                        "block_used": 31765277,
                                        "device": "/dev/vda1",
                                        "fstype": "ext4",
                                        "inode_available": 16499355,
                                        "inode_total": 20643840,
                                        "inode_used": 4144485,
                                        "mount": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/.ansible",
                                        "options": "rw,nosuid,nodev,relatime,discard,errors=remount-ro,bind",
                                        "size_available": 36193497088,
                                        "size_total": 166304071680,
                                        "uuid": "N/A"
                                    }
                                ],
                                "ansible_nodename": "0a8996d2b663",
                                "ansible_os_family": "Debian",
                                "ansible_pkg_mgr": "apt",
                                "ansible_proc_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-5.15.0-130-generic",
                                    "console": [
                                        "tty1",
                                        "ttyS0"
                                    ],
                                    "ro": true,
                                    "root": "UUID=5a569d86-b935-46dd-ae79-7a72a25b6a4c"
                                },
                                "ansible_processor": [
                                    "0",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "1",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "2",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "3",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "4",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "5",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "6",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "7",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "8",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "9",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "10",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "11",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "12",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "13",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "14",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor",
                                    "15",
                                    "AuthenticAMD",
                                    "AMD EPYC-Rome Processor"
                                ],
                                "ansible_processor_cores": 1,
                                "ansible_processor_count": 16,
                                "ansible_processor_nproc": 16,
                                "ansible_processor_threads_per_core": 1,
                                "ansible_processor_vcpus": 16,
                                "ansible_product_name": "NA",
                                "ansible_product_serial": "NA",
                                "ansible_product_uuid": "NA",
                                "ansible_product_version": "NA",
                                "ansible_python": {
                                    "executable": "/usr/local/lib/zuul/ansible/9/bin/python",
                                    "has_sslcontext": true,
                                    "type": "cpython",
                                    "version": {
                                        "major": 3,
                                        "micro": 14,
                                        "minor": 11,
                                        "releaselevel": "final",
                                        "serial": 0
                                    },
                                    "version_info": [
                                        3,
                                        11,
                                        14,
                                        "final",
                                        0
                                    ]
                                },
                                "ansible_python_version": "3.11.14",
                                "ansible_real_group_id": 0,
                                "ansible_real_user_id": 0,
                                "ansible_selinux": {
                                    "status": "disabled"
                                },
                                "ansible_selinux_python_present": true,
                                "ansible_service_mgr": "bwrap",
                                "ansible_swapfree_mb": 0,
                                "ansible_swaptotal_mb": 0,
                                "ansible_system": "Linux",
                                "ansible_system_capabilities": "N/A",
                                "ansible_system_capabilities_enforced": "N/A",
                                "ansible_system_vendor": "NA",
                                "ansible_uptime_seconds": 38960899,
                                "ansible_user_dir": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work",
                                "ansible_user_gecos": "root",
                                "ansible_user_gid": 0,
                                "ansible_user_id": "root",
                                "ansible_user_shell": "/bin/bash",
                                "ansible_user_uid": 0,
                                "ansible_userspace_architecture": "x86_64",
                                "ansible_userspace_bits": "64",
                                "ansible_virtualization_role": "host",
                                "ansible_virtualization_tech_guest": [],
                                "ansible_virtualization_tech_host": [
                                    "kvm"
                                ],
                                "ansible_virtualization_type": "kvm",
                                "gather_subset": [
                                    "all"
                                ],
                                "module_setup": true
                            },
                            "changed": false,
                            "deprecations": [],
                            "warnings": []
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:10.042749Z",
                            "start": "2026-04-04T10:06:08.545811Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-00000000002a",
                        "name": "Gathering Facts"
                    }
                },
                {
                    "hosts": {
                        "localhost": {
                            "action": "include_role",
                            "changed": false,
                            "include_args": {
                                "name": "set-zuul-log-path-fact"
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000006",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:10.066854Z",
                            "start": "2026-04-04T10:06:10.049387Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-000000000008",
                        "name": "Setup log path fact"
                    }
                },
                {
                    "hosts": {
                        "localhost": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "zuul_log_path": "5b9/oss/5b9aaae51dd74c078827d3a79afb7078"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000054",
                        "name": "set-zuul-log-path-fact",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/set-zuul-log-path-fact"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:10.111869Z",
                            "start": "2026-04-04T10:06:10.082322Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-000000000056",
                        "name": "Set log path for a build"
                    }
                },
                {
                    "hosts": {
                        "localhost": {
                            "action": "debug",
                            "changed": false,
                            "msg": "# Job Information\nAnsible Version: 2.16.16\nJob: atmosphere-molecule-aio-ovn\nPipeline: check\nExecutor: 0a8996d2b663\nTriggered by: https://github.com/vexxhost/atmosphere/pull/3740\nEvent ID: b34789e0-2f82-11f1-8be9-cf24059efa35\n"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000006",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:10.159658Z",
                            "start": "2026-04-04T10:06:10.119753Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-000000000009",
                        "name": "Print job information"
                    }
                },
                {
                    "hosts": {
                        "localhost": {
                            "action": "debug",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "failed": false,
                                    "msg": "# Node Information\nInventory Hostname: instance\nHostname: np0000163469\nUsername: zuul\nDistro: Ubuntu 22.04\nProvider: yul1\nRegion: ca-ymq-1\nLabel: ubuntu-jammy-16\nProduct Name: OpenStack Nova\nInterface IP: 199.204.45.153\n",
                                    "zj_item": "instance"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000006",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:10.270463Z",
                            "start": "2026-04-04T10:06:10.163777Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-00000000000a",
                        "name": "Print node information"
                    }
                },
                {
                    "hosts": {
                        "localhost": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/zuul-info",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/zuul-info",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 493,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/zuul-info",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/zuul-info",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-00000000000d",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:11.135325Z",
                            "start": "2026-04-04T10:06:10.279781Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-00000000000f",
                        "name": "Ensure Zuul Ansible directory exists"
                    }
                },
                {
                    "hosts": {
                        "localhost": {
                            "action": "copy",
                            "changed": true,
                            "checksum": "01cffe988e04b8e3849b8a16d3cde443c420de32",
                            "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/zuul-info/inventory.yaml",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "inventory.yaml",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "01cffe988e04b8e3849b8a16d3cde443c420de32",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/zuul-info",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/.ansible/tmp/ansible-tmp-1775297171.174392-84-180951747330301/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "0028479c0799a00fb4535b17983206da",
                            "mode": "0644",
                            "owner": "root",
                            "size": 38004,
                            "src": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/.ansible/tmp/ansible-tmp-1775297171.174392-84-180951747330301/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-00000000000d",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:12.282443Z",
                            "start": "2026-04-04T10:06:11.141126Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-000000000010",
                        "name": "Copy ansible inventory to logs dir"
                    }
                }
            ]
        },
        {
            "play": {
                "duration": {
                    "end": "2026-04-04T10:06:20.504122Z",
                    "start": "2026-04-04T10:06:12.311354Z"
                },
                "id": "0242ac17-0011-9f12-8fac-000000000011",
                "name": "all"
            },
            "tasks": [
                {
                    "hosts": {
                        "instance": {
                            "action": "stat",
                            "changed": false,
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "checksum_algorithm": "sha1",
                                    "follow": false,
                                    "get_attributes": true,
                                    "get_checksum": true,
                                    "get_mime": true,
                                    "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/5b9aaae51dd74c078827d3a79afb7078_id_rsa"
                                }
                            },
                            "stat": {
                                "exists": false
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:12.730660Z",
                            "start": "2026-04-04T10:06:12.326490Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-000000000018",
                        "name": "Check to see if ssh key was already created for this build"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "create-key-and-replace.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:12.933712Z",
                            "start": "2026-04-04T10:06:12.742369Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-000000000019",
                        "name": "Create a new key in workspace based on build UUID"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-keygen",
                                "-t",
                                "rsa",
                                "-N",
                                "",
                                "-C",
                                "zuul-build-sshkey",
                                "-f",
                                "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/5b9aaae51dd74c078827d3a79afb7078_id_rsa",
                                "-b",
                                "3072"
                            ],
                            "delta": "0:00:00.430052",
                            "end": "2026-04-04 10:06:13.803910",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-keygen -t rsa -N '' -C 'zuul-build-sshkey' -f /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/5b9aaae51dd74c078827d3a79afb7078_id_rsa -b 3072",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-9f12-8fac-00000000009c-0-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-04-04 10:06:13.373858",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Generating public/private rsa key pair.\nYour identification has been saved in /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/5b9aaae51dd74c078827d3a79afb7078_id_rsa\nYour public key has been saved in /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/5b9aaae51dd74c078827d3a79afb7078_id_rsa.pub\nThe key fingerprint is:\nSHA256:kd2+prqKROJlIs07MOUFWqnKMOdFVecvSFz8ZSxgFGs zuul-build-sshkey\nThe key's randomart image is:\n+---[RSA 3072]----+\n|  o. ....oBo .   |\n| o...  . B.o. +  |\n|.....   = E..+   |\n|+=...  . + o.    |\n|B+*.+   S . o    |\n|.*.B       . .   |\n|  + .       o    |\n|   o .     o     |\n|    . ..oo.      |\n+----[SHA256]-----+",
                            "stdout_lines": [
                                "Generating public/private rsa key pair.",
                                "Your identification has been saved in /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/5b9aaae51dd74c078827d3a79afb7078_id_rsa",
                                "Your public key has been saved in /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/5b9aaae51dd74c078827d3a79afb7078_id_rsa.pub",
                                "The key fingerprint is:",
                                "SHA256:kd2+prqKROJlIs07MOUFWqnKMOdFVecvSFz8ZSxgFGs zuul-build-sshkey",
                                "The key's randomart image is:",
                                "+---[RSA 3072]----+",
                                "|  o. ....oBo .   |",
                                "| o...  . B.o. +  |",
                                "|.....   = E..+   |",
                                "|+=...  . + o.    |",
                                "|B+*.+   S . o    |",
                                "|.*.B       . .   |",
                                "|  + .       o    |",
                                "|   o .     o     |",
                                "|    . ..oo.      |",
                                "+----[SHA256]-----+"
                            ],
                            "zuul_log_id": "0242ac17-0011-9f12-8fac-00000000009c-0-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:13.847376Z",
                            "start": "2026-04-04T10:06:12.967465Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-00000000009c",
                        "name": "Create Temp SSH key"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "remote-linux.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:13.889125Z",
                            "start": "2026-04-04T10:06:13.853541Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-00000000009d",
                        "name": "Remote setup ssh keys (linux)"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "lineinfile",
                            "changed": false,
                            "false_condition": "zuul_build_sshkey_cleanup",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:13.959501Z",
                            "start": "2026-04-04T10:06:13.934929Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-0000000000bc",
                        "name": "Remove previously added zuul-build-sshkey"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC18iToJ05crK3mrjrruwsI+Y3C4kUryzDSqSq7sYq2/oqCcCKb+/jIju4ha8TiGO+rrZ/Fl3kQvUWVuRXiMrjy9mbJTsAj2byBhx/Krgd3LAHS0wSheCF/xmlpClSjBIUX/i+8Gen90424txbIUIe2PxBalZ0FtWG6DK3roNbJCvl5LyxbkOEQwyGG7ixNdLzJshSoFmhPn8ggqRB0A2MkvyKOgkhh+sA7kQ7kthhYCy/LHYxB30GK1+quozfXfx9NGHj9Q7KZwjzMLb17qzPVAp56Gdtv7M8eDWuwk6nvulD4tX0qQk/hFBJtbiQCD5M8eUpP/qYXumLg3oT1jtTSWfylL06EPDXm8XgI7kgWGulM9Zacr8ddH7Aeq9KBCL0NukORM+plLD8oLZqKPnObHHCNHkX1NVm6sLPyW9IJbNrhnJRijngy6yG2PAW85lv0czGP3b0Yb847IdptBU7AwbE1qq3kRsZwg5b3RD+4n5lbSHhCRsTCGBIasf3F5OM= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "present",
                                    "user": "zuul",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC18iToJ05crK3mrjrruwsI+Y3C4kUryzDSqSq7sYq2/oqCcCKb+/jIju4ha8TiGO+rrZ/Fl3kQvUWVuRXiMrjy9mbJTsAj2byBhx/Krgd3LAHS0wSheCF/xmlpClSjBIUX/i+8Gen90424txbIUIe2PxBalZ0FtWG6DK3roNbJCvl5LyxbkOEQwyGG7ixNdLzJshSoFmhPn8ggqRB0A2MkvyKOgkhh+sA7kQ7kthhYCy/LHYxB30GK1+quozfXfx9NGHj9Q7KZwjzMLb17qzPVAp56Gdtv7M8eDWuwk6nvulD4tX0qQk/hFBJtbiQCD5M8eUpP/qYXumLg3oT1jtTSWfylL06EPDXm8XgI7kgWGulM9Zacr8ddH7Aeq9KBCL0NukORM+plLD8oLZqKPnObHHCNHkX1NVm6sLPyW9IJbNrhnJRijngy6yG2PAW85lv0czGP3b0Yb847IdptBU7AwbE1qq3kRsZwg5b3RD+4n5lbSHhCRsTCGBIasf3F5OM= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "present",
                            "user": "zuul",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:14.433071Z",
                            "start": "2026-04-04T10:06:13.968174Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-0000000000bd",
                        "name": "Enable access via build key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/home/zuul/.ssh"
                                },
                                "before": {
                                    "path": "/home/zuul/.ssh"
                                }
                            },
                            "gid": 1000,
                            "group": "zuul",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 448,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/home/zuul/.ssh",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0700",
                            "owner": "zuul",
                            "path": "/home/zuul/.ssh",
                            "size": 4096,
                            "state": "directory",
                            "uid": 1000
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:14.677209Z",
                            "start": "2026-04-04T10:06:14.441901Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-0000000000be",
                        "name": "Make sure user has a .ssh"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "copy",
                            "changed": true,
                            "checksum": "828ca506fa6f0e2d8abba6bce567ae0bba15b166",
                            "dest": "/home/zuul/.ssh/id_rsa",
                            "diff": [],
                            "gid": 1000,
                            "group": "zuul",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "5b9aaae51dd74c078827d3a79afb7078_id_rsa",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "828ca506fa6f0e2d8abba6bce567ae0bba15b166",
                                    "content": null,
                                    "dest": "/home/zuul/.ssh/id_rsa",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": false,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 384,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/home/zuul/.ansible/tmp/ansible-tmp-1775297174.720846-190-38587565683338/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "ba05648338dd962821a2e68efea9544b",
                            "mode": "0600",
                            "owner": "zuul",
                            "size": 2602,
                            "src": "/home/zuul/.ansible/tmp/ansible-tmp-1775297174.720846-190-38587565683338/source",
                            "state": "file",
                            "uid": 1000
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:15.341539Z",
                            "start": "2026-04-04T10:06:14.683277Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-0000000000bf",
                        "name": "Install build private key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "copy",
                            "changed": true,
                            "checksum": "3f2b0c7669d8fd5619a8784e986f30db57ba5c55",
                            "dest": "/home/zuul/.ssh/id_rsa.pub",
                            "diff": [],
                            "gid": 1000,
                            "group": "zuul",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "5b9aaae51dd74c078827d3a79afb7078_id_rsa.pub",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "3f2b0c7669d8fd5619a8784e986f30db57ba5c55",
                                    "content": null,
                                    "dest": "/home/zuul/.ssh/id_rsa.pub",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": false,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/home/zuul/.ansible/tmp/ansible-tmp-1775297175.3901675-200-35961837264081/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "6395b690cbbe8568d73b79c067186ca7",
                            "mode": "0644",
                            "owner": "zuul",
                            "size": 571,
                            "src": "/home/zuul/.ansible/tmp/ansible-tmp-1775297175.3901675-200-35961837264081/source",
                            "state": "file",
                            "uid": 1000
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:15.930678Z",
                            "start": "2026-04-04T10:06:15.350835Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-0000000000c0",
                        "name": "Install build public key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "include_tasks",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:15.965830Z",
                            "start": "2026-04-04T10:06:15.940864Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-00000000009e",
                        "name": "Remote setup ssh keys (windows)"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "sshagent_remove_keys",
                            "changed": true,
                            "invocation": {
                                "module_args": {
                                    "remove": "^(?!\\(stdin\\)).*"
                                }
                            },
                            "removed": [
                                "/etc/zuul/id_rsa"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-0000000000a3",
                        "name": "remove-zuul-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/remove-zuul-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:16.480404Z",
                            "start": "2026-04-04T10:06:16.057233Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-0000000000a5",
                        "name": "Remove master key from local agent"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-add",
                                "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/5b9aaae51dd74c078827d3a79afb7078_id_rsa"
                            ],
                            "delta": "0:00:00.015530",
                            "end": "2026-04-04 10:06:16.817339",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-add /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/5b9aaae51dd74c078827d3a79afb7078_id_rsa",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-9f12-8fac-0000000000ab-0-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-04-04 10:06:16.801809",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Identity added: /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/5b9aaae51dd74c078827d3a79afb7078_id_rsa (zuul-build-sshkey)",
                            "stdout_lines": [
                                "Identity added: /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/5b9aaae51dd74c078827d3a79afb7078_id_rsa (zuul-build-sshkey)"
                            ],
                            "zuul_log_id": "0242ac17-0011-9f12-8fac-0000000000ab-0-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:16.864813Z",
                            "start": "2026-04-04T10:06:16.560042Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-0000000000ab",
                        "name": "Add back temp key"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "ping",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "data": "pong"
                                }
                            },
                            "ping": "pong"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:17.154845Z",
                            "start": "2026-04-04T10:06:16.869615Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-0000000000ac",
                        "name": "Verify we can still SSH to all nodes"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "command",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:17.196302Z",
                            "start": "2026-04-04T10:06:17.162176Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-0000000000ad",
                        "name": "Verify we can still SSH to all nodes (windows)"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "zuul_console",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "path": "/tmp/console-{log_uuid}.log",
                                    "port": 19885,
                                    "state": "present"
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-00000000001b",
                        "name": "prepare-workspace",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/prepare-workspace"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:17.509874Z",
                            "start": "2026-04-04T10:06:17.214002Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-00000000001d",
                        "name": "Start zuul_console daemon."
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "synchronize",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": true
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-00000000001b",
                        "name": "prepare-workspace",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/prepare-workspace"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:19.401864Z",
                            "start": "2026-04-04T10:06:17.528607Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-00000000001e",
                        "name": "Synchronize src repos to workspace directory."
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": false,
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": null,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/home/zuul/zuul-output/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "absent",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "path": "/home/zuul/zuul-output/logs",
                                    "state": "absent",
                                    "zj_output_dir": "logs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": false,
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": null,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/home/zuul/zuul-output/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "absent",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "path": "/home/zuul/zuul-output/artifacts",
                                    "state": "absent",
                                    "zj_output_dir": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": false,
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": null,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/home/zuul/zuul-output/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "absent",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "path": "/home/zuul/zuul-output/docs",
                                    "state": "absent",
                                    "zj_output_dir": "docs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000020",
                        "name": "ensure-output-dirs",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/ensure-output-dirs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:19.950450Z",
                            "start": "2026-04-04T10:06:19.413584Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-000000000022",
                        "name": "Empty Zuul Output directories by removing them"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "mode": "0755",
                                            "path": "/home/zuul/zuul-output/logs",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "mode": "0775",
                                            "path": "/home/zuul/zuul-output/logs",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1000,
                                    "group": "zuul",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/home/zuul/zuul-output/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul",
                                    "path": "/home/zuul/zuul-output/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1000,
                                    "zj_output_dir": "logs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "mode": "0755",
                                            "path": "/home/zuul/zuul-output/artifacts",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "mode": "0775",
                                            "path": "/home/zuul/zuul-output/artifacts",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1000,
                                    "group": "zuul",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/home/zuul/zuul-output/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul",
                                    "path": "/home/zuul/zuul-output/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1000,
                                    "zj_output_dir": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "mode": "0755",
                                            "path": "/home/zuul/zuul-output/docs",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "mode": "0775",
                                            "path": "/home/zuul/zuul-output/docs",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1000,
                                    "group": "zuul",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/home/zuul/zuul-output/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul",
                                    "path": "/home/zuul/zuul-output/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1000,
                                    "zj_output_dir": "docs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-9f12-8fac-000000000020",
                        "name": "ensure-output-dirs",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/ensure-output-dirs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:20.504122Z",
                            "start": "2026-04-04T10:06:19.957937Z"
                        },
                        "id": "0242ac17-0011-9f12-8fac-000000000024",
                        "name": "Ensure Zuul Output directories exist"
                    }
                }
            ]
        }
    ],
    "stats": {
        "instance": {
            "changed": 8,
            "failures": 0,
            "ignored": 0,
            "ok": 15,
            "rescued": 0,
            "skipped": 3,
            "unreachable": 0
        },
        "localhost": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 6,
            "rescued": 0,
            "skipped": 0,
            "unreachable": 0
        }
    },
    "trusted": true
},
{
    "branch": "main",
    "index": "1",
    "phase": "pre",
    "playbook": "github.com/vexxhost/zuul-jobs/playbooks/molecule/pre.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-04-04T10:06:24.350871Z",
                    "start": "2026-04-04T10:06:21.529224Z"
                },
                "id": "0242ac17-0011-6f17-2114-000000000002",
                "name": "all"
            },
            "tasks": [
                {
                    "hosts": {
                        "instance": {
                            "action": "ansible.builtin.unarchive",
                            "changed": true,
                            "dest": "/usr/local/bin",
                            "extract_results": {
                                "cmd": [
                                    "/usr/bin/tar",
                                    "--extract",
                                    "-C",
                                    "/usr/local/bin",
                                    "-z",
                                    "--show-transformed-names",
                                    "--strip-components=1",
                                    "-f",
                                    "/home/zuul/.ansible/tmp/ansible-tmp-1775297181.6115482-5-104280563669696/uv-x86_64-unknown-linux-gnum8m3c7rz.tar.gz"
                                ],
                                "err": "",
                                "out": "",
                                "rc": 0
                            },
                            "gid": 0,
                            "group": "root",
                            "handler": "TgzArchive",
                            "invocation": {
                                "module_args": {
                                    "attributes": null,
                                    "copy": true,
                                    "creates": "/usr/local/bin/uv",
                                    "decrypt": true,
                                    "dest": "/usr/local/bin",
                                    "exclude": [],
                                    "extra_opts": [
                                        "--strip-components=1"
                                    ],
                                    "group": null,
                                    "include": [],
                                    "io_buffer_size": 65536,
                                    "keep_newer": false,
                                    "list_files": false,
                                    "mode": null,
                                    "owner": null,
                                    "remote_src": true,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "https://github.com/astral-sh/uv/releases/download/0.8.13/uv-x86_64-unknown-linux-gnu.tar.gz",
                                    "unsafe_writes": false,
                                    "validate_certs": true
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "size": 4096,
                            "src": "/home/zuul/.ansible/tmp/ansible-tmp-1775297181.6115482-5-104280563669696/uv-x86_64-unknown-linux-gnum8m3c7rz.tar.gz",
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-6f17-2114-000000000005",
                        "name": "setup-uv",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_0/github.com/vexxhost/zuul-jobs/roles/setup-uv"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:23.791395Z",
                            "start": "2026-04-04T10:06:21.551067Z"
                        },
                        "id": "0242ac17-0011-6f17-2114-000000000007",
                        "name": "Extract archive"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "/usr/local/bin/uv",
                                "--version"
                            ],
                            "delta": "0:00:00.014090",
                            "end": "2026-04-04 10:06:24.190729",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "/usr/local/bin/uv --version",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-6f17-2114-000000000008-1-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-04-04 10:06:24.176639",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "uv 0.8.13",
                            "stdout_lines": [
                                "uv 0.8.13"
                            ],
                            "zuul_log_id": "0242ac17-0011-6f17-2114-000000000008-1-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-6f17-2114-000000000005",
                        "name": "setup-uv",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_0/github.com/vexxhost/zuul-jobs/roles/setup-uv"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:24.350871Z",
                            "start": "2026-04-04T10:06:23.834480Z"
                        },
                        "id": "0242ac17-0011-6f17-2114-000000000008",
                        "name": "Print version"
                    }
                }
            ]
        }
    ],
    "stats": {
        "instance": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 2,
            "rescued": 0,
            "skipped": 0,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "stable/zed",
    "index": "2",
    "phase": "pre",
    "playbook": "github.com/vexxhost/atmosphere/test-playbooks/molecule/pre.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-04-04T10:06:36.620344Z",
                    "start": "2026-04-04T10:06:25.343380Z"
                },
                "id": "0242ac17-0011-8a28-9266-000000000002",
                "name": "all"
            },
            "tasks": [
                {
                    "hosts": {
                        "instance": {
                            "action": "ansible.builtin.apt",
                            "cache_update_time": 1775297190,
                            "cache_updated": true,
                            "changed": true,
                            "diff": {},
                            "invocation": {
                                "module_args": {
                                    "allow_change_held_packages": false,
                                    "allow_downgrade": false,
                                    "allow_unauthenticated": false,
                                    "autoclean": false,
                                    "autoremove": false,
                                    "cache_valid_time": 0,
                                    "clean": false,
                                    "deb": null,
                                    "default_release": null,
                                    "dpkg_options": "force-confdef,force-confold",
                                    "fail_on_autoremove": false,
                                    "force": false,
                                    "force_apt_get": false,
                                    "install_recommends": null,
                                    "lock_timeout": 60,
                                    "name": "jq",
                                    "only_upgrade": false,
                                    "package": [
                                        "jq"
                                    ],
                                    "policy_rc_d": null,
                                    "purge": false,
                                    "state": "present",
                                    "update_cache": true,
                                    "update_cache_retries": 5,
                                    "update_cache_retry_max_delay": 12,
                                    "upgrade": null
                                }
                            },
                            "stderr": "debconf: delaying package configuration, since apt-utils is not installed\n",
                            "stderr_lines": [
                                "debconf: delaying package configuration, since apt-utils is not installed"
                            ],
                            "stdout": "Reading package lists...\nBuilding dependency tree...\nReading state information...\nThe following additional packages will be installed:\n  libjq1 libonig5\nThe following NEW packages will be installed:\n  jq libjq1 libonig5\n0 upgraded, 3 newly installed, 0 to remove and 0 not upgraded.\nNeed to get 357 kB of archives.\nAfter this operation, 1087 kB of additional disk space will be used.\nGet:1 http://nova.clouds.archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB]\nGet:2 http://nova.clouds.archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB]\nGet:3 http://nova.clouds.archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB]\nFetched 357 kB in 2s (224 kB/s)\nSelecting previously unselected package libonig5:amd64.\r\n(Reading database ... \r(Reading database ... 5%\r(Reading database ... 10%\r(Reading database ... 15%\r(Reading database ... 20%\r(Reading database ... 25%\r(Reading database ... 30%\r(Reading database ... 35%\r(Reading database ... 40%\r(Reading database ... 45%\r(Reading database ... 50%\r(Reading database ... 55%\r(Reading database ... 60%\r(Reading database ... 65%\r(Reading database ... 70%\r(Reading database ... 75%\r(Reading database ... 80%\r(Reading database ... 85%\r(Reading database ... 90%\r(Reading database ... 95%\r(Reading database ... 100%\r(Reading database ... 30906 files and directories currently installed.)\r\nPreparing to unpack .../libonig5_6.9.7.1-2build1_amd64.deb ...\r\nUnpacking libonig5:amd64 (6.9.7.1-2build1) ...\r\nSelecting previously unselected package libjq1:amd64.\r\nPreparing to unpack .../libjq1_1.6-2.1ubuntu3.1_amd64.deb ...\r\nUnpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ...\r\nSelecting previously unselected package jq.\r\nPreparing to unpack .../jq_1.6-2.1ubuntu3.1_amd64.deb ...\r\nUnpacking jq (1.6-2.1ubuntu3.1) ...\r\nSetting up libonig5:amd64 (6.9.7.1-2build1) ...\r\nSetting up libjq1:amd64 (1.6-2.1ubuntu3.1) ...\r\nSetting up jq (1.6-2.1ubuntu3.1) ...\r\nProcessing triggers for libc-bin (2.35-0ubuntu3.13) ...\r\n",
                            "stdout_lines": [
                                "Reading package lists...",
                                "Building dependency tree...",
                                "Reading state information...",
                                "The following additional packages will be installed:",
                                "  libjq1 libonig5",
                                "The following NEW packages will be installed:",
                                "  jq libjq1 libonig5",
                                "0 upgraded, 3 newly installed, 0 to remove and 0 not upgraded.",
                                "Need to get 357 kB of archives.",
                                "After this operation, 1087 kB of additional disk space will be used.",
                                "Get:1 http://nova.clouds.archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB]",
                                "Get:2 http://nova.clouds.archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB]",
                                "Get:3 http://nova.clouds.archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB]",
                                "Fetched 357 kB in 2s (224 kB/s)",
                                "Selecting previously unselected package libonig5:amd64.",
                                "(Reading database ... ",
                                "(Reading database ... 5%",
                                "(Reading database ... 10%",
                                "(Reading database ... 15%",
                                "(Reading database ... 20%",
                                "(Reading database ... 25%",
                                "(Reading database ... 30%",
                                "(Reading database ... 35%",
                                "(Reading database ... 40%",
                                "(Reading database ... 45%",
                                "(Reading database ... 50%",
                                "(Reading database ... 55%",
                                "(Reading database ... 60%",
                                "(Reading database ... 65%",
                                "(Reading database ... 70%",
                                "(Reading database ... 75%",
                                "(Reading database ... 80%",
                                "(Reading database ... 85%",
                                "(Reading database ... 90%",
                                "(Reading database ... 95%",
                                "(Reading database ... 100%",
                                "(Reading database ... 30906 files and directories currently installed.)",
                                "Preparing to unpack .../libonig5_6.9.7.1-2build1_amd64.deb ...",
                                "Unpacking libonig5:amd64 (6.9.7.1-2build1) ...",
                                "Selecting previously unselected package libjq1:amd64.",
                                "Preparing to unpack .../libjq1_1.6-2.1ubuntu3.1_amd64.deb ...",
                                "Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ...",
                                "Selecting previously unselected package jq.",
                                "Preparing to unpack .../jq_1.6-2.1ubuntu3.1_amd64.deb ...",
                                "Unpacking jq (1.6-2.1ubuntu3.1) ...",
                                "Setting up libonig5:amd64 (6.9.7.1-2build1) ...",
                                "Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ...",
                                "Setting up jq (1.6-2.1ubuntu3.1) ...",
                                "Processing triggers for libc-bin (2.35-0ubuntu3.13) ..."
                            ]
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:36.620344Z",
                            "start": "2026-04-04T10:06:25.357159Z"
                        },
                        "id": "0242ac17-0011-8a28-9266-000000000004",
                        "name": "Install \"jq\" for log collection"
                    }
                }
            ]
        }
    ],
    "stats": {
        "instance": {
            "changed": 1,
            "failures": 0,
            "ignored": 0,
            "ok": 1,
            "rescued": 0,
            "skipped": 0,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "0",
    "phase": "run",
    "playbook": "github.com/vexxhost/zuul-jobs/playbooks/molecule/run.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-04-04T10:24:59.386986Z",
                    "start": "2026-04-04T10:06:37.404927Z"
                },
                "id": "0242ac17-0011-9b67-3b0e-000000000002",
                "name": "all"
            },
            "tasks": [
                {
                    "hosts": {
                        "instance": {
                            "action": "ansible.builtin.copy",
                            "changed": true,
                            "checksum": "01cffe988e04b8e3849b8a16d3cde443c420de32",
                            "dest": "src/github.com/vexxhost/atmosphere/inventory.yaml",
                            "diff": [],
                            "gid": 1000,
                            "group": "zuul",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "inventory.yaml",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "01cffe988e04b8e3849b8a16d3cde443c420de32",
                                    "content": null,
                                    "dest": "src/github.com/vexxhost/atmosphere",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/home/zuul/.ansible/tmp/ansible-tmp-1775297197.4627736-5-166178296132466/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "0028479c0799a00fb4535b17983206da",
                            "mode": "0644",
                            "owner": "zuul",
                            "size": 38004,
                            "src": "/home/zuul/.ansible/tmp/ansible-tmp-1775297197.4627736-5-166178296132466/source",
                            "state": "file",
                            "uid": 1000
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:38.354255Z",
                            "start": "2026-04-04T10:06:37.417789Z"
                        },
                        "id": "0242ac17-0011-9b67-3b0e-000000000004",
                        "name": "Copy inventory file for Zuul"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "ansible.builtin.replace",
                            "changed": true,
                            "invocation": {
                                "module_args": {
                                    "after": null,
                                    "attributes": null,
                                    "backup": false,
                                    "before": null,
                                    "encoding": "utf-8",
                                    "group": null,
                                    "mode": null,
                                    "owner": null,
                                    "path": "src/github.com/vexxhost/atmosphere/inventory.yaml",
                                    "regexp": "(^\\s*ansible_host:\\s*).*$",
                                    "replace": "\\1\"{{ nodepool.private_ipv4 }}\"",
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "msg": "1 replacements made",
                            "rc": 0
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:06:38.701619Z",
                            "start": "2026-04-04T10:06:38.360697Z"
                        },
                        "id": "0242ac17-0011-9b67-3b0e-000000000005",
                        "name": "Switch \"ansible_host\" to private IP"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "uv",
                                "run",
                                "molecule",
                                "test",
                                "--destroy",
                                "never",
                                "-s",
                                "aio"
                            ],
                            "delta": "0:18:19.814348",
                            "end": "2026-04-04 10:24:58.908199",
                            "failed": true,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "uv run molecule test --destroy never -s aio",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": "src/github.com/vexxhost/atmosphere",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-9b67-3b0e-000000000006-1-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "non-zero return code",
                            "rc": 2,
                            "start": "2026-04-04 10:06:39.093851",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Using CPython 3.10.12 interpreter at: /usr/bin/python3\nCreating virtual environment at: .venv\n   Building atmosphere @ file:///home/zuul/src/github.com/vexxhost/atmosphere\nDownloading pygments (1.2MiB)\nDownloading cryptography (4.2MiB)\nDownloading setuptools (1.1MiB)\nDownloading rjsonnet (1.2MiB)\nDownloading openstacksdk (1.7MiB)\nDownloading netaddr (2.2MiB)\nDownloading ansible-core (2.1MiB)\nDownloading kubernetes (1.9MiB)\n   Building pyperclip==1.9.0\n Downloading rjsonnet\n Downloading netaddr\n Downloading pygments\n Downloading cryptography\n Downloading setuptools\n Downloading kubernetes\n Downloading ansible-core\n Downloading openstacksdk\n      Built pyperclip==1.9.0\n      Built atmosphere @ file:///home/zuul/src/github.com/vexxhost/atmosphere\nInstalled 79 packages in 44ms\nWARNING  Molecule scenarios should migrate to 'extensions/molecule'\nINFO     [aio > discovery] scenario test matrix: dependency, cleanup, destroy, syntax, create, prepare, converge, idempotence, side_effect, verify, cleanup, destroy\nINFO     [aio > prerun] Performing prerun with role_name_check=0...\nINFO     [aio > dependency] Executing\nWARNING  [aio > dependency] Missing roles requirements file: requirements.yml\nWARNING  [aio > dependency] Missing collections requirements file: collections.yml\nWARNING  [aio > dependency] Executed: 2 missing (Remove from test_sequence to suppress)\nINFO     [aio > cleanup] Executing\nWARNING  [aio > cleanup] Executed: Missing playbook (Remove from test_sequence to suppress)\nINFO     [aio > destroy] Executing\nWARNING  [aio > destroy] Skipping, '--destroy=never' requested.\nINFO     [aio > destroy] Executed: Successful\nINFO     [aio > syntax] Executing\n\nplaybook: /home/zuul/src/github.com/vexxhost/atmosphere/molecule/aio/converge.yml\nINFO     [aio > syntax] Executed: Successful\nINFO     [aio > create] Executing\nWARNING  [aio > create] Executed: Missing playbook (Remove from test_sequence to suppress)\nINFO     [aio > prepare] Executing\n\nPLAY [Prepare] *****************************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:07:25 +0000 (0:00:00.026)       0:00:00.026 ********\n[WARNING]: Platform linux on host instance is using the discovered Python\ninterpreter at /usr/bin/python3.10, but future installation of another Python\ninterpreter could change the meaning of that path. See\nhttps://docs.ansible.com/ansible-\ncore/2.17/reference_appendices/interpreter_discovery.html for more information.\nok: [instance]\n\nTASK [Configure short hostname] ************************************************\nSaturday 04 April 2026  10:07:26 +0000 (0:00:01.127)       0:00:01.153 ********\nchanged: [instance]\n\nTASK [Ensure hostname inside hosts file] ***************************************\nSaturday 04 April 2026  10:07:27 +0000 (0:00:00.656)       0:00:01.810 ********\n[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created\nwith a mode of 0700, this may cause issues when running as another user. To\navoid this, create the remote_tmp dir with the correct permissions manually\nchanged: [instance]\n\nTASK [Install \"dirmngr\" for GPG keyserver operations] **************************\nSaturday 04 April 2026  10:07:27 +0000 (0:00:00.284)       0:00:02.094 ********\nok: [instance]\n\nTASK [Purge \"snapd\" package] ***************************************************\nSaturday 04 April 2026  10:07:28 +0000 (0:00:01.133)       0:00:03.228 ********\nok: [instance]\n\nPLAY [Generate workspace for Atmosphere] ***************************************\n\nTASK [Create folders for workspace] ********************************************\nSaturday 04 April 2026  10:07:29 +0000 (0:00:00.673)       0:00:03.901 ********\nchanged: [localhost] => (item=group_vars)\nchanged: [localhost] => (item=group_vars/all)\nchanged: [localhost] => (item=group_vars/controllers)\nchanged: [localhost] => (item=group_vars/cephs)\nchanged: [localhost] => (item=group_vars/computes)\nchanged: [localhost] => (item=host_vars)\n\nPLAY [Generate Ceph control plane configuration for workspace] *****************\n\nTASK [Ensure the Ceph control plane configuration file exists] *****************\nSaturday 04 April 2026  10:07:30 +0000 (0:00:00.974)       0:00:04.875 ********\nchanged: [localhost]\n\nTASK [Load the current Ceph control plane configuration into a variable] *******\nSaturday 04 April 2026  10:07:30 +0000 (0:00:00.165)       0:00:05.041 ********\nok: [localhost]\n\nTASK [Generate Ceph control plane values for missing variables] ****************\nSaturday 04 April 2026  10:07:30 +0000 (0:00:00.029)       0:00:05.070 ********\nok: [localhost] => (item={'key': 'ceph_fsid', 'value': '9772977c-b199-5d3f-94e7-44f40285818e'})\nok: [localhost] => (item={'key': 'ceph_mon_public_network', 'value': '10.96.240.0/24'})\n\nTASK [Write new Ceph control plane configuration file to disk] *****************\nSaturday 04 April 2026  10:07:30 +0000 (0:00:00.050)       0:00:05.121 ********\nchanged: [localhost]\n\nPLAY [Generate Ceph OSD configuration for workspace] ***************************\n\nTASK [Ensure the Ceph OSDs configuration file exists] **************************\nSaturday 04 April 2026  10:07:31 +0000 (0:00:00.545)       0:00:05.666 ********\nchanged: [localhost]\n\nTASK [Load the current Ceph OSDs configuration into a variable] ****************\nSaturday 04 April 2026  10:07:31 +0000 (0:00:00.169)       0:00:05.836 ********\nok: [localhost]\n\nTASK [Generate Ceph OSDs values for missing variables] *************************\nSaturday 04 April 2026  10:07:31 +0000 (0:00:00.028)       0:00:05.864 ********\nok: [localhost] => (item={'key': 'ceph_osd_devices', 'value': ['/dev/vdb', '/dev/vdc', '/dev/vdd']})\n\nTASK [Write new Ceph OSDs configuration file to disk] **************************\nSaturday 04 April 2026  10:07:31 +0000 (0:00:00.031)       0:00:05.896 ********\nchanged: [localhost]\n\nPLAY [Generate Kubernetes configuration for workspace] *************************\n\nTASK [Ensure the Kubernetes configuration file exists] *************************\nSaturday 04 April 2026  10:07:31 +0000 (0:00:00.350)       0:00:06.246 ********\nchanged: [localhost]\n\nTASK [Load the current Kubernetes configuration into a variable] ***************\nSaturday 04 April 2026  10:07:32 +0000 (0:00:00.179)       0:00:06.426 ********\nok: [localhost]\n\nTASK [Generate Kubernetes values for missing variables] ************************\nSaturday 04 April 2026  10:07:32 +0000 (0:00:00.029)       0:00:06.455 ********\nok: [localhost] => (item={'key': 'kubernetes_hostname', 'value': '10.96.240.10'})\nok: [localhost] => (item={'key': 'kubernetes_keepalived_vrid', 'value': 42})\nok: [localhost] => (item={'key': 'kubernetes_keepalived_vip', 'value': '10.96.240.10'})\n\nTASK [Write new Kubernetes configuration file to disk] *************************\nSaturday 04 April 2026  10:07:32 +0000 (0:00:00.043)       0:00:06.499 ********\nchanged: [localhost]\n\nPLAY [Generate Keepalived configuration for workspace] *************************\n\nTASK [Ensure the Keeaplived configuration file exists] *************************\nSaturday 04 April 2026  10:07:32 +0000 (0:00:00.356)       0:00:06.855 ********\nchanged: [localhost]\n\nTASK [Load the current Keepalived configuration into a variable] ***************\nSaturday 04 April 2026  10:07:32 +0000 (0:00:00.170)       0:00:07.026 ********\nok: [localhost]\n\nTASK [Generate Keepalived values for missing variables] ************************\nSaturday 04 April 2026  10:07:32 +0000 (0:00:00.028)       0:00:07.054 ********\nok: [localhost] => (item={'key': 'keepalived_interface', 'value': 'br-ex'})\nok: [localhost] => (item={'key': 'keepalived_vip', 'value': '10.96.250.10'})\n\nTASK [Write new Keepalived configuration file to disk] *************************\nSaturday 04 April 2026  10:07:32 +0000 (0:00:00.040)       0:00:07.095 ********\nchanged: [localhost]\n\nPLAY [Generate endpoints for workspace] ****************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:07:33 +0000 (0:00:00.358)       0:00:07.454 ********\nok: [localhost]\n\nTASK [Ensure the endpoints file exists] ****************************************\nSaturday 04 April 2026  10:07:33 +0000 (0:00:00.682)       0:00:08.136 ********\nchanged: [localhost]\n\nTASK [Load the current endpoints into a variable] ******************************\nSaturday 04 April 2026  10:07:34 +0000 (0:00:00.170)       0:00:08.306 ********\nok: [localhost]\n\nTASK [Generate endpoint skeleton for missing variables] ************************\nSaturday 04 April 2026  10:07:34 +0000 (0:00:00.030)       0:00:08.336 ********\nok: [localhost] => (item=keycloak_host)\nok: [localhost] => (item=kube_prometheus_stack_grafana_host)\nok: [localhost] => (item=kube_prometheus_stack_alertmanager_host)\nok: [localhost] => (item=kube_prometheus_stack_prometheus_host)\nok: [localhost] => (item=openstack_helm_endpoints_region_name)\nok: [localhost] => (item=openstack_helm_endpoints_keystone_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_glance_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_cinder_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_placement_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_barbican_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_neutron_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_nova_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_nova_novnc_host)\nok: [localhost] => (item=openstack_helm_endpoints_ironic_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_designate_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_octavia_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_magnum_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_magnum_registry_host)\nok: [localhost] => (item=openstack_helm_endpoints_senlin_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_heat_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_heat_cfn_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_horizon_api_host)\nok: [localhost] => (item=openstack_helm_endpoints_rgw_host)\nok: [localhost] => (item=openstack_helm_endpoints_manila_api_host)\n\nTASK [Write new endpoints file to disk] ****************************************\nSaturday 04 April 2026  10:07:34 +0000 (0:00:00.817)       0:00:09.153 ********\nchanged: [localhost]\n\nTASK [Ensure the endpoints file exists] ****************************************\nSaturday 04 April 2026  10:07:35 +0000 (0:00:00.343)       0:00:09.497 ********\nchanged: [localhost]\n\nPLAY [Generate Neutron configuration for workspace] ****************************\n\nTASK [Ensure the Neutron configuration file exists] ****************************\nSaturday 04 April 2026  10:07:35 +0000 (0:00:00.180)       0:00:09.677 ********\nchanged: [localhost]\n\nTASK [Load the current Neutron configuration into a variable] ******************\nSaturday 04 April 2026  10:07:35 +0000 (0:00:00.173)       0:00:09.850 ********\nok: [localhost]\n\nTASK [Generate Neutron values for missing variables] ***************************\nSaturday 04 April 2026  10:07:35 +0000 (0:00:00.033)       0:00:09.884 ********\nok: [localhost] => (item={'key': 'neutron_networks', 'value': [{'name': 'public', 'external': True, 'shared': True, 'mtu_size': 1500, 'port_security_enabled': True, 'provider_network_type': 'flat', 'provider_physical_network': 'external', 'subnets': [{'name': 'public-subnet', 'cidr': '10.96.250.0/24', 'gateway_ip': '10.96.250.10', 'allocation_pool_start': '10.96.250.200', 'allocation_pool_end': '10.96.250.220', 'enable_dhcp': True}]}]})\n\nTASK [Write new Neutron configuration file to disk] ****************************\nSaturday 04 April 2026  10:07:35 +0000 (0:00:00.042)       0:00:09.926 ********\nchanged: [localhost]\n\nPLAY [Generate Nova configuration for workspace] *******************************\n\nTASK [Ensure the Nova configuration file exists] *******************************\nSaturday 04 April 2026  10:07:36 +0000 (0:00:00.349)       0:00:10.276 ********\nchanged: [localhost]\n\nTASK [Load the current Nova configuration into a variable] *********************\nSaturday 04 April 2026  10:07:36 +0000 (0:00:00.171)       0:00:10.447 ********\nok: [localhost]\n\nTASK [Generate Nova values for missing variables] ******************************\nSaturday 04 April 2026  10:07:36 +0000 (0:00:00.031)       0:00:10.478 ********\nok: [localhost] => (item={'key': 'nova_flavors', 'value': [{'name': 'm1.tiny', 'ram': 512, 'disk': 1, 'vcpus': 1}, {'name': 'm1.small', 'ram': 2048, 'disk': 20, 'vcpus': 1}, {'name': 'm1.medium', 'ram': 4096, 'disk': 40, 'vcpus': 2}, {'name': 'm1.large', 'ram': 8192, 'disk': 80, 'vcpus': 4}, {'name': 'm1.xlarge', 'ram': 16384, 'disk': 160, 'vcpus': 8}]})\n\nTASK [Write new Nova configuration file to disk] *******************************\nSaturday 04 April 2026  10:07:36 +0000 (0:00:00.046)       0:00:10.525 ********\nchanged: [localhost]\n\nPLAY [Generate secrets for workspace] ******************************************\n\nTASK [Ensure the secrets file exists] ******************************************\nSaturday 04 April 2026  10:07:36 +0000 (0:00:00.348)       0:00:10.873 ********\nchanged: [localhost]\n\nTASK [Load the current secrets into a variable] ********************************\nSaturday 04 April 2026  10:07:36 +0000 (0:00:00.166)       0:00:11.040 ********\nok: [localhost]\n\nTASK [Generate secrets for missing variables] **********************************\nSaturday 04 April 2026  10:07:36 +0000 (0:00:00.035)       0:00:11.076 ********\nok: [localhost] => (item=heat_auth_encryption_key)\nok: [localhost] => (item=keepalived_password)\nok: [localhost] => (item=keycloak_admin_password)\nok: [localhost] => (item=keycloak_database_password)\nok: [localhost] => (item=keystone_keycloak_client_secret)\nok: [localhost] => (item=keystone_oidc_crypto_passphrase)\nok: [localhost] => (item=kube_prometheus_stack_grafana_admin_password)\nok: [localhost] => (item=octavia_heartbeat_key)\nok: [localhost] => (item=openstack_helm_endpoints_rabbitmq_admin_password)\nok: [localhost] => (item=openstack_helm_endpoints_memcached_secret_key)\nok: [localhost] => (item=openstack_helm_endpoints_keystone_admin_password)\nok: [localhost] => (item=openstack_helm_endpoints_keystone_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_keystone_rabbitmq_password)\nok: [localhost] => (item=openstack_helm_endpoints_glance_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_glance_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_glance_rabbitmq_password)\nok: [localhost] => (item=openstack_helm_endpoints_cinder_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_cinder_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_cinder_rabbitmq_password)\nok: [localhost] => (item=openstack_helm_endpoints_placement_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_placement_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_barbican_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_barbican_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_neutron_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_neutron_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_neutron_rabbitmq_password)\nok: [localhost] => (item=openstack_helm_endpoints_neutron_metadata_secret)\nok: [localhost] => (item=openstack_helm_endpoints_nova_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_nova_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_nova_rabbitmq_password)\nok: [localhost] => (item=openstack_helm_endpoints_ironic_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_ironic_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_ironic_rabbitmq_password)\nok: [localhost] => (item=openstack_helm_endpoints_designate_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_designate_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_designate_rabbitmq_password)\nok: [localhost] => (item=openstack_helm_endpoints_octavia_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_octavia_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_octavia_rabbitmq_password)\nok: [localhost] => (item=openstack_helm_endpoints_magnum_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_magnum_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_magnum_rabbitmq_password)\nok: [localhost] => (item=openstack_helm_endpoints_senlin_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_senlin_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_senlin_rabbitmq_password)\nok: [localhost] => (item=openstack_helm_endpoints_heat_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_heat_trustee_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_heat_stack_user_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_heat_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_heat_rabbitmq_password)\nok: [localhost] => (item=openstack_helm_endpoints_horizon_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_tempest_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_openstack_exporter_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_rgw_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_manila_keystone_password)\nok: [localhost] => (item=openstack_helm_endpoints_manila_mariadb_password)\nok: [localhost] => (item=openstack_helm_endpoints_staffeln_mariadb_password)\n\nTASK [Generate base64 encoded secrets] *****************************************\nSaturday 04 April 2026  10:07:37 +0000 (0:00:00.417)       0:00:11.493 ********\nok: [localhost] => (item=barbican_kek)\n\nTASK [Generate temporary files for generating keys for missing variables] ******\nSaturday 04 April 2026  10:07:37 +0000 (0:00:00.053)       0:00:11.546 ********\nchanged: [localhost] => (item=manila_ssh_key)\nchanged: [localhost] => (item=nova_ssh_key)\n\nTASK [Generate SSH keys for missing variables] *********************************\nSaturday 04 April 2026  10:07:37 +0000 (0:00:00.410)       0:00:11.957 ********\nchanged: [localhost] => (item=manila_ssh_key)\nchanged: [localhost] => (item=nova_ssh_key)\n\nTASK [Set values for SSH keys] *************************************************\nSaturday 04 April 2026  10:07:40 +0000 (0:00:02.988)       0:00:14.945 ********\nok: [localhost] => (item=manila_ssh_key)\nok: [localhost] => (item=nova_ssh_key)\n\nTASK [Delete the temporary files generated for SSH keys] ***********************\nSaturday 04 April 2026  10:07:40 +0000 (0:00:00.053)       0:00:14.998 ********\nchanged: [localhost] => (item=manila_ssh_key)\nchanged: [localhost] => (item=nova_ssh_key)\n\nTASK [Write new secrets file to disk] ******************************************\nSaturday 04 April 2026  10:07:41 +0000 (0:00:00.318)       0:00:15.317 ********\nchanged: [localhost]\n\nTASK [Encrypt secrets file with Vault password] ********************************\nSaturday 04 April 2026  10:07:41 +0000 (0:00:00.348)       0:00:15.665 ********\nskipping: [localhost]\n\nPLAY [Setup networking] ********************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:07:41 +0000 (0:00:00.034)       0:00:15.699 ********\nok: [instance]\n\nTASK [Create bridge for management network] ************************************\nSaturday 04 April 2026  10:07:42 +0000 (0:00:00.681)       0:00:16.381 ********\nok: [instance]\n\nTASK [Create fake interface for management bridge] *****************************\nSaturday 04 April 2026  10:07:42 +0000 (0:00:00.362)       0:00:16.743 ********\nok: [instance]\n\nTASK [Assign dummy interface to management bridge] *****************************\nSaturday 04 April 2026  10:07:42 +0000 (0:00:00.206)       0:00:16.950 ********\nok: [instance]\n\nTASK [Assign IP address for management bridge] *********************************\nSaturday 04 April 2026  10:07:42 +0000 (0:00:00.186)       0:00:17.136 ********\nok: [instance]\n\nTASK [Bring up interfaces] *****************************************************\nSaturday 04 April 2026  10:07:43 +0000 (0:00:00.208)       0:00:17.345 ********\nok: [instance] => (item=br-mgmt)\nok: [instance] => (item=dummy0)\n\nPLAY [Create devices for Ceph] *************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:07:43 +0000 (0:00:00.386)       0:00:17.731 ********\nok: [instance]\n\nTASK [Install depedencies] *****************************************************\nSaturday 04 April 2026  10:07:44 +0000 (0:00:00.704)       0:00:18.436 ********\nchanged: [instance]\n\nTASK [Start up service] ********************************************************\nSaturday 04 April 2026  10:08:03 +0000 (0:00:19.794)       0:00:38.231 ********\nok: [instance]\n\nTASK [Generate lvm.conf] *******************************************************\nSaturday 04 April 2026  10:08:04 +0000 (0:00:00.544)       0:00:38.776 ********\nok: [instance]\n\nTASK [Write /etc/lvm/lvm.conf] *************************************************\nSaturday 04 April 2026  10:08:04 +0000 (0:00:00.203)       0:00:38.979 ********\nchanged: [instance]\n\nTASK [Get list of all loopback devices] ****************************************\nSaturday 04 April 2026  10:08:05 +0000 (0:00:00.441)       0:00:39.421 ********\nok: [instance]\n\nTASK [Fail if there is any existing loopback devices] **************************\nSaturday 04 April 2026  10:08:05 +0000 (0:00:00.198)       0:00:39.619 ********\nskipping: [instance]\n\nTASK [Create devices for Ceph] *************************************************\nSaturday 04 April 2026  10:08:05 +0000 (0:00:00.026)       0:00:39.646 ********\nchanged: [instance] => (item=osd0)\nchanged: [instance] => (item=osd1)\nchanged: [instance] => (item=osd2)\n\nTASK [Set permissions on loopback devices] *************************************\nSaturday 04 April 2026  10:08:05 +0000 (0:00:00.530)       0:00:40.176 ********\nchanged: [instance] => (item=osd0)\nchanged: [instance] => (item=osd1)\nchanged: [instance] => (item=osd2)\n\nTASK [Start loop devices] ******************************************************\nSaturday 04 April 2026  10:08:06 +0000 (0:00:00.525)       0:00:40.702 ********\nchanged: [instance] => (item=osd0)\nchanged: [instance] => (item=osd1)\nchanged: [instance] => (item=osd2)\n\nTASK [Create a volume group for each loop device] ******************************\nSaturday 04 April 2026  10:08:07 +0000 (0:00:00.729)       0:00:41.431 ********\nchanged: [instance] => (item=osd0)\nchanged: [instance] => (item=osd1)\nchanged: [instance] => (item=osd2)\n\nTASK [Create a logical volume for each loop device] ****************************\nSaturday 04 April 2026  10:08:10 +0000 (0:00:03.109)       0:00:44.540 ********\nchanged: [instance] => (item=ceph-instance-osd0)\nchanged: [instance] => (item=ceph-instance-osd1)\nchanged: [instance] => (item=ceph-instance-osd2)\n\nPLAY [controllers] *************************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:08:12 +0000 (0:00:01.851)       0:00:46.392 ********\nok: [instance]\n\nTASK [Set masquerade rule] *****************************************************\nSaturday 04 April 2026  10:08:13 +0000 (0:00:00.913)       0:00:47.305 ********\nchanged: [instance]\n\nPLAY RECAP *********************************************************************\ninstance                   : ok=24   changed=10   unreachable=0    failed=0    skipped=1    rescued=0    ignored=0\nlocalhost                  : ok=40   changed=21   unreachable=0    failed=0    skipped=1    rescued=0    ignored=0\n\nSaturday 04 April 2026  10:08:13 +0000 (0:00:00.443)       0:00:47.748 ********\n===============================================================================\nInstall depedencies ---------------------------------------------------- 19.79s\nCreate a volume group for each loop device ------------------------------ 3.11s\nGenerate SSH keys for missing variables --------------------------------- 2.99s\nCreate a logical volume for each loop device ---------------------------- 1.85s\nInstall \"dirmngr\" for GPG keyserver operations -------------------------- 1.13s\nGathering Facts --------------------------------------------------------- 1.13s\nCreate folders for workspace -------------------------------------------- 0.97s\nGathering Facts --------------------------------------------------------- 0.91s\nGenerate endpoint skeleton for missing variables ------------------------ 0.82s\nStart loop devices ------------------------------------------------------ 0.73s\nGathering Facts --------------------------------------------------------- 0.70s\nGathering Facts --------------------------------------------------------- 0.68s\nGathering Facts --------------------------------------------------------- 0.68s\nPurge \"snapd\" package --------------------------------------------------- 0.67s\nConfigure short hostname ------------------------------------------------ 0.66s\nWrite new Ceph control plane configuration file to disk ----------------- 0.55s\nStart up service -------------------------------------------------------- 0.54s\nCreate devices for Ceph ------------------------------------------------- 0.53s\nSet permissions on loopback devices ------------------------------------- 0.53s\nSet masquerade rule ----------------------------------------------------- 0.44s\nINFO     [aio > prepare] Executed: Successful\nINFO     [aio > converge] Executing\n\nPLAY [all] *********************************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:08:16 +0000 (0:00:00.065)       0:00:00.065 ********\n[WARNING]: Platform linux on host instance is using the discovered Python\ninterpreter at /usr/bin/python3.10, but future installation of another Python\ninterpreter could change the meaning of that path. See\nhttps://docs.ansible.com/ansible-\ncore/2.17/reference_appendices/interpreter_discovery.html for more information.\nok: [instance]\n\nTASK [Set a fact with the \"atmosphere_images\" for other plays] *****************\nSaturday 04 April 2026  10:08:17 +0000 (0:00:01.170)       0:00:01.236 ********\nok: [instance]\n\nPLAY [Deploy Ceph monitors & managers] *****************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:08:17 +0000 (0:00:00.173)       0:00:01.409 ********\nok: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:08:18 +0000 (0:00:00.904)       0:00:02.313 ********\nok: [instance]\n\nTASK [vexxhost.containers.package : Update state for tar] **********************\nSaturday 04 April 2026  10:08:18 +0000 (0:00:00.287)       0:00:02.601 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.directory : Create directory (/var/lib/downloads)] ***\nSaturday 04 April 2026  10:08:18 +0000 (0:00:00.044)       0:00:02.645 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:08:19 +0000 (0:00:00.289)       0:00:02.935 ********\nok: [instance] => {\n    \"msg\": \"https://github.com/opencontainers/runc/releases/download/v1.4.0/runc.amd64\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:08:19 +0000 (0:00:00.074)       0:00:03.010 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:08:19 +0000 (0:00:00.628)       0:00:03.639 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.package : Update state for tar] **********************\nSaturday 04 April 2026  10:08:19 +0000 (0:00:00.048)       0:00:03.687 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:08:19 +0000 (0:00:00.047)       0:00:03.735 ********\nok: [instance]\n\nTASK [vexxhost.containers.package : Update state for tar] **********************\nSaturday 04 April 2026  10:08:20 +0000 (0:00:00.199)       0:00:03.935 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:08:21 +0000 (0:00:01.197)       0:00:05.132 ********\nok: [instance] => {\n    \"msg\": \"https://github.com/containerd/containerd/releases/download/v2.2.0/containerd-2.2.0-linux-amd64.tar.gz\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:08:21 +0000 (0:00:00.068)       0:00:05.200 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:08:22 +0000 (0:00:00.677)       0:00:05.878 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.containerd : Install SELinux packages] ***************\nSaturday 04 April 2026  10:08:25 +0000 (0:00:03.845)       0:00:09.724 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Set SELinux to permissive at runtime] ***\nSaturday 04 April 2026  10:08:25 +0000 (0:00:00.030)       0:00:09.755 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Persist SELinux permissive mode] ********\nSaturday 04 April 2026  10:08:25 +0000 (0:00:00.036)       0:00:09.791 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Install AppArmor packages] **************\nSaturday 04 April 2026  10:08:25 +0000 (0:00:00.037)       0:00:09.829 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.containerd : Create systemd service file for containerd] ***\nSaturday 04 April 2026  10:08:31 +0000 (0:00:05.469)       0:00:15.298 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.containerd : Create folders for configuration] *******\nSaturday 04 April 2026  10:08:31 +0000 (0:00:00.524)       0:00:15.822 ********\nchanged: [instance] => (item={'path': '/etc/containerd'})\nchanged: [instance] => (item={'path': '/var/lib/containerd', 'mode': '0o700'})\nchanged: [instance] => (item={'path': '/run/containerd', 'mode': '0o711'})\nchanged: [instance] => (item={'path': '/run/containerd/io.containerd.grpc.v1.cri', 'mode': '0o700'})\nchanged: [instance] => (item={'path': '/run/containerd/io.containerd.sandbox.controller.v1.shim', 'mode': '0o700'})\n\nTASK [vexxhost.containers.containerd : Create containerd config file] **********\nSaturday 04 April 2026  10:08:32 +0000 (0:00:00.934)       0:00:16.757 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.containerd : Force any restarts if necessary] ********\nSaturday 04 April 2026  10:08:33 +0000 (0:00:00.561)       0:00:17.318 ********\n\nRUNNING HANDLER [vexxhost.containers.containerd : Reload systemd] **************\nSaturday 04 April 2026  10:08:33 +0000 (0:00:00.009)       0:00:17.328 ********\nok: [instance]\n\nRUNNING HANDLER [vexxhost.containers.containerd : Restart containerd] **********\nSaturday 04 April 2026  10:08:34 +0000 (0:00:00.904)       0:00:18.233 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.containerd : Enable and start service] ***************\nSaturday 04 April 2026  10:08:34 +0000 (0:00:00.445)       0:00:18.679 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:08:35 +0000 (0:00:00.584)       0:00:19.264 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:08:35 +0000 (0:00:00.204)       0:00:19.469 ********\nok: [instance] => {\n    \"msg\": \"https://download.docker.com/linux/static/stable/x86_64/docker-24.0.9.tgz\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:08:35 +0000 (0:00:00.057)       0:00:19.526 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:08:36 +0000 (0:00:00.917)       0:00:20.444 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.docker : Install AppArmor packages] ******************\nSaturday 04 April 2026  10:08:40 +0000 (0:00:04.359)       0:00:24.804 ********\nok: [instance]\n\nTASK [vexxhost.containers.docker : Ensure group \"docker\" exists] ***************\nSaturday 04 April 2026  10:08:41 +0000 (0:00:00.967)       0:00:25.771 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.docker : Create systemd service file for docker] *****\nSaturday 04 April 2026  10:08:42 +0000 (0:00:00.368)       0:00:26.139 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.docker : Create folders for configuration] ***********\nSaturday 04 April 2026  10:08:42 +0000 (0:00:00.449)       0:00:26.589 ********\nchanged: [instance] => (item={'path': '/etc/docker'})\nchanged: [instance] => (item={'path': '/var/lib/docker', 'mode': '0o710'})\nchanged: [instance] => (item={'path': '/run/docker', 'mode': '0o711'})\n\nTASK [vexxhost.containers.docker : Create systemd socket file for docker] ******\nSaturday 04 April 2026  10:08:43 +0000 (0:00:00.541)       0:00:27.131 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.docker : Create docker daemon config file] ***********\nSaturday 04 April 2026  10:08:43 +0000 (0:00:00.463)       0:00:27.594 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.docker : Force any restarts if necessary] ************\nSaturday 04 April 2026  10:08:44 +0000 (0:00:00.487)       0:00:28.082 ********\n\nRUNNING HANDLER [vexxhost.containers.containerd : Reload systemd] **************\nSaturday 04 April 2026  10:08:44 +0000 (0:00:00.008)       0:00:28.090 ********\nok: [instance]\n\nRUNNING HANDLER [vexxhost.containers.docker : Restart docker] ******************\nSaturday 04 April 2026  10:08:44 +0000 (0:00:00.715)       0:00:28.806 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.docker : Enable and start service] *******************\nSaturday 04 April 2026  10:08:46 +0000 (0:00:01.363)       0:00:30.170 ********\nchanged: [instance]\n\nTASK [vexxhost.ceph.cephadm : Gather variables for each operating system] ******\nSaturday 04 April 2026  10:08:46 +0000 (0:00:00.539)       0:00:30.709 ********\nok: [instance] => (item=/home/zuul/.ansible/collections/ansible_collections/vexxhost/ceph/roles/cephadm/vars/ubuntu-22.04.yml)\n\nTASK [vexxhost.ceph.cephadm : Install packages] ********************************\nSaturday 04 April 2026  10:08:46 +0000 (0:00:00.062)       0:00:30.771 ********\nchanged: [instance]\n\nTASK [vexxhost.ceph.cephadm : Ensure services are started] *********************\nSaturday 04 April 2026  10:08:52 +0000 (0:00:05.186)       0:00:35.958 ********\nok: [instance] => (item=chronyd)\nok: [instance] => (item=sshd)\n\nTASK [vexxhost.ceph.cephadm : Download \"cephadm\"] ******************************\nSaturday 04 April 2026  10:08:52 +0000 (0:00:00.635)       0:00:36.594 ********\nchanged: [instance]\n\nTASK [vexxhost.ceph.cephadm : Remove cephadm from old path] ********************\nSaturday 04 April 2026  10:08:53 +0000 (0:00:00.794)       0:00:37.388 ********\nok: [instance]\n\nTASK [vexxhost.ceph.cephadm : Ensure \"cephadm\" user is present] ****************\nSaturday 04 April 2026  10:08:53 +0000 (0:00:00.197)       0:00:37.586 ********\nchanged: [instance]\n\nTASK [vexxhost.ceph.cephadm : Allow \"cephadm\" user to have passwordless sudo] ***\nSaturday 04 April 2026  10:08:54 +0000 (0:00:00.539)       0:00:38.125 ********\nchanged: [instance]\n\nTASK [vexxhost.ceph.mon : Get `cephadm ls` status] *****************************\nSaturday 04 April 2026  10:08:54 +0000 (0:00:00.441)       0:00:38.567 ********\nok: [instance]\n\nTASK [vexxhost.ceph.mon : Parse the `cephadm ls` output] ***********************\nSaturday 04 April 2026  10:08:56 +0000 (0:00:01.679)       0:00:40.247 ********\nok: [instance]\n\nTASK [vexxhost.ceph.mon : Assimilate existing configs in `ceph.conf`] **********\nSaturday 04 April 2026  10:08:56 +0000 (0:00:00.054)       0:00:40.301 ********\nskipping: [instance]\n\nTASK [vexxhost.ceph.mon : Adopt monitor to cluster] ****************************\nSaturday 04 April 2026  10:08:56 +0000 (0:00:00.042)       0:00:40.344 ********\nskipping: [instance]\n\nTASK [vexxhost.ceph.mon : Adopt manager to cluster] ****************************\nSaturday 04 April 2026  10:08:56 +0000 (0:00:00.041)       0:00:40.386 ********\nskipping: [instance]\n\nTASK [vexxhost.ceph.mon : Enable \"cephadm\" mgr module] *************************\nSaturday 04 April 2026  10:08:56 +0000 (0:00:00.041)       0:00:40.428 ********\nskipping: [instance]\n\nTASK [vexxhost.ceph.mon : Set orchestrator backend to \"cephadm\"] ***************\nSaturday 04 April 2026  10:08:56 +0000 (0:00:00.038)       0:00:40.466 ********\nskipping: [instance]\n\nTASK [vexxhost.ceph.mon : Use `cephadm` user for cephadm] **********************\nSaturday 04 April 2026  10:08:56 +0000 (0:00:00.039)       0:00:40.506 ********\nskipping: [instance]\n\nTASK [vexxhost.ceph.mon : Generate \"cephadm\" key] ******************************\nSaturday 04 April 2026  10:08:56 +0000 (0:00:00.042)       0:00:40.548 ********\nskipping: [instance]\n\nTASK [vexxhost.ceph.mon : Set Ceph Monitor IP address] *************************\nSaturday 04 April 2026  10:08:56 +0000 (0:00:00.040)       0:00:40.589 ********\nok: [instance]\n\nTASK [vexxhost.ceph.mon : Check if any node is bootstrapped] *******************\nSaturday 04 April 2026  10:08:56 +0000 (0:00:00.115)       0:00:40.705 ********\nok: [instance] => (item=instance)\n\nTASK [vexxhost.ceph.mon : Select pre-existing bootstrap node if exists] ********\nSaturday 04 April 2026  10:08:57 +0000 (0:00:00.229)       0:00:40.934 ********\nok: [instance]\n\nTASK [vexxhost.ceph.mon : Bootstrap cluster] ***********************************\nSaturday 04 April 2026  10:08:57 +0000 (0:00:00.057)       0:00:40.991 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/ceph/roles/mon/tasks/bootstrap-ceph.yml for instance\n\nTASK [vexxhost.ceph.mon : Generate temporary file for \"ceph.conf\"] *************\nSaturday 04 April 2026  10:08:57 +0000 (0:00:00.081)       0:00:41.072 ********\nchanged: [instance]\n\nTASK [vexxhost.ceph.mon : Include extra configuration values] ******************\nSaturday 04 April 2026  10:08:57 +0000 (0:00:00.343)       0:00:41.416 ********\nchanged: [instance] => (item={'option': 'mon allow pool size one', 'section': 'global', 'value': True})\nchanged: [instance] => (item={'option': 'osd crush chooseleaf type', 'section': 'global', 'value': 0})\nchanged: [instance] => (item={'option': 'auth allow insecure global id reclaim', 'section': 'mon', 'value': False})\n\nTASK [vexxhost.ceph.mon : Run Bootstrap coomand] *******************************\nSaturday 04 April 2026  10:08:58 +0000 (0:00:00.682)       0:00:42.098 ********\nok: [instance]\n\nTASK [vexxhost.ceph.mon : Remove temporary file for \"ceph.conf\"] ***************\nSaturday 04 April 2026  10:11:16 +0000 (0:02:18.663)       0:03:00.762 ********\nchanged: [instance]\n\nTASK [vexxhost.ceph.mon : Set bootstrap node] **********************************\nSaturday 04 April 2026  10:11:17 +0000 (0:00:00.233)       0:03:00.996 ********\nok: [instance]\n\nTASK [Install Ceph host] *******************************************************\nSaturday 04 April 2026  10:11:17 +0000 (0:00:00.052)       0:03:01.049 ********\nincluded: vexxhost.ceph.cephadm_host for instance\n\nTASK [vexxhost.ceph.cephadm_host : Get public SSH key for \"cephadm\" user] ******\nSaturday 04 April 2026  10:11:17 +0000 (0:00:00.087)       0:03:01.136 ********\nok: [instance]\n\nTASK [vexxhost.ceph.cephadm_host : Set fact with public SSH key for \"cephadm\" user] ***\nSaturday 04 April 2026  10:11:19 +0000 (0:00:01.717)       0:03:02.854 ********\nok: [instance] => (item=instance)\n\nTASK [vexxhost.ceph.cephadm_host : Set authorized key for \"cephadm\"] ***********\nSaturday 04 April 2026  10:11:19 +0000 (0:00:00.066)       0:03:02.920 ********\nok: [instance]\n\nTASK [vexxhost.ceph.cephadm_host : Add new host to Ceph] ***********************\nSaturday 04 April 2026  10:11:19 +0000 (0:00:00.461)       0:03:03.381 ********\nok: [instance]\n\nTASK [vexxhost.ceph.mon : Configure \"mon\" label for monitors] ******************\nSaturday 04 April 2026  10:11:21 +0000 (0:00:02.391)       0:03:05.773 ********\nok: [instance]\n\nTASK [vexxhost.ceph.mon : Validate monitor exist] ******************************\nSaturday 04 April 2026  10:11:23 +0000 (0:00:01.955)       0:03:07.729 ********\nok: [instance]\n\nTASK [Install Ceph host] *******************************************************\nSaturday 04 April 2026  10:11:35 +0000 (0:00:11.173)       0:03:18.903 ********\nincluded: vexxhost.ceph.cephadm_host for instance\n\nTASK [vexxhost.ceph.cephadm_host : Get public SSH key for \"cephadm\" user] ******\nSaturday 04 April 2026  10:11:35 +0000 (0:00:00.109)       0:03:19.012 ********\nskipping: [instance]\n\nTASK [vexxhost.ceph.cephadm_host : Set fact with public SSH key for \"cephadm\" user] ***\nSaturday 04 April 2026  10:11:35 +0000 (0:00:00.057)       0:03:19.070 ********\nskipping: [instance] => (item=instance)\nskipping: [instance]\n\nTASK [vexxhost.ceph.cephadm_host : Set authorized key for \"cephadm\"] ***********\nSaturday 04 April 2026  10:11:35 +0000 (0:00:00.066)       0:03:19.137 ********\nok: [instance]\n\nTASK [vexxhost.ceph.cephadm_host : Add new host to Ceph] ***********************\nSaturday 04 April 2026  10:11:35 +0000 (0:00:00.262)       0:03:19.400 ********\nok: [instance]\n\nTASK [vexxhost.ceph.mgr : Configure \"mgr\" label for managers] ******************\nSaturday 04 April 2026  10:11:38 +0000 (0:00:02.920)       0:03:22.320 ********\nok: [instance]\n\nTASK [vexxhost.ceph.mgr : Validate manager exist] ******************************\nSaturday 04 April 2026  10:11:41 +0000 (0:00:02.780)       0:03:25.101 ********\nok: [instance]\n\nTASK [vexxhost.ceph.mgr : Enable the Ceph Manager prometheus module] ***********\nSaturday 04 April 2026  10:11:45 +0000 (0:00:04.581)       0:03:29.683 ********\nok: [instance]\n\nPLAY [Deploy Ceph OSDs] ********************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:11:48 +0000 (0:00:02.653)       0:03:32.336 ********\nok: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:11:49 +0000 (0:00:01.023)       0:03:33.359 ********\nok: [instance]\n\nTASK [vexxhost.containers.package : Update state for tar] **********************\nSaturday 04 April 2026  10:11:49 +0000 (0:00:00.219)       0:03:33.578 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.directory : Create directory (/var/lib/downloads)] ***\nSaturday 04 April 2026  10:11:49 +0000 (0:00:00.048)       0:03:33.627 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:11:50 +0000 (0:00:00.218)       0:03:33.845 ********\nok: [instance] => {\n    \"msg\": \"https://github.com/opencontainers/runc/releases/download/v1.4.0/runc.amd64\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:11:50 +0000 (0:00:00.060)       0:03:33.906 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:11:50 +0000 (0:00:00.304)       0:03:34.210 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.package : Update state for tar] **********************\nSaturday 04 April 2026  10:11:50 +0000 (0:00:00.042)       0:03:34.253 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:11:50 +0000 (0:00:00.043)       0:03:34.296 ********\nok: [instance]\n\nTASK [vexxhost.containers.package : Update state for tar] **********************\nSaturday 04 April 2026  10:11:50 +0000 (0:00:00.206)       0:03:34.503 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:11:51 +0000 (0:00:01.120)       0:03:35.623 ********\nok: [instance] => {\n    \"msg\": \"https://github.com/containerd/containerd/releases/download/v2.2.0/containerd-2.2.0-linux-amd64.tar.gz\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:11:51 +0000 (0:00:00.077)       0:03:35.700 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:11:52 +0000 (0:00:00.377)       0:03:36.078 ********\nok: [instance]\n\nTASK [vexxhost.containers.containerd : Install SELinux packages] ***************\nSaturday 04 April 2026  10:11:54 +0000 (0:00:01.975)       0:03:38.053 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Set SELinux to permissive at runtime] ***\nSaturday 04 April 2026  10:11:54 +0000 (0:00:00.027)       0:03:38.080 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Persist SELinux permissive mode] ********\nSaturday 04 April 2026  10:11:54 +0000 (0:00:00.031)       0:03:38.112 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Install AppArmor packages] **************\nSaturday 04 April 2026  10:11:54 +0000 (0:00:00.028)       0:03:38.140 ********\nok: [instance]\n\nTASK [vexxhost.containers.containerd : Create systemd service file for containerd] ***\nSaturday 04 April 2026  10:11:55 +0000 (0:00:01.167)       0:03:39.307 ********\nok: [instance]\n\nTASK [vexxhost.containers.containerd : Create folders for configuration] *******\nSaturday 04 April 2026  10:11:55 +0000 (0:00:00.470)       0:03:39.778 ********\nok: [instance] => (item={'path': '/etc/containerd'})\nok: [instance] => (item={'path': '/var/lib/containerd', 'mode': '0o700'})\nok: [instance] => (item={'path': '/run/containerd', 'mode': '0o711'})\nok: [instance] => (item={'path': '/run/containerd/io.containerd.grpc.v1.cri', 'mode': '0o700'})\nok: [instance] => (item={'path': '/run/containerd/io.containerd.sandbox.controller.v1.shim', 'mode': '0o700'})\n\nTASK [vexxhost.containers.containerd : Create containerd config file] **********\nSaturday 04 April 2026  10:11:56 +0000 (0:00:00.945)       0:03:40.724 ********\nok: [instance]\n\nTASK [vexxhost.containers.containerd : Force any restarts if necessary] ********\nSaturday 04 April 2026  10:11:57 +0000 (0:00:00.540)       0:03:41.265 ********\n\nTASK [vexxhost.containers.containerd : Enable and start service] ***************\nSaturday 04 April 2026  10:11:57 +0000 (0:00:00.007)       0:03:41.272 ********\nok: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:11:57 +0000 (0:00:00.372)       0:03:41.644 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:11:58 +0000 (0:00:00.220)       0:03:41.864 ********\nok: [instance] => {\n    \"msg\": \"https://download.docker.com/linux/static/stable/x86_64/docker-24.0.9.tgz\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:11:58 +0000 (0:00:00.062)       0:03:41.927 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:11:58 +0000 (0:00:00.374)       0:03:42.301 ********\nok: [instance]\n\nTASK [vexxhost.containers.docker : Install AppArmor packages] ******************\nSaturday 04 April 2026  10:12:01 +0000 (0:00:03.073)       0:03:45.375 ********\nok: [instance]\n\nTASK [vexxhost.containers.docker : Ensure group \"docker\" exists] ***************\nSaturday 04 April 2026  10:12:02 +0000 (0:00:01.174)       0:03:46.549 ********\nok: [instance]\n\nTASK [vexxhost.containers.docker : Create systemd service file for docker] *****\nSaturday 04 April 2026  10:12:02 +0000 (0:00:00.225)       0:03:46.774 ********\nok: [instance]\n\nTASK [vexxhost.containers.docker : Create folders for configuration] ***********\nSaturday 04 April 2026  10:12:03 +0000 (0:00:00.470)       0:03:47.244 ********\nok: [instance] => (item={'path': '/etc/docker'})\nok: [instance] => (item={'path': '/var/lib/docker', 'mode': '0o710'})\nok: [instance] => (item={'path': '/run/docker', 'mode': '0o711'})\n\nTASK [vexxhost.containers.docker : Create systemd socket file for docker] ******\nSaturday 04 April 2026  10:12:04 +0000 (0:00:00.593)       0:03:47.838 ********\nok: [instance]\n\nTASK [vexxhost.containers.docker : Create docker daemon config file] ***********\nSaturday 04 April 2026  10:12:04 +0000 (0:00:00.433)       0:03:48.272 ********\nok: [instance]\n\nTASK [vexxhost.containers.docker : Force any restarts if necessary] ************\nSaturday 04 April 2026  10:12:04 +0000 (0:00:00.428)       0:03:48.700 ********\n\nTASK [vexxhost.containers.docker : Enable and start service] *******************\nSaturday 04 April 2026  10:12:04 +0000 (0:00:00.007)       0:03:48.707 ********\nok: [instance]\n\nTASK [vexxhost.ceph.cephadm : Gather variables for each operating system] ******\nSaturday 04 April 2026  10:12:05 +0000 (0:00:00.362)       0:03:49.070 ********\nok: [instance] => (item=/home/zuul/.ansible/collections/ansible_collections/vexxhost/ceph/roles/cephadm/vars/ubuntu-22.04.yml)\n\nTASK [vexxhost.ceph.cephadm : Install packages] ********************************\nSaturday 04 April 2026  10:12:05 +0000 (0:00:00.064)       0:03:49.135 ********\nok: [instance]\n\nTASK [vexxhost.ceph.cephadm : Ensure services are started] *********************\nSaturday 04 April 2026  10:12:06 +0000 (0:00:01.120)       0:03:50.255 ********\nok: [instance] => (item=chronyd)\nok: [instance] => (item=sshd)\n\nTASK [vexxhost.ceph.cephadm : Download \"cephadm\"] ******************************\nSaturday 04 April 2026  10:12:07 +0000 (0:00:00.704)       0:03:50.960 ********\nok: [instance]\n\nTASK [vexxhost.ceph.cephadm : Remove cephadm from old path] ********************\nSaturday 04 April 2026  10:12:07 +0000 (0:00:00.800)       0:03:51.761 ********\nok: [instance]\n\nTASK [vexxhost.ceph.cephadm : Ensure \"cephadm\" user is present] ****************\nSaturday 04 April 2026  10:12:08 +0000 (0:00:00.215)       0:03:51.976 ********\nok: [instance]\n\nTASK [vexxhost.ceph.cephadm : Allow \"cephadm\" user to have passwordless sudo] ***\nSaturday 04 April 2026  10:12:08 +0000 (0:00:00.284)       0:03:52.260 ********\nok: [instance]\n\nTASK [vexxhost.ceph.osd : Get monitor status] **********************************\nSaturday 04 April 2026  10:12:08 +0000 (0:00:00.233)       0:03:52.493 ********\nok: [instance] => (item=instance)\n\nTASK [vexxhost.ceph.osd : Select admin host] ***********************************\nSaturday 04 April 2026  10:12:08 +0000 (0:00:00.242)       0:03:52.736 ********\nok: [instance]\n\nTASK [vexxhost.ceph.osd : Get `cephadm ls` status] *****************************\nSaturday 04 April 2026  10:12:08 +0000 (0:00:00.057)       0:03:52.794 ********\nok: [instance]\n\nTASK [vexxhost.ceph.osd : Parse the `cephadm ls` output] ***********************\nSaturday 04 April 2026  10:12:14 +0000 (0:00:05.365)       0:03:58.159 ********\nok: [instance]\n\nTASK [Install Ceph host] *******************************************************\nSaturday 04 April 2026  10:12:14 +0000 (0:00:00.058)       0:03:58.218 ********\nincluded: vexxhost.ceph.cephadm_host for instance\n\nTASK [vexxhost.ceph.cephadm_host : Get public SSH key for \"cephadm\" user] ******\nSaturday 04 April 2026  10:12:14 +0000 (0:00:00.067)       0:03:58.285 ********\nskipping: [instance]\n\nTASK [vexxhost.ceph.cephadm_host : Set fact with public SSH key for \"cephadm\" user] ***\nSaturday 04 April 2026  10:12:14 +0000 (0:00:00.050)       0:03:58.335 ********\nskipping: [instance] => (item=instance)\nskipping: [instance]\n\nTASK [vexxhost.ceph.cephadm_host : Set authorized key for \"cephadm\"] ***********\nSaturday 04 April 2026  10:12:14 +0000 (0:00:00.067)       0:03:58.403 ********\nok: [instance]\n\nTASK [vexxhost.ceph.cephadm_host : Add new host to Ceph] ***********************\nSaturday 04 April 2026  10:12:14 +0000 (0:00:00.264)       0:03:58.668 ********\nok: [instance]\n\nTASK [vexxhost.ceph.osd : Adopt OSDs to cluster] *******************************\nSaturday 04 April 2026  10:12:17 +0000 (0:00:02.205)       0:04:00.874 ********\nskipping: [instance]\n\nTASK [vexxhost.ceph.osd : Wait until OSD added to cephadm] *********************\nSaturday 04 April 2026  10:12:17 +0000 (0:00:00.036)       0:04:00.911 ********\nskipping: [instance]\n\nTASK [vexxhost.ceph.osd : Ensure all OSDs are non-legacy] **********************\nSaturday 04 April 2026  10:12:17 +0000 (0:00:00.041)       0:04:00.953 ********\nok: [instance]\n\nTASK [vexxhost.ceph.osd : Get `ceph-volume lvm list` status] *******************\nSaturday 04 April 2026  10:12:22 +0000 (0:00:05.367)       0:04:06.320 ********\nok: [instance]\n\nTASK [vexxhost.ceph.osd : Install OSDs] ****************************************\nSaturday 04 April 2026  10:12:35 +0000 (0:00:13.156)       0:04:19.477 ********\nok: [instance] => (item=/dev/ceph-instance-osd0/data)\nok: [instance] => (item=/dev/ceph-instance-osd1/data)\nok: [instance] => (item=/dev/ceph-instance-osd2/data)\n\nTASK [vexxhost.ceph.osd : Get mon dump] ****************************************\nSaturday 04 April 2026  10:14:38 +0000 (0:02:03.319)       0:06:22.797 ********\nok: [instance]\n\nTASK [vexxhost.ceph.osd : Mark require osd release] ****************************\nSaturday 04 April 2026  10:14:41 +0000 (0:00:02.775)       0:06:25.572 ********\nok: [instance]\n\nTASK [vexxhost.ceph.osd : Wait for all OSD to be running] **********************\nSaturday 04 April 2026  10:14:43 +0000 (0:00:02.088)       0:06:27.661 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/ceph/roles/osd/tasks/check-osds.yml for instance\n\nTASK [vexxhost.ceph.osd : Set the retry count] *********************************\nSaturday 04 April 2026  10:14:43 +0000 (0:00:00.094)       0:06:27.755 ********\nok: [instance]\n\nTASK [vexxhost.ceph.osd : Get `ceph orch ps`] **********************************\nSaturday 04 April 2026  10:14:43 +0000 (0:00:00.052)       0:06:27.807 ********\nok: [instance]\n\nTASK [vexxhost.ceph.osd : OSD daemon list] *************************************\nSaturday 04 April 2026  10:14:45 +0000 (0:00:01.910)       0:06:29.718 ********\nok: [instance]\n\nTASK [vexxhost.ceph.osd : Fail if any OSD not running] *************************\nSaturday 04 April 2026  10:14:45 +0000 (0:00:00.054)       0:06:29.772 ********\nskipping: [instance] => (item=1)\nskipping: [instance] => (item=1)\nskipping: [instance] => (item=1)\nskipping: [instance]\n\nTASK [vexxhost.ceph.osd : Fail if any duplicate OSD ID] ************************\nSaturday 04 April 2026  10:14:46 +0000 (0:00:00.054)       0:06:29.827 ********\nskipping: [instance]\n\nPLAY [all] *********************************************************************\n\nTASK [Ensure RBD kernel module is loaded] **************************************\nSaturday 04 April 2026  10:14:46 +0000 (0:00:00.055)       0:06:29.882 ********\nchanged: [instance]\n\nPLAY [all] *********************************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:14:46 +0000 (0:00:00.402)       0:06:30.285 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.sysctl : Configure sysctl values] ********************\nSaturday 04 April 2026  10:14:47 +0000 (0:00:01.018)       0:06:31.303 ********\nchanged: [instance] => (item={'name': 'fs.aio-max-nr', 'value': 1048576})\nchanged: [instance] => (item={'name': 'net.ipv4.tcp_timestamps', 'value': 0})\nchanged: [instance] => (item={'name': 'net.ipv4.tcp_sack', 'value': 1})\nchanged: [instance] => (item={'name': 'net.core.netdev_budget', 'value': 1000})\nchanged: [instance] => (item={'name': 'net.core.netdev_max_backlog', 'value': 250000})\nchanged: [instance] => (item={'name': 'net.core.rmem_max', 'value': 4194304})\nchanged: [instance] => (item={'name': 'net.core.wmem_max', 'value': 4194304})\nchanged: [instance] => (item={'name': 'net.core.rmem_default', 'value': 4194304})\nchanged: [instance] => (item={'name': 'net.core.wmem_default', 'value': 4194304})\nchanged: [instance] => (item={'name': 'net.core.optmem_max', 'value': 4194304})\nchanged: [instance] => (item={'name': 'net.ipv4.tcp_rmem', 'value': '4096 87380 4194304'})\nchanged: [instance] => (item={'name': 'net.ipv4.tcp_wmem', 'value': '4096 65536 4194304'})\nchanged: [instance] => (item={'name': 'net.ipv4.tcp_low_latency', 'value': 1})\nchanged: [instance] => (item={'name': 'net.ipv4.tcp_adv_win_scale', 'value': 1})\nchanged: [instance] => (item={'name': 'net.ipv4.neigh.default.gc_thresh1', 'value': 128})\nchanged: [instance] => (item={'name': 'net.ipv4.neigh.default.gc_thresh2', 'value': 28872})\nchanged: [instance] => (item={'name': 'net.ipv4.neigh.default.gc_thresh3', 'value': 32768})\nchanged: [instance] => (item={'name': 'net.ipv6.neigh.default.gc_thresh1', 'value': 128})\nchanged: [instance] => (item={'name': 'net.ipv6.neigh.default.gc_thresh2', 'value': 28872})\nchanged: [instance] => (item={'name': 'net.ipv6.neigh.default.gc_thresh3', 'value': 32768})\n\nTASK [vexxhost.atmosphere.ethtool : Create folder for persistent configuration] ***\nSaturday 04 April 2026  10:14:53 +0000 (0:00:05.917)       0:06:37.221 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.ethtool : Install persistent \"ethtool\" tuning] *******\nSaturday 04 April 2026  10:14:53 +0000 (0:00:00.207)       0:06:37.429 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.ethtool : Run \"ethtool\" tuning] **********************\nSaturday 04 April 2026  10:14:54 +0000 (0:00:00.471)       0:06:37.900 ********\nok: [instance]\n\nTASK [Set a fact with the \"atmosphere_images\" for other plays] *****************\nSaturday 04 April 2026  10:14:54 +0000 (0:00:00.273)       0:06:38.173 ********\nok: [instance]\n\nPLAY [Configure Kubernetes VIP] ************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:14:54 +0000 (0:00:00.054)       0:06:38.228 ********\nok: [instance]\n\nTASK [vexxhost.containers.directory : Create directory (/etc/kubernetes/manifests)] ***\nSaturday 04 April 2026  10:14:55 +0000 (0:00:00.971)       0:06:39.199 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kube_vip : Uninstall legacy HA stack] ****************\nSaturday 04 April 2026  10:14:55 +0000 (0:00:00.217)       0:06:39.416 ********\nok: [instance] => (item=/etc/keepalived/keepalived.conf)\nok: [instance] => (item=/etc/keepalived/check_apiserver.sh)\nok: [instance] => (item=/etc/kubernetes/manifests/keepalived.yaml)\nok: [instance] => (item=/etc/haproxy/haproxy.cfg)\nok: [instance] => (item=/etc/kubernetes/manifests/haproxy.yaml)\n\nTASK [vexxhost.kubernetes.kube_vip : Switch API server to run on port 6443] ****\nSaturday 04 April 2026  10:14:56 +0000 (0:00:00.922)       0:06:40.339 ********\nfailed: [instance] (item=/etc/kubernetes/manifests/kube-apiserver.yaml) => {\"ansible_loop_var\": \"item\", \"changed\": false, \"item\": \"/etc/kubernetes/manifests/kube-apiserver.yaml\", \"msg\": \"Path /etc/kubernetes/manifests/kube-apiserver.yaml does not exist !\", \"rc\": 257}\nfailed: [instance] (item=/etc/kubernetes/controller-manager.conf) => {\"ansible_loop_var\": \"item\", \"changed\": false, \"item\": \"/etc/kubernetes/controller-manager.conf\", \"msg\": \"Path /etc/kubernetes/controller-manager.conf does not exist !\", \"rc\": 257}\nfailed: [instance] (item=/etc/kubernetes/scheduler.conf) => {\"ansible_loop_var\": \"item\", \"changed\": false, \"item\": \"/etc/kubernetes/scheduler.conf\", \"msg\": \"Path /etc/kubernetes/scheduler.conf does not exist !\", \"rc\": 257}\n...ignoring\n\nTASK [vexxhost.kubernetes.kube_vip : Check if super-admin.conf exists] *********\nSaturday 04 April 2026  10:14:57 +0000 (0:00:00.591)       0:06:40.930 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kube_vip : Check if kubeadm has already run] *********\nSaturday 04 April 2026  10:14:57 +0000 (0:00:00.191)       0:06:41.122 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kube_vip : Set fact with KUBECONFIG path] ************\nSaturday 04 April 2026  10:14:57 +0000 (0:00:00.191)       0:06:41.314 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kube_vip : Set fact with KUBECONFIG path (with super-admin.conf)] ***\nSaturday 04 April 2026  10:14:57 +0000 (0:00:00.034)       0:06:41.348 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kube_vip : Upload Kubernetes manifest] ***************\nSaturday 04 April 2026  10:14:57 +0000 (0:00:00.045)       0:06:41.394 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kube_vip : Ensure kube-vip configuration file] *******\nSaturday 04 April 2026  10:14:58 +0000 (0:00:00.541)       0:06:41.935 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kube_vip : Flush handlers] ***************************\nSaturday 04 April 2026  10:14:58 +0000 (0:00:00.198)       0:06:42.133 ********\n\nPLAY [Install Kubernetes] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:14:58 +0000 (0:00:00.061)       0:06:42.195 ********\nok: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:14:59 +0000 (0:00:01.144)       0:06:43.339 ********\nok: [instance]\n\nTASK [vexxhost.containers.package : Update state for tar] **********************\nSaturday 04 April 2026  10:14:59 +0000 (0:00:00.405)       0:06:43.744 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.directory : Create directory (/var/lib/downloads)] ***\nSaturday 04 April 2026  10:14:59 +0000 (0:00:00.046)       0:06:43.791 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:15:00 +0000 (0:00:00.222)       0:06:44.013 ********\nok: [instance] => {\n    \"msg\": \"https://github.com/opencontainers/runc/releases/download/v1.4.0/runc.amd64\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:15:00 +0000 (0:00:00.057)       0:06:44.071 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:15:00 +0000 (0:00:00.348)       0:06:44.419 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:15:00 +0000 (0:00:00.051)       0:06:44.471 ********\nok: [instance]\n\nTASK [vexxhost.containers.package : Update state for tar] **********************\nSaturday 04 April 2026  10:15:00 +0000 (0:00:00.216)       0:06:44.688 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:15:02 +0000 (0:00:01.162)       0:06:45.850 ********\nok: [instance] => {\n    \"msg\": \"https://github.com/containerd/containerd/releases/download/v2.2.0/containerd-2.2.0-linux-amd64.tar.gz\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:15:02 +0000 (0:00:00.093)       0:06:45.943 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:15:02 +0000 (0:00:00.379)       0:06:46.323 ********\nok: [instance]\n\nTASK [vexxhost.containers.containerd : Install SELinux packages] ***************\nSaturday 04 April 2026  10:15:04 +0000 (0:00:02.189)       0:06:48.513 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Set SELinux to permissive at runtime] ***\nSaturday 04 April 2026  10:15:04 +0000 (0:00:00.037)       0:06:48.550 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Persist SELinux permissive mode] ********\nSaturday 04 April 2026  10:15:04 +0000 (0:00:00.039)       0:06:48.589 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Install AppArmor packages] **************\nSaturday 04 April 2026  10:15:04 +0000 (0:00:00.042)       0:06:48.632 ********\nok: [instance]\n\nTASK [vexxhost.containers.containerd : Create systemd service file for containerd] ***\nSaturday 04 April 2026  10:15:05 +0000 (0:00:01.195)       0:06:49.827 ********\nok: [instance]\n\nTASK [vexxhost.containers.containerd : Create folders for configuration] *******\nSaturday 04 April 2026  10:15:06 +0000 (0:00:00.451)       0:06:50.279 ********\nok: [instance] => (item={'path': '/etc/containerd'})\nok: [instance] => (item={'path': '/var/lib/containerd', 'mode': '0o700'})\nok: [instance] => (item={'path': '/run/containerd', 'mode': '0o711'})\nok: [instance] => (item={'path': '/run/containerd/io.containerd.grpc.v1.cri', 'mode': '0o700'})\nok: [instance] => (item={'path': '/run/containerd/io.containerd.sandbox.controller.v1.shim', 'mode': '0o700'})\n\nTASK [vexxhost.containers.containerd : Create containerd config file] **********\nSaturday 04 April 2026  10:15:07 +0000 (0:00:00.946)       0:06:51.226 ********\nok: [instance]\n\nTASK [vexxhost.containers.containerd : Force any restarts if necessary] ********\nSaturday 04 April 2026  10:15:07 +0000 (0:00:00.545)       0:06:51.771 ********\n\nTASK [vexxhost.containers.containerd : Enable and start service] ***************\nSaturday 04 April 2026  10:15:07 +0000 (0:00:00.006)       0:06:51.778 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes_upgrade_check : Retrieve the \"kubeadm-config\" ConfigMap] ***\nSaturday 04 April 2026  10:15:08 +0000 (0:00:00.376)       0:06:52.154 ********\nfatal: [instance]: FAILED! => {\"changed\": false, \"msg\": \"Failed to import the required Python library (kubernetes) on instance's Python /usr/bin/python3.10. Please read the module documentation and install it in the appropriate location. If the required library is installed, but Ansible is using the wrong Python interpreter, please consult the documentation on ansible_python_interpreter\"}\n...ignoring\n\nTASK [vexxhost.kubernetes.kubernetes_upgrade_check : Parse the ClusterConfiguration] ***\nSaturday 04 April 2026  10:15:08 +0000 (0:00:00.670)       0:06:52.825 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes_upgrade_check : Retrieve the current Kubernetes version] ***\nSaturday 04 April 2026  10:15:09 +0000 (0:00:00.036)       0:06:52.862 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes_upgrade_check : Extract major, minor, and patch versions] ***\nSaturday 04 April 2026  10:15:09 +0000 (0:00:00.033)       0:06:52.896 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes_upgrade_check : Fail if we're jumping more than one minor version] ***\nSaturday 04 April 2026  10:15:09 +0000 (0:00:00.039)       0:06:52.935 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes_upgrade_check : Set fact if we need to upgrade] ***\nSaturday 04 April 2026  10:15:09 +0000 (0:00:00.044)       0:06:52.979 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:15:09 +0000 (0:00:00.058)       0:06:53.037 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:15:09 +0000 (0:00:00.219)       0:06:53.256 ********\nok: [instance] => {\n    \"msg\": \"https://dl.k8s.io/release/v1.28.13/bin/linux/amd64/kubeadm\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:15:09 +0000 (0:00:00.043)       0:06:53.300 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:15:10 +0000 (0:00:00.743)       0:06:54.044 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:15:10 +0000 (0:00:00.049)       0:06:54.093 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:15:10 +0000 (0:00:00.213)       0:06:54.306 ********\nok: [instance] => {\n    \"msg\": \"https://dl.k8s.io/release/v1.28.13/bin/linux/amd64/kubectl\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:15:10 +0000 (0:00:00.047)       0:06:54.354 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:15:11 +0000 (0:00:00.873)       0:06:55.228 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Install SELinux packages] ***************\nSaturday 04 April 2026  10:15:11 +0000 (0:00:00.054)       0:06:55.282 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Set SELinux to permissive at runtime] ***\nSaturday 04 April 2026  10:15:11 +0000 (0:00:00.038)       0:06:55.320 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Persist SELinux permissive mode] ********\nSaturday 04 April 2026  10:15:11 +0000 (0:00:00.035)       0:06:55.356 ********\nskipping: [instance]\n\nTASK [vexxhost.containers.containerd : Install AppArmor packages] **************\nSaturday 04 April 2026  10:15:11 +0000 (0:00:00.027)       0:06:55.383 ********\nok: [instance]\n\nTASK [vexxhost.containers.containerd : Create systemd service file for containerd] ***\nSaturday 04 April 2026  10:15:12 +0000 (0:00:01.095)       0:06:56.479 ********\nok: [instance]\n\nTASK [vexxhost.containers.containerd : Create folders for configuration] *******\nSaturday 04 April 2026  10:15:13 +0000 (0:00:00.458)       0:06:56.937 ********\nok: [instance] => (item={'path': '/etc/containerd'})\nok: [instance] => (item={'path': '/var/lib/containerd', 'mode': '0o700'})\nok: [instance] => (item={'path': '/run/containerd', 'mode': '0o711'})\nok: [instance] => (item={'path': '/run/containerd/io.containerd.grpc.v1.cri', 'mode': '0o700'})\nok: [instance] => (item={'path': '/run/containerd/io.containerd.sandbox.controller.v1.shim', 'mode': '0o700'})\n\nTASK [vexxhost.containers.containerd : Create containerd config file] **********\nSaturday 04 April 2026  10:15:14 +0000 (0:00:00.949)       0:06:57.887 ********\nok: [instance]\n\nTASK [vexxhost.containers.containerd : Force any restarts if necessary] ********\nSaturday 04 April 2026  10:15:14 +0000 (0:00:00.504)       0:06:58.391 ********\n\nTASK [vexxhost.containers.containerd : Enable and start service] ***************\nSaturday 04 April 2026  10:15:14 +0000 (0:00:00.008)       0:06:58.400 ********\nok: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:15:14 +0000 (0:00:00.377)       0:06:58.778 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:15:15 +0000 (0:00:00.214)       0:06:58.992 ********\nok: [instance] => {\n    \"msg\": \"https://github.com/kubernetes-sigs/cri-tools/releases/download/v1.34.0/crictl-v1.34.0-linux-amd64.tar.gz\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:15:15 +0000 (0:00:00.061)       0:06:59.053 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:15:15 +0000 (0:00:00.574)       0:06:59.628 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:15:17 +0000 (0:00:01.425)       0:07:01.054 ********\nok: [instance] => {\n    \"msg\": \"https://github.com/kubernetes-sigs/cri-tools/releases/download/v1.34.0/critest-v1.34.0-linux-amd64.tar.gz\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:15:17 +0000 (0:00:00.060)       0:07:01.115 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:15:17 +0000 (0:00:00.544)       0:07:01.659 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.cri_tools : Create crictl config] ********************\nSaturday 04 April 2026  10:15:19 +0000 (0:00:01.646)       0:07:03.305 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.directory : Create directory (/opt/cni/bin)] *********\nSaturday 04 April 2026  10:15:19 +0000 (0:00:00.468)       0:07:03.774 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:15:20 +0000 (0:00:00.222)       0:07:03.997 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:15:20 +0000 (0:00:00.274)       0:07:04.271 ********\nok: [instance] => {\n    \"msg\": \"https://github.com/containernetworking/plugins/releases/download/v1.8.0/cni-plugins-linux-amd64-v1.8.0.tgz\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:15:20 +0000 (0:00:00.070)       0:07:04.342 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:15:21 +0000 (0:00:00.892)       0:07:05.234 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.cni_plugins : Gather variables for each operating system] ***\nSaturday 04 April 2026  10:15:24 +0000 (0:00:02.917)       0:07:08.151 ********\nok: [instance] => (item=/home/zuul/.ansible/collections/ansible_collections/vexxhost/containers/roles/cni_plugins/vars/debian.yml)\n\nTASK [vexxhost.containers.cni_plugins : Install additional packages] ***********\nSaturday 04 April 2026  10:15:24 +0000 (0:00:00.083)       0:07:08.235 ********\nok: [instance]\n\nTASK [vexxhost.containers.cni_plugins : Ensure IPv6 is enabled] ****************\nSaturday 04 April 2026  10:15:25 +0000 (0:00:01.062)       0:07:09.297 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.cni_plugins : Enable kernel modules on-boot] *********\nSaturday 04 April 2026  10:15:25 +0000 (0:00:00.208)       0:07:09.506 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.cni_plugins : Enable kernel modules in runtime] ******\nSaturday 04 April 2026  10:15:26 +0000 (0:00:00.476)       0:07:09.982 ********\nchanged: [instance] => (item=br_netfilter)\nok: [instance] => (item=ip_tables)\nchanged: [instance] => (item=ip6_tables)\nok: [instance] => (item=nf_conntrack)\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:15:26 +0000 (0:00:00.845)       0:07:10.828 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:15:27 +0000 (0:00:00.234)       0:07:11.062 ********\nok: [instance] => {\n    \"msg\": \"https://dl.k8s.io/release/v1.28.13/bin/linux/amd64/kubelet\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:15:27 +0000 (0:00:00.055)       0:07:11.118 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:15:28 +0000 (0:00:01.131)       0:07:12.250 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Gather variables for each operating system] ***\nSaturday 04 April 2026  10:15:28 +0000 (0:00:00.044)       0:07:12.294 ********\nok: [instance] => (item=/home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/kubelet/vars/debian.yml)\n\nTASK [vexxhost.kubernetes.kubelet : Install coreutils] *************************\nSaturday 04 April 2026  10:15:28 +0000 (0:00:00.065)       0:07:12.360 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Install additional packages] ***************\nSaturday 04 April 2026  10:15:28 +0000 (0:00:00.050)       0:07:12.410 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Configure sysctl values] *******************\nSaturday 04 April 2026  10:15:32 +0000 (0:00:04.314)       0:07:16.725 ********\nchanged: [instance] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\nchanged: [instance] => (item={'name': 'net.bridge.bridge-nf-call-iptables', 'value': 1})\nchanged: [instance] => (item={'name': 'net.bridge.bridge-nf-call-ip6tables', 'value': 1})\nchanged: [instance] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 0})\nchanged: [instance] => (item={'name': 'fs.inotify.max_queued_events', 'value': 1048576})\nchanged: [instance] => (item={'name': 'fs.inotify.max_user_instances', 'value': 8192})\nchanged: [instance] => (item={'name': 'fs.inotify.max_user_watches', 'value': 1048576})\n\nTASK [vexxhost.kubernetes.kubelet : Create folders for kubernetes configuration] ***\nSaturday 04 April 2026  10:15:36 +0000 (0:00:03.344)       0:07:20.069 ********\nchanged: [instance] => (item=/etc/systemd/system/kubelet.service.d)\nok: [instance] => (item=/etc/kubernetes)\nok: [instance] => (item=/etc/kubernetes/manifests)\n\nTASK [vexxhost.kubernetes.kubelet : Add kubelet systemd service config] ********\nSaturday 04 April 2026  10:15:36 +0000 (0:00:00.555)       0:07:20.625 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Add kubeadm dropin for kubelet systemd service config] ***\nSaturday 04 April 2026  10:15:37 +0000 (0:00:00.470)       0:07:21.096 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Check swap status] *************************\nSaturday 04 April 2026  10:15:37 +0000 (0:00:00.463)       0:07:21.559 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Disable swap] ******************************\nSaturday 04 April 2026  10:15:37 +0000 (0:00:00.204)       0:07:21.764 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Remove swapfile from /etc/fstab] ***********\nSaturday 04 April 2026  10:15:37 +0000 (0:00:00.038)       0:07:21.803 ********\nok: [instance] => (item=swap)\nok: [instance] => (item=none)\n\nTASK [vexxhost.kubernetes.kubelet : Create noswap systemd service config file] ***\nSaturday 04 April 2026  10:15:38 +0000 (0:00:00.595)       0:07:22.398 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Enable noswap service] *********************\nSaturday 04 April 2026  10:15:39 +0000 (0:00:00.481)       0:07:22.879 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Force any restarts if necessary] ***********\nSaturday 04 April 2026  10:15:39 +0000 (0:00:00.645)       0:07:23.524 ********\n\nRUNNING HANDLER [vexxhost.kubernetes.kubelet : Reload systemd] *****************\nSaturday 04 April 2026  10:15:39 +0000 (0:00:00.007)       0:07:23.532 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Enable and start kubelet service] **********\nSaturday 04 April 2026  10:15:40 +0000 (0:00:00.758)       0:07:24.290 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Ensure availability of dbus on Debian] *****\nSaturday 04 April 2026  10:15:41 +0000 (0:00:00.600)       0:07:24.890 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Configure short hostname] ******************\nSaturday 04 April 2026  10:15:42 +0000 (0:00:01.010)       0:07:25.901 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kubelet : Ensure hostname inside hosts file] *********\nSaturday 04 April 2026  10:15:42 +0000 (0:00:00.766)       0:07:26.667 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Remove kubernetes repository] ***********\nSaturday 04 April 2026  10:15:43 +0000 (0:00:00.246)       0:07:26.913 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Setup control plane] ********************\nSaturday 04 April 2026  10:15:43 +0000 (0:00:00.544)       0:07:27.458 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/kubernetes/tasks/control-plane.yml for instance\n\nTASK [vexxhost.kubernetes.kubernetes : Bootstrap cluster] **********************\nSaturday 04 April 2026  10:15:43 +0000 (0:00:00.079)       0:07:27.538 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/kubernetes/tasks/bootstrap-cluster.yml for instance\n\nTASK [vexxhost.kubernetes.kubernetes : Check if any control plane is bootstrapped] ***\nSaturday 04 April 2026  10:15:43 +0000 (0:00:00.083)       0:07:27.622 ********\nok: [instance] => (item=instance)\n\nTASK [vexxhost.kubernetes.kubernetes : Pick node from pre-existing cluster] ****\nSaturday 04 April 2026  10:15:44 +0000 (0:00:00.236)       0:07:27.858 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Select first node to initialize cluster] ***\nSaturday 04 April 2026  10:15:44 +0000 (0:00:00.039)       0:07:27.898 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Print selected bootstrap node] **********\nSaturday 04 April 2026  10:15:44 +0000 (0:00:00.051)       0:07:27.950 ********\nok: [instance] => {\n    \"msg\": \"instance\"\n}\n\nTASK [vexxhost.kubernetes.kubernetes : Upload cluster configuration for bootstrap node] ***\nSaturday 04 April 2026  10:15:44 +0000 (0:00:00.049)       0:07:27.999 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Create pki folder] **********************\nSaturday 04 April 2026  10:15:44 +0000 (0:00:00.566)       0:07:28.566 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Create kubernetes ca key] ***************\nSaturday 04 April 2026  10:15:44 +0000 (0:00:00.037)       0:07:28.603 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Create kubernetes ca cert] **************\nSaturday 04 April 2026  10:15:44 +0000 (0:00:00.032)       0:07:28.635 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Create etcd-ca key] *********************\nSaturday 04 April 2026  10:15:44 +0000 (0:00:00.039)       0:07:28.675 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Create etcd-ca cert] ********************\nSaturday 04 April 2026  10:15:44 +0000 (0:00:00.037)       0:07:28.712 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Create front-proxy-ca key] **************\nSaturday 04 April 2026  10:15:44 +0000 (0:00:00.044)       0:07:28.757 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Create front-proxy-ca cert] *************\nSaturday 04 April 2026  10:15:44 +0000 (0:00:00.041)       0:07:28.798 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Initialize cluster] *********************\nSaturday 04 April 2026  10:15:45 +0000 (0:00:00.040)       0:07:28.839 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Check if the node is already part of the cluster] ***\nSaturday 04 April 2026  10:16:14 +0000 (0:00:29.735)       0:07:58.574 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Join cluster] ***************************\nSaturday 04 April 2026  10:16:14 +0000 (0:00:00.200)       0:07:58.775 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Create folder for admin configuration] ***\nSaturday 04 April 2026  10:16:14 +0000 (0:00:00.042)       0:07:58.818 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Creating a symlink for admin configuration file] ***\nSaturday 04 April 2026  10:16:15 +0000 (0:00:00.202)       0:07:59.020 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Add bash autocomplete for kubectl] ******\nSaturday 04 April 2026  10:16:15 +0000 (0:00:00.220)       0:07:59.240 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Install PIP] ****************************\nSaturday 04 April 2026  10:16:15 +0000 (0:00:00.196)       0:07:59.437 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Install Kubernetes Python package using pip on supported systems] ***\nSaturday 04 April 2026  10:16:20 +0000 (0:00:05.003)       0:08:04.440 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Enable EPEL repository] *****************\nSaturday 04 April 2026  10:16:25 +0000 (0:00:04.696)       0:08:09.137 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Install Kubernetes Python package using package manager on supported systems] ***\nSaturday 04 April 2026  10:16:25 +0000 (0:00:00.046)       0:08:09.184 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Allow workload on control plane node] ***\nSaturday 04 April 2026  10:16:25 +0000 (0:00:00.036)       0:08:09.220 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Remove kube-proxy resources] ************\nSaturday 04 April 2026  10:16:26 +0000 (0:00:00.794)       0:08:10.015 ********\nskipping: [instance] => (item=DaemonSet)\nskipping: [instance] => (item=ConfigMap)\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Start an upgrade] ***********************\nSaturday 04 April 2026  10:16:26 +0000 (0:00:00.043)       0:08:10.059 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Check if the Kubernetes API services is up to date] ***\nSaturday 04 April 2026  10:16:26 +0000 (0:00:00.044)       0:08:10.104 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Trigger an upgrade of the Kubernetes API services] ***\nSaturday 04 April 2026  10:16:26 +0000 (0:00:00.039)       0:08:10.143 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Setup nodes] ****************************\nSaturday 04 April 2026  10:16:26 +0000 (0:00:00.046)       0:08:10.190 ********\nskipping: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Upgrade & restart Kubelet node for upgrade (if needed)] ***\nSaturday 04 April 2026  10:16:26 +0000 (0:00:00.039)       0:08:10.229 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.kubernetes : Set node selector for CoreDNS components] ***\nSaturday 04 April 2026  10:16:26 +0000 (0:00:00.521)       0:08:10.751 ********\nchanged: [instance]\n\nPLAY [Install control-plane components] ****************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:16:27 +0000 (0:00:00.951)       0:08:11.703 ********\nok: [instance]\n\nTASK [vexxhost.containers.forget_package : Forget package] *********************\nSaturday 04 April 2026  10:16:28 +0000 (0:00:00.969)       0:08:12.672 ********\nok: [instance]\n\nTASK [vexxhost.containers.package : Update state for tar] **********************\nSaturday 04 April 2026  10:16:29 +0000 (0:00:00.194)       0:08:12.867 ********\nok: [instance]\n\nTASK [vexxhost.containers.directory : Create directory (/var/lib/downloads)] ***\nSaturday 04 April 2026  10:16:30 +0000 (0:00:01.110)       0:08:13.977 ********\nok: [instance]\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:16:30 +0000 (0:00:00.459)       0:08:14.436 ********\nok: [instance] => {\n    \"msg\": \"https://get.helm.sh/helm-v3.11.2-linux-amd64.tar.gz\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:16:30 +0000 (0:00:00.042)       0:08:14.478 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:16:31 +0000 (0:00:00.514)       0:08:14.992 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.helm : Remove Helm repository] ***********************\nSaturday 04 April 2026  10:16:32 +0000 (0:00:01.474)       0:08:16.467 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.helm : Add bash autocomplete for helm] ***************\nSaturday 04 April 2026  10:16:32 +0000 (0:00:00.336)       0:08:16.803 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.helm : Get Helm plugins dir] *************************\nSaturday 04 April 2026  10:16:33 +0000 (0:00:00.190)       0:08:16.994 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.helm : Create Helm plugins directory if it does not exist] ***\nSaturday 04 April 2026  10:16:33 +0000 (0:00:00.245)       0:08:17.240 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.helm : Remove Helm diff plugin installed by kubernetes.core.helm_plugin] ***\nSaturday 04 April 2026  10:16:33 +0000 (0:00:00.200)       0:08:17.440 ********\nok: [instance]\n\nTASK [Install plugin] **********************************************************\nSaturday 04 April 2026  10:16:33 +0000 (0:00:00.195)       0:08:17.636 ********\nincluded: vexxhost.containers.download_artifact for instance\n\nTASK [vexxhost.containers.download_artifact : Starting download of file] *******\nSaturday 04 April 2026  10:16:33 +0000 (0:00:00.051)       0:08:17.688 ********\nok: [instance] => {\n    \"msg\": \"https://github.com/databus23/helm-diff/releases/download/v3.8.1/helm-diff-linux-amd64.tgz\"\n}\n\nTASK [vexxhost.containers.download_artifact : Download item] *******************\nSaturday 04 April 2026  10:16:33 +0000 (0:00:00.039)       0:08:17.728 ********\nchanged: [instance]\n\nTASK [vexxhost.containers.download_artifact : Extract archive] *****************\nSaturday 04 April 2026  10:16:34 +0000 (0:00:00.968)       0:08:18.696 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:16:36 +0000 (0:00:02.026)       0:08:20.722 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:16:36 +0000 (0:00:00.042)       0:08:20.765 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.cilium : Get current Kubernetes version] *************\nSaturday 04 April 2026  10:16:37 +0000 (0:00:00.634)       0:08:21.400 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.cilium : Deploy Helm chart] **************************\nSaturday 04 April 2026  10:16:38 +0000 (0:00:00.870)       0:08:22.271 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:16:41 +0000 (0:00:02.598)       0:08:24.869 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:16:41 +0000 (0:00:00.032)       0:08:24.902 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.envoy_gateway : Deploy Helm chart] *******************\nSaturday 04 April 2026  10:16:41 +0000 (0:00:00.481)       0:08:25.383 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.envoy_gateway : Create GatewayClass] *****************\nSaturday 04 April 2026  10:17:23 +0000 (0:00:42.116)       0:09:07.500 ********\nchanged: [instance]\n\nPLAY [all] *********************************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:17:24 +0000 (0:00:00.803)       0:09:08.303 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.kubernetes_node_labels : Add labels to node] *********\nSaturday 04 April 2026  10:17:25 +0000 (0:00:01.127)       0:09:09.430 ********\nchanged: [instance]\n\nPLAY [all] *********************************************************************\n\nTASK [Uninstall unattended-upgrades] *******************************************\nSaturday 04 April 2026  10:17:26 +0000 (0:00:00.688)       0:09:10.119 ********\nok: [instance]\n\nPLAY [controllers] *************************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:17:27 +0000 (0:00:00.915)       0:09:11.034 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:17:28 +0000 (0:00:01.222)       0:09:12.257 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:17:28 +0000 (0:00:00.045)       0:09:12.303 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.local_path_provisioner : Deploy Helm chart] **********\nSaturday 04 April 2026  10:17:28 +0000 (0:00:00.465)       0:09:12.768 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:17:30 +0000 (0:00:01.905)       0:09:14.673 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ceph_csi_rbd : Collect \"ceph mon dump\" output from a monitor] ***\nSaturday 04 April 2026  10:17:30 +0000 (0:00:00.051)       0:09:14.725 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ceph_csi_rbd : Generate fact with list of Ceph monitors] ***\nSaturday 04 April 2026  10:17:30 +0000 (0:00:00.050)       0:09:14.776 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ceph_csi_rbd : Create Ceph pool] *********************\nSaturday 04 April 2026  10:17:30 +0000 (0:00:00.036)       0:09:14.812 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ceph_csi_rbd : Create user client.kube] **************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.051)       0:09:14.863 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ceph_csi_rbd : Retrieve keyring for client.kube] *****\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.056)       0:09:14.920 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ceph_csi_rbd : Store keyring inside fact] ************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.057)       0:09:14.977 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ceph_csi_rbd : Suspend the existing HelmRelease] *****\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.043)       0:09:15.021 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ceph_csi_rbd : Remove the existing HelmRelease] ******\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.035)       0:09:15.057 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ceph_csi_rbd : Deploy Helm chart] ********************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.038)       0:09:15.095 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.powerstore_csi : Clone PowerStore CSI from GitHub] ***\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.049)       0:09:15.145 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.powerstore_csi : Create Secret] **********************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.041)       0:09:15.187 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.powerstore_csi : Create StorageClass] ****************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.039)       0:09:15.226 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.powerstore_csi : Deploy Helm chart] ******************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.036)       0:09:15.263 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.multipathd : Add backports PPA] **********************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.056)       0:09:15.319 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.multipathd : Install the multipathd package] *********\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.049)       0:09:15.369 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.multipathd : Install the configuration file] *********\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.042)       0:09:15.411 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.portworx : Detect if InitiatorName is set] ***********\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.047)       0:09:15.459 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.portworx : Generate a new InitiatorName] *************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.042)       0:09:15.501 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.portworx : Write the new InitiatorName] **************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.039)       0:09:15.540 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.portworx : Create namespace] *************************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.042)       0:09:15.582 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.portworx : Install Portworx] *************************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.040)       0:09:15.623 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.portworx : Wait till the CRDs are created] ***********\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.032)       0:09:15.655 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.portworx : Create Portworx Storage Cluster] **********\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.036)       0:09:15.691 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.storpool_csi : Deploy CSI RBAC] **********************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.057)       0:09:15.748 ********\nskipping: [instance] => (item={'name': 'controllerplugin'})\nskipping: [instance] => (item={'name': 'nodeplugin'})\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.storpool_csi : Deploy CSI] ***************************\nSaturday 04 April 2026  10:17:31 +0000 (0:00:00.043)       0:09:15.792 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.storpool_csi : Create StorageClass] ******************\nSaturday 04 April 2026  10:17:32 +0000 (0:00:00.038)       0:09:15.831 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ibm_block_csi_driver : Deploy CSI] *******************\nSaturday 04 April 2026  10:17:32 +0000 (0:00:00.051)       0:09:15.882 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ibm_block_csi_driver : Create Secret] ****************\nSaturday 04 April 2026  10:17:32 +0000 (0:00:00.033)       0:09:15.915 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.ibm_block_csi_driver : Create StorageClass] **********\nSaturday 04 April 2026  10:17:32 +0000 (0:00:00.033)       0:09:15.949 ********\nskipping: [instance]\n\nPLAY [Deploy Infrastructure] ***************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:17:32 +0000 (0:00:00.143)       0:09:16.093 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.cert_manager : Suspend the existing HelmRelease] *****\nSaturday 04 April 2026  10:17:33 +0000 (0:00:01.216)       0:09:17.309 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.cert_manager : Remove the existing HelmRelease] ******\nSaturday 04 April 2026  10:17:34 +0000 (0:00:00.658)       0:09:17.968 ********\nok: [instance]\n\nTASK [Deploy Helm chart] *******************************************************\nSaturday 04 April 2026  10:17:34 +0000 (0:00:00.740)       0:09:18.709 ********\nincluded: vexxhost.kubernetes.cert_manager for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:17:34 +0000 (0:00:00.063)       0:09:18.772 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:17:34 +0000 (0:00:00.054)       0:09:18.827 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.cert_manager : Deploy Helm chart] ********************\nSaturday 04 April 2026  10:17:35 +0000 (0:00:00.501)       0:09:19.328 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.cluster_issuer : Create self-signed cluster issuer] ***\nSaturday 04 April 2026  10:18:07 +0000 (0:00:31.837)       0:09:51.166 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.cluster_issuer : Import tasks for ClusterIssuer type] ***\nSaturday 04 April 2026  10:18:08 +0000 (0:00:00.706)       0:09:51.872 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/atmosphere/roles/cluster_issuer/tasks/type/self-signed/main.yml for instance\n\nTASK [vexxhost.atmosphere.cluster_issuer : Create ClusterIssuer] ***************\nSaturday 04 April 2026  10:18:08 +0000 (0:00:00.050)       0:09:51.922 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.cluster_issuer : Wait till the secret is created] ****\nSaturday 04 April 2026  10:18:08 +0000 (0:00:00.775)       0:09:52.698 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.cluster_issuer : Copy CA certificate on host] ********\nSaturday 04 April 2026  10:18:10 +0000 (0:00:01.868)       0:09:54.566 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.cluster_issuer : Flush all handlers] *****************\nSaturday 04 April 2026  10:18:11 +0000 (0:00:00.530)       0:09:55.097 ********\n\nRUNNING HANDLER [vexxhost.atmosphere.cluster_issuer : Update CA certificates on host] ***\nSaturday 04 April 2026  10:18:11 +0000 (0:00:00.012)       0:09:55.109 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:18:13 +0000 (0:00:01.738)       0:09:56.848 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:18:13 +0000 (0:00:00.046)       0:09:56.894 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.ingress_nginx : Remove the existing HelmRelease] *****\nSaturday 04 April 2026  10:18:13 +0000 (0:00:00.510)       0:09:57.404 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.ingress_nginx : Deploy Helm chart] *******************\nSaturday 04 April 2026  10:18:14 +0000 (0:00:00.786)       0:09:58.191 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:18:35 +0000 (0:00:21.074)       0:10:19.265 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:18:35 +0000 (0:00:00.044)       0:10:19.310 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.rabbitmq_cluster_operator : Suspend the existing HelmRelease] ***\nSaturday 04 April 2026  10:18:35 +0000 (0:00:00.462)       0:10:19.772 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.rabbitmq_cluster_operator : Remove the existing HelmRelease] ***\nSaturday 04 April 2026  10:18:36 +0000 (0:00:00.897)       0:10:20.670 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.rabbitmq_cluster_operator : Deploy Helm chart] *******\nSaturday 04 April 2026  10:18:37 +0000 (0:00:00.859)       0:10:21.530 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:18:43 +0000 (0:00:05.317)       0:10:26.848 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:18:43 +0000 (0:00:00.052)       0:10:26.901 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster_operator : Suspend the existing HelmRelease] ***\nSaturday 04 April 2026  10:18:43 +0000 (0:00:00.456)       0:10:27.357 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster_operator : Remove the existing HelmRelease] ***\nSaturday 04 April 2026  10:18:44 +0000 (0:00:00.685)       0:10:28.042 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster_operator : Deploy Helm chart] ***\nSaturday 04 April 2026  10:18:45 +0000 (0:00:00.799)       0:10:28.842 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Check if the Percona XtraDB cluster secret exists] ***\nSaturday 04 April 2026  10:18:49 +0000 (0:00:04.973)       0:10:33.815 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Create a secret] ************\nSaturday 04 April 2026  10:18:50 +0000 (0:00:00.755)       0:10:34.571 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Check if the Percona XtraDB cluster exists] ***\nSaturday 04 April 2026  10:18:52 +0000 (0:00:01.280)       0:10:35.852 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Get current status of the cluster] ***\nSaturday 04 April 2026  10:18:52 +0000 (0:00:00.717)       0:10:36.569 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Assert that the cluster is healthy before upgrade] ***\nSaturday 04 April 2026  10:18:52 +0000 (0:00:00.039)       0:10:36.609 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Stop PXC-operator] **********\nSaturday 04 April 2026  10:18:52 +0000 (0:00:00.044)       0:10:36.653 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Change the cluster Statefulset image to 8.0] ***\nSaturday 04 April 2026  10:18:52 +0000 (0:00:00.041)       0:10:36.695 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Wait until the cluster Statefulset rollout] ***\nSaturday 04 April 2026  10:18:52 +0000 (0:00:00.040)       0:10:36.736 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Update pxc cluster spec] ****\nSaturday 04 April 2026  10:18:52 +0000 (0:00:00.047)       0:10:36.784 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Start PXC-operator] *********\nSaturday 04 April 2026  10:18:52 +0000 (0:00:00.043)       0:10:36.827 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Apply Percona XtraDB cluster] ***\nSaturday 04 April 2026  10:18:53 +0000 (0:00:00.042)       0:10:36.869 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.percona_xtradb_cluster : Create percona haproxy metric service] ***\nSaturday 04 April 2026  10:20:12 +0000 (0:01:19.271)       0:11:56.140 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Retrieve list of all the needed endpoints] ***\nSaturday 04 April 2026  10:20:14 +0000 (0:00:01.788)       0:11:57.929 ********\nskipping: [instance]\n\nTASK [Create RabbitMQ cluster] *************************************************\nSaturday 04 April 2026  10:20:14 +0000 (0:00:00.046)       0:11:57.976 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Grab RabbitMQ cluster secret] ***\nSaturday 04 April 2026  10:20:14 +0000 (0:00:00.039)       0:11:58.016 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Cache fact with RabbitMQ cluster credentials] ***\nSaturday 04 April 2026  10:20:14 +0000 (0:00:00.035)       0:11:58.052 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Grab Percona XtraDB cluster secret] ***\nSaturday 04 April 2026  10:20:14 +0000 (0:00:00.039)       0:11:58.091 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Cache fact with Percona XtraDB password] ***\nSaturday 04 April 2026  10:20:15 +0000 (0:00:00.908)       0:11:58.999 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Reset value for OpenStack_Helm endpoints] ***\nSaturday 04 April 2026  10:20:15 +0000 (0:00:00.041)       0:11:59.041 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Generate OpenStack-Helm endpoints] ***\nSaturday 04 April 2026  10:20:15 +0000 (0:00:00.055)       0:11:59.097 ********\nok: [instance] => (item=oslo_db)\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Clean-up facts] ***********\nSaturday 04 April 2026  10:20:15 +0000 (0:00:00.045)       0:11:59.142 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:20:15 +0000 (0:00:00.060)       0:11:59.203 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:20:15 +0000 (0:00:00.048)       0:11:59.251 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.keycloak : Get the Kuberentes service for Percona XtraDB Cluster] ***\nSaturday 04 April 2026  10:20:15 +0000 (0:00:00.519)       0:11:59.771 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.keycloak : Install MySQL python package] *************\nSaturday 04 April 2026  10:20:16 +0000 (0:00:00.725)       0:12:00.496 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.keycloak : Check MySQL ready] ************************\nSaturday 04 April 2026  10:20:18 +0000 (0:00:02.032)       0:12:02.528 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.keycloak : Create Keycloak database] *****************\nSaturday 04 April 2026  10:20:19 +0000 (0:00:00.469)       0:12:02.997 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.keycloak : Create a Keycloak user] *******************\nSaturday 04 April 2026  10:20:19 +0000 (0:00:00.510)       0:12:03.508 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.keycloak : Disable pxc strict mode] ******************\nSaturday 04 April 2026  10:20:20 +0000 (0:00:00.512)       0:12:04.020 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.keycloak : Deploy Helm chart] ************************\nSaturday 04 April 2026  10:20:20 +0000 (0:00:00.539)       0:12:04.560 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.keycloak : Wait until keycloak ready] ****************\nSaturday 04 April 2026  10:22:24 +0000 (0:02:03.477)       0:14:08.037 ********\nok: [instance]\n\nTASK [Create Keycloak Ingress] *************************************************\nSaturday 04 April 2026  10:22:24 +0000 (0:00:00.742)       0:14:08.780 ********\nincluded: ingress for instance\n\nTASK [vexxhost.atmosphere.ingress : Create Ingress keycloak] *******************\nSaturday 04 April 2026  10:22:25 +0000 (0:00:00.060)       0:14:08.841 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.keycloak : Enable pxc strict mode] *******************\nSaturday 04 April 2026  10:22:25 +0000 (0:00:00.931)       0:14:09.772 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.keepalived : Deploy service] *************************\nSaturday 04 April 2026  10:22:26 +0000 (0:00:00.284)       0:14:10.057 ********\nchanged: [instance]\n\nPLAY [Deploy Monitoring] *******************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:22:27 +0000 (0:00:00.904)       0:14:10.962 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:22:28 +0000 (0:00:01.497)       0:14:12.459 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:22:28 +0000 (0:00:00.049)       0:14:12.508 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.node_feature_discovery : Suspend the existing HelmRelease] ***\nSaturday 04 April 2026  10:22:29 +0000 (0:00:00.515)       0:14:13.024 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.node_feature_discovery : Remove the existing HelmRelease] ***\nSaturday 04 April 2026  10:22:30 +0000 (0:00:00.870)       0:14:13.895 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.node_feature_discovery : Deploy Helm chart] **********\nSaturday 04 April 2026  10:22:30 +0000 (0:00:00.883)       0:14:14.779 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:22:34 +0000 (0:00:03.662)       0:14:18.441 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:22:34 +0000 (0:00:00.055)       0:14:18.496 ********\nchanged: [instance]\n\nTASK [atmosphere.common.secretgen_controller : Deploy secretgen-controller] ****\nSaturday 04 April 2026  10:22:35 +0000 (0:00:00.569)       0:14:19.065 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Uninstall the legacy \"ethtool-exporter\"] ***\nSaturday 04 April 2026  10:22:36 +0000 (0:00:01.450)       0:14:20.516 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Suspend the existing HelmRelease] ***\nSaturday 04 April 2026  10:22:37 +0000 (0:00:01.188)       0:14:21.704 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Remove the existing HelmRelease] ***\nSaturday 04 April 2026  10:22:38 +0000 (0:00:01.088)       0:14:22.792 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Wait until Keycloak service is ready] ***\nSaturday 04 April 2026  10:22:39 +0000 (0:00:00.814)       0:14:23.607 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Create Keycloak realm] *******\nSaturday 04 April 2026  10:22:40 +0000 (0:00:00.709)       0:14:24.317 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Add client roles in \"id_token\"] ***\nSaturday 04 April 2026  10:22:43 +0000 (0:00:03.141)       0:14:27.458 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Retrieve \"etcd\" CA certificate] ***\nSaturday 04 April 2026  10:22:44 +0000 (0:00:00.807)       0:14:28.266 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Retrieve \"etcd\" client certificate] ***\nSaturday 04 April 2026  10:22:44 +0000 (0:00:00.356)       0:14:28.622 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Retrieve \"etcd\" client key] ***\nSaturday 04 April 2026  10:22:45 +0000 (0:00:00.206)       0:14:28.829 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Create secrets for monitoring] ***\nSaturday 04 April 2026  10:22:45 +0000 (0:00:00.199)       0:14:29.029 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Generate client secret passwords] ***\nSaturday 04 April 2026  10:22:46 +0000 (0:00:00.982)       0:14:30.012 ********\nchanged: [instance] => (item=alertmanager)\nchanged: [instance] => (item=grafana)\nchanged: [instance] => (item=prometheus)\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Collect all client secrets] ***\nSaturday 04 April 2026  10:23:03 +0000 (0:00:17.453)       0:14:47.465 ********\nok: [instance] => (item=alertmanager)\nok: [instance] => (item=grafana)\nok: [instance] => (item=prometheus)\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Create Keycloak clients] *****\nSaturday 04 April 2026  10:23:05 +0000 (0:00:01.995)       0:14:49.461 ********\nchanged: [instance] => (item=None)\nchanged: [instance] => (item=None)\nchanged: [instance] => (item=None)\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Create Keycloak roles] *******\nSaturday 04 April 2026  10:23:07 +0000 (0:00:01.760)       0:14:51.222 ********\nchanged: [instance] => (item=None)\nchanged: [instance] => (item=None)\nchanged: [instance] => (item=None)\nchanged: [instance] => (item=None)\nchanged: [instance] => (item=None)\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Generate cookie secrets] *****\nSaturday 04 April 2026  10:23:10 +0000 (0:00:02.847)       0:14:54.069 ********\nchanged: [instance] => (item=alertmanager)\nchanged: [instance] => (item=prometheus)\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Generate OAuth2 proxy configuration] ***\nSaturday 04 April 2026  10:23:21 +0000 (0:00:11.624)       0:15:05.693 ********\nchanged: [instance] => (item=alertmanager)\nchanged: [instance] => (item=prometheus)\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Create certificate issuer] ***\nSaturday 04 April 2026  10:23:33 +0000 (0:00:11.853)       0:15:17.547 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Install all CRDs] ************\nSaturday 04 April 2026  10:23:34 +0000 (0:00:00.700)       0:15:18.247 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Deploy additional dashboards] ***\nSaturday 04 April 2026  10:23:41 +0000 (0:00:07.042)       0:15:25.290 ********\nchanged: [instance] => (item={'name': 'haproxy', 'state': 'present'})\nchanged: [instance] => (item={'name': 'goldpinger', 'state': 'present'})\nchanged: [instance] => (item={'name': 'node-exporter-full', 'state': 'present'})\nchanged: [instance] => (item={'name': 'ceph-cluster', 'state': 'present'})\nchanged: [instance] => (item={'name': 'ceph-cluster-advanced', 'state': 'present'})\nchanged: [instance] => (item={'name': 'hosts-overview', 'state': 'present'})\nchanged: [instance] => (item={'name': 'host-details', 'state': 'present'})\nchanged: [instance] => (item={'name': 'pool-overview', 'state': 'present'})\nchanged: [instance] => (item={'name': 'pool-detail', 'state': 'present'})\nchanged: [instance] => (item={'name': 'osds-overview', 'state': 'present'})\nchanged: [instance] => (item={'name': 'osd-device-details', 'state': 'present'})\nchanged: [instance] => (item={'name': 'rbd-overview', 'state': 'present'})\nchanged: [instance] => (item={'name': 'rbd-details', 'state': 'present'})\n\nTASK [vexxhost.atmosphere.kube_prometheus_stack : Deploy Helm chart] ***********\nSaturday 04 April 2026  10:23:57 +0000 (0:00:15.632)       0:15:40.923 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:24:30 +0000 (0:00:33.357)       0:16:14.280 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:24:30 +0000 (0:00:00.069)       0:16:14.350 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.loki : Deploy Helm chart] ****************************\nSaturday 04 April 2026  10:24:31 +0000 (0:00:00.584)       0:16:14.935 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:24:35 +0000 (0:00:04.252)       0:16:19.187 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:24:35 +0000 (0:00:00.053)       0:16:19.240 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.vector : Deploy Helm chart] **************************\nSaturday 04 April 2026  10:24:35 +0000 (0:00:00.530)       0:16:19.770 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:24:38 +0000 (0:00:02.076)       0:16:21.846 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:24:38 +0000 (0:00:00.062)       0:16:21.908 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.goldpinger : Deploy Helm chart] **********************\nSaturday 04 April 2026  10:24:38 +0000 (0:00:00.584)       0:16:22.493 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.ipmi_exporter : Deploy service] **********************\nSaturday 04 April 2026  10:24:43 +0000 (0:00:04.443)       0:16:26.936 ********\nchanged: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:24:43 +0000 (0:00:00.888)       0:16:27.825 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:24:44 +0000 (0:00:00.056)       0:16:27.881 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.prometheus_pushgateway : Suspend the existing HelmRelease] ***\nSaturday 04 April 2026  10:24:44 +0000 (0:00:00.482)       0:16:28.363 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.prometheus_pushgateway : Remove the existing HelmRelease] ***\nSaturday 04 April 2026  10:24:45 +0000 (0:00:00.760)       0:16:29.124 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.prometheus_pushgateway : Deploy Helm chart] **********\nSaturday 04 April 2026  10:24:46 +0000 (0:00:00.808)       0:16:29.933 ********\nchanged: [instance]\n\nPLAY [Deploy OpenStack] ********************************************************\n\nTASK [Gathering Facts] *********************************************************\nSaturday 04 April 2026  10:24:48 +0000 (0:00:02.224)       0:16:32.158 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Retrieve list of all the needed endpoints] ***\nSaturday 04 April 2026  10:24:50 +0000 (0:00:01.856)       0:16:34.014 ********\nok: [instance]\n\nTASK [Create RabbitMQ cluster] *************************************************\nSaturday 04 April 2026  10:24:50 +0000 (0:00:00.087)       0:16:34.102 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Grab RabbitMQ cluster secret] ***\nSaturday 04 April 2026  10:24:50 +0000 (0:00:00.047)       0:16:34.149 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Cache fact with RabbitMQ cluster credentials] ***\nSaturday 04 April 2026  10:24:50 +0000 (0:00:00.053)       0:16:34.202 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Grab Percona XtraDB cluster secret] ***\nSaturday 04 April 2026  10:24:50 +0000 (0:00:00.048)       0:16:34.251 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Cache fact with Percona XtraDB password] ***\nSaturday 04 April 2026  10:24:50 +0000 (0:00:00.045)       0:16:34.296 ********\nskipping: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Reset value for OpenStack_Helm endpoints] ***\nSaturday 04 April 2026  10:24:50 +0000 (0:00:00.048)       0:16:34.344 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Generate OpenStack-Helm endpoints] ***\nSaturday 04 April 2026  10:24:50 +0000 (0:00:00.059)       0:16:34.404 ********\nok: [instance] => (item=oslo_cache)\n\nTASK [vexxhost.atmosphere.openstack_helm_endpoints : Clean-up facts] ***********\nSaturday 04 April 2026  10:24:50 +0000 (0:00:00.064)       0:16:34.469 ********\nok: [instance]\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***\nSaturday 04 April 2026  10:24:50 +0000 (0:00:00.080)       0:16:34.549 ********\nincluded: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance\n\nTASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************\nSaturday 04 April 2026  10:24:50 +0000 (0:00:00.065)       0:16:34.615 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.memcached : Suspend the existing HelmRelease] ********\nSaturday 04 April 2026  10:24:51 +0000 (0:00:00.534)       0:16:35.150 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.memcached : Remove the existing HelmRelease] *********\nSaturday 04 April 2026  10:24:52 +0000 (0:00:01.256)       0:16:36.406 ********\nok: [instance]\n\nTASK [vexxhost.atmosphere.memcached : Deploy Helm chart] ***********************\nSaturday 04 April 2026  10:24:54 +0000 (0:00:01.903)       0:16:38.310 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.memcached : Apply manifests for monitoring] **********\nSaturday 04 April 2026  10:24:56 +0000 (0:00:01.965)       0:16:40.276 ********\nchanged: [instance]\n\nTASK [vexxhost.atmosphere.openstacksdk : Install openstacksdk] *****************\nSaturday 04 April 2026  10:24:57 +0000 (0:00:00.842)       0:16:41.118 ********\nfatal: [instance]: FAILED! => {\"changed\": false, \"cmd\": [\"/usr/bin/python3.10\", \"-m\", \"pip.__main__\", \"install\", \"openstacksdk==0.61.0\"], \"msg\": \"stdout: Collecting openstacksdk==0.61.0\\n  Downloading openstacksdk-0.61.0-py3-none-any.whl (1.4 MB)\\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501                             0.4/1.4 MB 9.3 MB/s eta 0:00:01\\n\\n:stderr: ERROR: THESE PACKAGES DO NOT MATCH THE HASHES FROM THE REQUIREMENTS FILE. If you have updated the package versions, please update the hashes. Otherwise, examine the package contents carefully; someone may have tampered with them.\\n    openstacksdk==0.61.0 from https://files.pythonhosted.org/packages/c2/31/a3501e87efc4ad83ab998502d681dde5fefbfe7d75225cabe268f815dad6/openstacksdk-0.61.0-py3-none-any.whl#sha256=9894d3d510563dcfc50c4755287dbfbf98def1f37caf2cfc15e9d0e1fd5d9a41:\\n        Expected sha256 9894d3d510563dcfc50c4755287dbfbf98def1f37caf2cfc15e9d0e1fd5d9a41\\n             Got        062b9b99adeb179395c6cf10283baad7165cce13a5f6f2aebc05254c78edf341\\n\\n\"}\n\nPLAY RECAP *********************************************************************\ninstance                   : ok=346  changed=123  unreachable=0    failed=1    skipped=107  rescued=0    ignored=2\n\nSaturday 04 April 2026  10:24:58 +0000 (0:00:01.201)       0:16:42.320 ********\n===============================================================================\nvexxhost.ceph.mon : Run Bootstrap coomand ----------------------------- 138.66s\nvexxhost.atmosphere.keycloak : Deploy Helm chart ---------------------- 123.48s\nvexxhost.ceph.osd : Install OSDs -------------------------------------- 123.32s\nvexxhost.atmosphere.percona_xtradb_cluster : Apply Percona XtraDB cluster -- 79.27s\nvexxhost.kubernetes.envoy_gateway : Deploy Helm chart ------------------ 42.12s\nvexxhost.atmosphere.kube_prometheus_stack : Deploy Helm chart ---------- 33.36s\nvexxhost.kubernetes.cert_manager : Deploy Helm chart ------------------- 31.84s\nvexxhost.kubernetes.kubernetes : Initialize cluster -------------------- 29.74s\nvexxhost.atmosphere.ingress_nginx : Deploy Helm chart ------------------ 21.07s\nvexxhost.atmosphere.kube_prometheus_stack : Generate client secret passwords -- 17.45s\nvexxhost.atmosphere.kube_prometheus_stack : Deploy additional dashboards -- 15.63s\nvexxhost.ceph.osd : Get `ceph-volume lvm list` status ------------------ 13.16s\nvexxhost.atmosphere.kube_prometheus_stack : Generate OAuth2 proxy configuration -- 11.85s\nvexxhost.atmosphere.kube_prometheus_stack : Generate cookie secrets ---- 11.62s\nvexxhost.ceph.mon : Validate monitor exist ----------------------------- 11.17s\nvexxhost.atmosphere.kube_prometheus_stack : Install all CRDs ------------ 7.04s\nvexxhost.atmosphere.sysctl : Configure sysctl values -------------------- 5.92s\nvexxhost.containers.containerd : Install AppArmor packages -------------- 5.47s\nvexxhost.ceph.osd : Ensure all OSDs are non-legacy ---------------------- 5.37s\nvexxhost.ceph.osd : Get `cephadm ls` status ----------------------------- 5.37s\nCRITICAL Ansible return code was 2, command was: ansible-playbook --inventory /home/zuul/.ansible/tmp/molecule.v9Wo.aio/inventory --skip-tags molecule-notest,notest --inventory=/home/zuul/src/github.com/vexxhost/atmosphere/inventory.yaml /home/zuul/src/github.com/vexxhost/atmosphere/molecule/aio/converge.yml\nERROR    [aio > converge] \u001b[31mExecuted: Failed\u001b[0m\nERROR    Ansible return code was 2, command was: ansible-playbook --inventory /home/zuul/.ansible/tmp/molecule.v9Wo.aio/inventory --skip-tags molecule-notest,notest --inventory=/home/zuul/src/github.com/vexxhost/atmosphere/inventory.yaml /home/zuul/src/github.com/vexxhost/atmosphere/molecule/aio/converge.yml",
                            "stdout_lines": [
                                "Using CPython 3.10.12 interpreter at: /usr/bin/python3",
                                "Creating virtual environment at: .venv",
                                "   Building atmosphere @ file:///home/zuul/src/github.com/vexxhost/atmosphere",
                                "Downloading pygments (1.2MiB)",
                                "Downloading cryptography (4.2MiB)",
                                "Downloading setuptools (1.1MiB)",
                                "Downloading rjsonnet (1.2MiB)",
                                "Downloading openstacksdk (1.7MiB)",
                                "Downloading netaddr (2.2MiB)",
                                "Downloading ansible-core (2.1MiB)",
                                "Downloading kubernetes (1.9MiB)",
                                "   Building pyperclip==1.9.0",
                                " Downloading rjsonnet",
                                " Downloading netaddr",
                                " Downloading pygments",
                                " Downloading cryptography",
                                " Downloading setuptools",
                                " Downloading kubernetes",
                                " Downloading ansible-core",
                                " Downloading openstacksdk",
                                "      Built pyperclip==1.9.0",
                                "      Built atmosphere @ file:///home/zuul/src/github.com/vexxhost/atmosphere",
                                "Installed 79 packages in 44ms",
                                "WARNING  Molecule scenarios should migrate to 'extensions/molecule'",
                                "INFO     [aio > discovery] scenario test matrix: dependency, cleanup, destroy, syntax, create, prepare, converge, idempotence, side_effect, verify, cleanup, destroy",
                                "INFO     [aio > prerun] Performing prerun with role_name_check=0...",
                                "INFO     [aio > dependency] Executing",
                                "WARNING  [aio > dependency] Missing roles requirements file: requirements.yml",
                                "WARNING  [aio > dependency] Missing collections requirements file: collections.yml",
                                "WARNING  [aio > dependency] Executed: 2 missing (Remove from test_sequence to suppress)",
                                "INFO     [aio > cleanup] Executing",
                                "WARNING  [aio > cleanup] Executed: Missing playbook (Remove from test_sequence to suppress)",
                                "INFO     [aio > destroy] Executing",
                                "WARNING  [aio > destroy] Skipping, '--destroy=never' requested.",
                                "INFO     [aio > destroy] Executed: Successful",
                                "INFO     [aio > syntax] Executing",
                                "",
                                "playbook: /home/zuul/src/github.com/vexxhost/atmosphere/molecule/aio/converge.yml",
                                "INFO     [aio > syntax] Executed: Successful",
                                "INFO     [aio > create] Executing",
                                "WARNING  [aio > create] Executed: Missing playbook (Remove from test_sequence to suppress)",
                                "INFO     [aio > prepare] Executing",
                                "",
                                "PLAY [Prepare] *****************************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:07:25 +0000 (0:00:00.026)       0:00:00.026 ********",
                                "[WARNING]: Platform linux on host instance is using the discovered Python",
                                "interpreter at /usr/bin/python3.10, but future installation of another Python",
                                "interpreter could change the meaning of that path. See",
                                "https://docs.ansible.com/ansible-",
                                "core/2.17/reference_appendices/interpreter_discovery.html for more information.",
                                "ok: [instance]",
                                "",
                                "TASK [Configure short hostname] ************************************************",
                                "Saturday 04 April 2026  10:07:26 +0000 (0:00:01.127)       0:00:01.153 ********",
                                "changed: [instance]",
                                "",
                                "TASK [Ensure hostname inside hosts file] ***************************************",
                                "Saturday 04 April 2026  10:07:27 +0000 (0:00:00.656)       0:00:01.810 ********",
                                "[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created",
                                "with a mode of 0700, this may cause issues when running as another user. To",
                                "avoid this, create the remote_tmp dir with the correct permissions manually",
                                "changed: [instance]",
                                "",
                                "TASK [Install \"dirmngr\" for GPG keyserver operations] **************************",
                                "Saturday 04 April 2026  10:07:27 +0000 (0:00:00.284)       0:00:02.094 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Purge \"snapd\" package] ***************************************************",
                                "Saturday 04 April 2026  10:07:28 +0000 (0:00:01.133)       0:00:03.228 ********",
                                "ok: [instance]",
                                "",
                                "PLAY [Generate workspace for Atmosphere] ***************************************",
                                "",
                                "TASK [Create folders for workspace] ********************************************",
                                "Saturday 04 April 2026  10:07:29 +0000 (0:00:00.673)       0:00:03.901 ********",
                                "changed: [localhost] => (item=group_vars)",
                                "changed: [localhost] => (item=group_vars/all)",
                                "changed: [localhost] => (item=group_vars/controllers)",
                                "changed: [localhost] => (item=group_vars/cephs)",
                                "changed: [localhost] => (item=group_vars/computes)",
                                "changed: [localhost] => (item=host_vars)",
                                "",
                                "PLAY [Generate Ceph control plane configuration for workspace] *****************",
                                "",
                                "TASK [Ensure the Ceph control plane configuration file exists] *****************",
                                "Saturday 04 April 2026  10:07:30 +0000 (0:00:00.974)       0:00:04.875 ********",
                                "changed: [localhost]",
                                "",
                                "TASK [Load the current Ceph control plane configuration into a variable] *******",
                                "Saturday 04 April 2026  10:07:30 +0000 (0:00:00.165)       0:00:05.041 ********",
                                "ok: [localhost]",
                                "",
                                "TASK [Generate Ceph control plane values for missing variables] ****************",
                                "Saturday 04 April 2026  10:07:30 +0000 (0:00:00.029)       0:00:05.070 ********",
                                "ok: [localhost] => (item={'key': 'ceph_fsid', 'value': '9772977c-b199-5d3f-94e7-44f40285818e'})",
                                "ok: [localhost] => (item={'key': 'ceph_mon_public_network', 'value': '10.96.240.0/24'})",
                                "",
                                "TASK [Write new Ceph control plane configuration file to disk] *****************",
                                "Saturday 04 April 2026  10:07:30 +0000 (0:00:00.050)       0:00:05.121 ********",
                                "changed: [localhost]",
                                "",
                                "PLAY [Generate Ceph OSD configuration for workspace] ***************************",
                                "",
                                "TASK [Ensure the Ceph OSDs configuration file exists] **************************",
                                "Saturday 04 April 2026  10:07:31 +0000 (0:00:00.545)       0:00:05.666 ********",
                                "changed: [localhost]",
                                "",
                                "TASK [Load the current Ceph OSDs configuration into a variable] ****************",
                                "Saturday 04 April 2026  10:07:31 +0000 (0:00:00.169)       0:00:05.836 ********",
                                "ok: [localhost]",
                                "",
                                "TASK [Generate Ceph OSDs values for missing variables] *************************",
                                "Saturday 04 April 2026  10:07:31 +0000 (0:00:00.028)       0:00:05.864 ********",
                                "ok: [localhost] => (item={'key': 'ceph_osd_devices', 'value': ['/dev/vdb', '/dev/vdc', '/dev/vdd']})",
                                "",
                                "TASK [Write new Ceph OSDs configuration file to disk] **************************",
                                "Saturday 04 April 2026  10:07:31 +0000 (0:00:00.031)       0:00:05.896 ********",
                                "changed: [localhost]",
                                "",
                                "PLAY [Generate Kubernetes configuration for workspace] *************************",
                                "",
                                "TASK [Ensure the Kubernetes configuration file exists] *************************",
                                "Saturday 04 April 2026  10:07:31 +0000 (0:00:00.350)       0:00:06.246 ********",
                                "changed: [localhost]",
                                "",
                                "TASK [Load the current Kubernetes configuration into a variable] ***************",
                                "Saturday 04 April 2026  10:07:32 +0000 (0:00:00.179)       0:00:06.426 ********",
                                "ok: [localhost]",
                                "",
                                "TASK [Generate Kubernetes values for missing variables] ************************",
                                "Saturday 04 April 2026  10:07:32 +0000 (0:00:00.029)       0:00:06.455 ********",
                                "ok: [localhost] => (item={'key': 'kubernetes_hostname', 'value': '10.96.240.10'})",
                                "ok: [localhost] => (item={'key': 'kubernetes_keepalived_vrid', 'value': 42})",
                                "ok: [localhost] => (item={'key': 'kubernetes_keepalived_vip', 'value': '10.96.240.10'})",
                                "",
                                "TASK [Write new Kubernetes configuration file to disk] *************************",
                                "Saturday 04 April 2026  10:07:32 +0000 (0:00:00.043)       0:00:06.499 ********",
                                "changed: [localhost]",
                                "",
                                "PLAY [Generate Keepalived configuration for workspace] *************************",
                                "",
                                "TASK [Ensure the Keeaplived configuration file exists] *************************",
                                "Saturday 04 April 2026  10:07:32 +0000 (0:00:00.356)       0:00:06.855 ********",
                                "changed: [localhost]",
                                "",
                                "TASK [Load the current Keepalived configuration into a variable] ***************",
                                "Saturday 04 April 2026  10:07:32 +0000 (0:00:00.170)       0:00:07.026 ********",
                                "ok: [localhost]",
                                "",
                                "TASK [Generate Keepalived values for missing variables] ************************",
                                "Saturday 04 April 2026  10:07:32 +0000 (0:00:00.028)       0:00:07.054 ********",
                                "ok: [localhost] => (item={'key': 'keepalived_interface', 'value': 'br-ex'})",
                                "ok: [localhost] => (item={'key': 'keepalived_vip', 'value': '10.96.250.10'})",
                                "",
                                "TASK [Write new Keepalived configuration file to disk] *************************",
                                "Saturday 04 April 2026  10:07:32 +0000 (0:00:00.040)       0:00:07.095 ********",
                                "changed: [localhost]",
                                "",
                                "PLAY [Generate endpoints for workspace] ****************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:07:33 +0000 (0:00:00.358)       0:00:07.454 ********",
                                "ok: [localhost]",
                                "",
                                "TASK [Ensure the endpoints file exists] ****************************************",
                                "Saturday 04 April 2026  10:07:33 +0000 (0:00:00.682)       0:00:08.136 ********",
                                "changed: [localhost]",
                                "",
                                "TASK [Load the current endpoints into a variable] ******************************",
                                "Saturday 04 April 2026  10:07:34 +0000 (0:00:00.170)       0:00:08.306 ********",
                                "ok: [localhost]",
                                "",
                                "TASK [Generate endpoint skeleton for missing variables] ************************",
                                "Saturday 04 April 2026  10:07:34 +0000 (0:00:00.030)       0:00:08.336 ********",
                                "ok: [localhost] => (item=keycloak_host)",
                                "ok: [localhost] => (item=kube_prometheus_stack_grafana_host)",
                                "ok: [localhost] => (item=kube_prometheus_stack_alertmanager_host)",
                                "ok: [localhost] => (item=kube_prometheus_stack_prometheus_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_region_name)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_keystone_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_glance_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_cinder_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_placement_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_barbican_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_neutron_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_nova_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_nova_novnc_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_ironic_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_designate_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_octavia_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_magnum_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_magnum_registry_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_senlin_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_heat_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_heat_cfn_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_horizon_api_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_rgw_host)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_manila_api_host)",
                                "",
                                "TASK [Write new endpoints file to disk] ****************************************",
                                "Saturday 04 April 2026  10:07:34 +0000 (0:00:00.817)       0:00:09.153 ********",
                                "changed: [localhost]",
                                "",
                                "TASK [Ensure the endpoints file exists] ****************************************",
                                "Saturday 04 April 2026  10:07:35 +0000 (0:00:00.343)       0:00:09.497 ********",
                                "changed: [localhost]",
                                "",
                                "PLAY [Generate Neutron configuration for workspace] ****************************",
                                "",
                                "TASK [Ensure the Neutron configuration file exists] ****************************",
                                "Saturday 04 April 2026  10:07:35 +0000 (0:00:00.180)       0:00:09.677 ********",
                                "changed: [localhost]",
                                "",
                                "TASK [Load the current Neutron configuration into a variable] ******************",
                                "Saturday 04 April 2026  10:07:35 +0000 (0:00:00.173)       0:00:09.850 ********",
                                "ok: [localhost]",
                                "",
                                "TASK [Generate Neutron values for missing variables] ***************************",
                                "Saturday 04 April 2026  10:07:35 +0000 (0:00:00.033)       0:00:09.884 ********",
                                "ok: [localhost] => (item={'key': 'neutron_networks', 'value': [{'name': 'public', 'external': True, 'shared': True, 'mtu_size': 1500, 'port_security_enabled': True, 'provider_network_type': 'flat', 'provider_physical_network': 'external', 'subnets': [{'name': 'public-subnet', 'cidr': '10.96.250.0/24', 'gateway_ip': '10.96.250.10', 'allocation_pool_start': '10.96.250.200', 'allocation_pool_end': '10.96.250.220', 'enable_dhcp': True}]}]})",
                                "",
                                "TASK [Write new Neutron configuration file to disk] ****************************",
                                "Saturday 04 April 2026  10:07:35 +0000 (0:00:00.042)       0:00:09.926 ********",
                                "changed: [localhost]",
                                "",
                                "PLAY [Generate Nova configuration for workspace] *******************************",
                                "",
                                "TASK [Ensure the Nova configuration file exists] *******************************",
                                "Saturday 04 April 2026  10:07:36 +0000 (0:00:00.349)       0:00:10.276 ********",
                                "changed: [localhost]",
                                "",
                                "TASK [Load the current Nova configuration into a variable] *********************",
                                "Saturday 04 April 2026  10:07:36 +0000 (0:00:00.171)       0:00:10.447 ********",
                                "ok: [localhost]",
                                "",
                                "TASK [Generate Nova values for missing variables] ******************************",
                                "Saturday 04 April 2026  10:07:36 +0000 (0:00:00.031)       0:00:10.478 ********",
                                "ok: [localhost] => (item={'key': 'nova_flavors', 'value': [{'name': 'm1.tiny', 'ram': 512, 'disk': 1, 'vcpus': 1}, {'name': 'm1.small', 'ram': 2048, 'disk': 20, 'vcpus': 1}, {'name': 'm1.medium', 'ram': 4096, 'disk': 40, 'vcpus': 2}, {'name': 'm1.large', 'ram': 8192, 'disk': 80, 'vcpus': 4}, {'name': 'm1.xlarge', 'ram': 16384, 'disk': 160, 'vcpus': 8}]})",
                                "",
                                "TASK [Write new Nova configuration file to disk] *******************************",
                                "Saturday 04 April 2026  10:07:36 +0000 (0:00:00.046)       0:00:10.525 ********",
                                "changed: [localhost]",
                                "",
                                "PLAY [Generate secrets for workspace] ******************************************",
                                "",
                                "TASK [Ensure the secrets file exists] ******************************************",
                                "Saturday 04 April 2026  10:07:36 +0000 (0:00:00.348)       0:00:10.873 ********",
                                "changed: [localhost]",
                                "",
                                "TASK [Load the current secrets into a variable] ********************************",
                                "Saturday 04 April 2026  10:07:36 +0000 (0:00:00.166)       0:00:11.040 ********",
                                "ok: [localhost]",
                                "",
                                "TASK [Generate secrets for missing variables] **********************************",
                                "Saturday 04 April 2026  10:07:36 +0000 (0:00:00.035)       0:00:11.076 ********",
                                "ok: [localhost] => (item=heat_auth_encryption_key)",
                                "ok: [localhost] => (item=keepalived_password)",
                                "ok: [localhost] => (item=keycloak_admin_password)",
                                "ok: [localhost] => (item=keycloak_database_password)",
                                "ok: [localhost] => (item=keystone_keycloak_client_secret)",
                                "ok: [localhost] => (item=keystone_oidc_crypto_passphrase)",
                                "ok: [localhost] => (item=kube_prometheus_stack_grafana_admin_password)",
                                "ok: [localhost] => (item=octavia_heartbeat_key)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_rabbitmq_admin_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_memcached_secret_key)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_keystone_admin_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_keystone_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_keystone_rabbitmq_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_glance_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_glance_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_glance_rabbitmq_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_cinder_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_cinder_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_cinder_rabbitmq_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_placement_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_placement_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_barbican_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_barbican_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_neutron_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_neutron_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_neutron_rabbitmq_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_neutron_metadata_secret)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_nova_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_nova_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_nova_rabbitmq_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_ironic_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_ironic_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_ironic_rabbitmq_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_designate_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_designate_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_designate_rabbitmq_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_octavia_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_octavia_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_octavia_rabbitmq_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_magnum_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_magnum_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_magnum_rabbitmq_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_senlin_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_senlin_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_senlin_rabbitmq_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_heat_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_heat_trustee_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_heat_stack_user_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_heat_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_heat_rabbitmq_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_horizon_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_tempest_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_openstack_exporter_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_rgw_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_manila_keystone_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_manila_mariadb_password)",
                                "ok: [localhost] => (item=openstack_helm_endpoints_staffeln_mariadb_password)",
                                "",
                                "TASK [Generate base64 encoded secrets] *****************************************",
                                "Saturday 04 April 2026  10:07:37 +0000 (0:00:00.417)       0:00:11.493 ********",
                                "ok: [localhost] => (item=barbican_kek)",
                                "",
                                "TASK [Generate temporary files for generating keys for missing variables] ******",
                                "Saturday 04 April 2026  10:07:37 +0000 (0:00:00.053)       0:00:11.546 ********",
                                "changed: [localhost] => (item=manila_ssh_key)",
                                "changed: [localhost] => (item=nova_ssh_key)",
                                "",
                                "TASK [Generate SSH keys for missing variables] *********************************",
                                "Saturday 04 April 2026  10:07:37 +0000 (0:00:00.410)       0:00:11.957 ********",
                                "changed: [localhost] => (item=manila_ssh_key)",
                                "changed: [localhost] => (item=nova_ssh_key)",
                                "",
                                "TASK [Set values for SSH keys] *************************************************",
                                "Saturday 04 April 2026  10:07:40 +0000 (0:00:02.988)       0:00:14.945 ********",
                                "ok: [localhost] => (item=manila_ssh_key)",
                                "ok: [localhost] => (item=nova_ssh_key)",
                                "",
                                "TASK [Delete the temporary files generated for SSH keys] ***********************",
                                "Saturday 04 April 2026  10:07:40 +0000 (0:00:00.053)       0:00:14.998 ********",
                                "changed: [localhost] => (item=manila_ssh_key)",
                                "changed: [localhost] => (item=nova_ssh_key)",
                                "",
                                "TASK [Write new secrets file to disk] ******************************************",
                                "Saturday 04 April 2026  10:07:41 +0000 (0:00:00.318)       0:00:15.317 ********",
                                "changed: [localhost]",
                                "",
                                "TASK [Encrypt secrets file with Vault password] ********************************",
                                "Saturday 04 April 2026  10:07:41 +0000 (0:00:00.348)       0:00:15.665 ********",
                                "skipping: [localhost]",
                                "",
                                "PLAY [Setup networking] ********************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:07:41 +0000 (0:00:00.034)       0:00:15.699 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Create bridge for management network] ************************************",
                                "Saturday 04 April 2026  10:07:42 +0000 (0:00:00.681)       0:00:16.381 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Create fake interface for management bridge] *****************************",
                                "Saturday 04 April 2026  10:07:42 +0000 (0:00:00.362)       0:00:16.743 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Assign dummy interface to management bridge] *****************************",
                                "Saturday 04 April 2026  10:07:42 +0000 (0:00:00.206)       0:00:16.950 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Assign IP address for management bridge] *********************************",
                                "Saturday 04 April 2026  10:07:42 +0000 (0:00:00.186)       0:00:17.136 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Bring up interfaces] *****************************************************",
                                "Saturday 04 April 2026  10:07:43 +0000 (0:00:00.208)       0:00:17.345 ********",
                                "ok: [instance] => (item=br-mgmt)",
                                "ok: [instance] => (item=dummy0)",
                                "",
                                "PLAY [Create devices for Ceph] *************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:07:43 +0000 (0:00:00.386)       0:00:17.731 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Install depedencies] *****************************************************",
                                "Saturday 04 April 2026  10:07:44 +0000 (0:00:00.704)       0:00:18.436 ********",
                                "changed: [instance]",
                                "",
                                "TASK [Start up service] ********************************************************",
                                "Saturday 04 April 2026  10:08:03 +0000 (0:00:19.794)       0:00:38.231 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Generate lvm.conf] *******************************************************",
                                "Saturday 04 April 2026  10:08:04 +0000 (0:00:00.544)       0:00:38.776 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Write /etc/lvm/lvm.conf] *************************************************",
                                "Saturday 04 April 2026  10:08:04 +0000 (0:00:00.203)       0:00:38.979 ********",
                                "changed: [instance]",
                                "",
                                "TASK [Get list of all loopback devices] ****************************************",
                                "Saturday 04 April 2026  10:08:05 +0000 (0:00:00.441)       0:00:39.421 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Fail if there is any existing loopback devices] **************************",
                                "Saturday 04 April 2026  10:08:05 +0000 (0:00:00.198)       0:00:39.619 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [Create devices for Ceph] *************************************************",
                                "Saturday 04 April 2026  10:08:05 +0000 (0:00:00.026)       0:00:39.646 ********",
                                "changed: [instance] => (item=osd0)",
                                "changed: [instance] => (item=osd1)",
                                "changed: [instance] => (item=osd2)",
                                "",
                                "TASK [Set permissions on loopback devices] *************************************",
                                "Saturday 04 April 2026  10:08:05 +0000 (0:00:00.530)       0:00:40.176 ********",
                                "changed: [instance] => (item=osd0)",
                                "changed: [instance] => (item=osd1)",
                                "changed: [instance] => (item=osd2)",
                                "",
                                "TASK [Start loop devices] ******************************************************",
                                "Saturday 04 April 2026  10:08:06 +0000 (0:00:00.525)       0:00:40.702 ********",
                                "changed: [instance] => (item=osd0)",
                                "changed: [instance] => (item=osd1)",
                                "changed: [instance] => (item=osd2)",
                                "",
                                "TASK [Create a volume group for each loop device] ******************************",
                                "Saturday 04 April 2026  10:08:07 +0000 (0:00:00.729)       0:00:41.431 ********",
                                "changed: [instance] => (item=osd0)",
                                "changed: [instance] => (item=osd1)",
                                "changed: [instance] => (item=osd2)",
                                "",
                                "TASK [Create a logical volume for each loop device] ****************************",
                                "Saturday 04 April 2026  10:08:10 +0000 (0:00:03.109)       0:00:44.540 ********",
                                "changed: [instance] => (item=ceph-instance-osd0)",
                                "changed: [instance] => (item=ceph-instance-osd1)",
                                "changed: [instance] => (item=ceph-instance-osd2)",
                                "",
                                "PLAY [controllers] *************************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:08:12 +0000 (0:00:01.851)       0:00:46.392 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Set masquerade rule] *****************************************************",
                                "Saturday 04 April 2026  10:08:13 +0000 (0:00:00.913)       0:00:47.305 ********",
                                "changed: [instance]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "instance                   : ok=24   changed=10   unreachable=0    failed=0    skipped=1    rescued=0    ignored=0",
                                "localhost                  : ok=40   changed=21   unreachable=0    failed=0    skipped=1    rescued=0    ignored=0",
                                "",
                                "Saturday 04 April 2026  10:08:13 +0000 (0:00:00.443)       0:00:47.748 ********",
                                "===============================================================================",
                                "Install depedencies ---------------------------------------------------- 19.79s",
                                "Create a volume group for each loop device ------------------------------ 3.11s",
                                "Generate SSH keys for missing variables --------------------------------- 2.99s",
                                "Create a logical volume for each loop device ---------------------------- 1.85s",
                                "Install \"dirmngr\" for GPG keyserver operations -------------------------- 1.13s",
                                "Gathering Facts --------------------------------------------------------- 1.13s",
                                "Create folders for workspace -------------------------------------------- 0.97s",
                                "Gathering Facts --------------------------------------------------------- 0.91s",
                                "Generate endpoint skeleton for missing variables ------------------------ 0.82s",
                                "Start loop devices ------------------------------------------------------ 0.73s",
                                "Gathering Facts --------------------------------------------------------- 0.70s",
                                "Gathering Facts --------------------------------------------------------- 0.68s",
                                "Gathering Facts --------------------------------------------------------- 0.68s",
                                "Purge \"snapd\" package --------------------------------------------------- 0.67s",
                                "Configure short hostname ------------------------------------------------ 0.66s",
                                "Write new Ceph control plane configuration file to disk ----------------- 0.55s",
                                "Start up service -------------------------------------------------------- 0.54s",
                                "Create devices for Ceph ------------------------------------------------- 0.53s",
                                "Set permissions on loopback devices ------------------------------------- 0.53s",
                                "Set masquerade rule ----------------------------------------------------- 0.44s",
                                "INFO     [aio > prepare] Executed: Successful",
                                "INFO     [aio > converge] Executing",
                                "",
                                "PLAY [all] *********************************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:08:16 +0000 (0:00:00.065)       0:00:00.065 ********",
                                "[WARNING]: Platform linux on host instance is using the discovered Python",
                                "interpreter at /usr/bin/python3.10, but future installation of another Python",
                                "interpreter could change the meaning of that path. See",
                                "https://docs.ansible.com/ansible-",
                                "core/2.17/reference_appendices/interpreter_discovery.html for more information.",
                                "ok: [instance]",
                                "",
                                "TASK [Set a fact with the \"atmosphere_images\" for other plays] *****************",
                                "Saturday 04 April 2026  10:08:17 +0000 (0:00:01.170)       0:00:01.236 ********",
                                "ok: [instance]",
                                "",
                                "PLAY [Deploy Ceph monitors & managers] *****************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:08:17 +0000 (0:00:00.173)       0:00:01.409 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:08:18 +0000 (0:00:00.904)       0:00:02.313 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.package : Update state for tar] **********************",
                                "Saturday 04 April 2026  10:08:18 +0000 (0:00:00.287)       0:00:02.601 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.directory : Create directory (/var/lib/downloads)] ***",
                                "Saturday 04 April 2026  10:08:18 +0000 (0:00:00.044)       0:00:02.645 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:08:19 +0000 (0:00:00.289)       0:00:02.935 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://github.com/opencontainers/runc/releases/download/v1.4.0/runc.amd64\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:08:19 +0000 (0:00:00.074)       0:00:03.010 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:08:19 +0000 (0:00:00.628)       0:00:03.639 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.package : Update state for tar] **********************",
                                "Saturday 04 April 2026  10:08:19 +0000 (0:00:00.048)       0:00:03.687 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:08:19 +0000 (0:00:00.047)       0:00:03.735 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.package : Update state for tar] **********************",
                                "Saturday 04 April 2026  10:08:20 +0000 (0:00:00.199)       0:00:03.935 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:08:21 +0000 (0:00:01.197)       0:00:05.132 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://github.com/containerd/containerd/releases/download/v2.2.0/containerd-2.2.0-linux-amd64.tar.gz\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:08:21 +0000 (0:00:00.068)       0:00:05.200 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:08:22 +0000 (0:00:00.677)       0:00:05.878 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Install SELinux packages] ***************",
                                "Saturday 04 April 2026  10:08:25 +0000 (0:00:03.845)       0:00:09.724 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Set SELinux to permissive at runtime] ***",
                                "Saturday 04 April 2026  10:08:25 +0000 (0:00:00.030)       0:00:09.755 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Persist SELinux permissive mode] ********",
                                "Saturday 04 April 2026  10:08:25 +0000 (0:00:00.036)       0:00:09.791 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Install AppArmor packages] **************",
                                "Saturday 04 April 2026  10:08:25 +0000 (0:00:00.037)       0:00:09.829 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Create systemd service file for containerd] ***",
                                "Saturday 04 April 2026  10:08:31 +0000 (0:00:05.469)       0:00:15.298 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Create folders for configuration] *******",
                                "Saturday 04 April 2026  10:08:31 +0000 (0:00:00.524)       0:00:15.822 ********",
                                "changed: [instance] => (item={'path': '/etc/containerd'})",
                                "changed: [instance] => (item={'path': '/var/lib/containerd', 'mode': '0o700'})",
                                "changed: [instance] => (item={'path': '/run/containerd', 'mode': '0o711'})",
                                "changed: [instance] => (item={'path': '/run/containerd/io.containerd.grpc.v1.cri', 'mode': '0o700'})",
                                "changed: [instance] => (item={'path': '/run/containerd/io.containerd.sandbox.controller.v1.shim', 'mode': '0o700'})",
                                "",
                                "TASK [vexxhost.containers.containerd : Create containerd config file] **********",
                                "Saturday 04 April 2026  10:08:32 +0000 (0:00:00.934)       0:00:16.757 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Force any restarts if necessary] ********",
                                "Saturday 04 April 2026  10:08:33 +0000 (0:00:00.561)       0:00:17.318 ********",
                                "",
                                "RUNNING HANDLER [vexxhost.containers.containerd : Reload systemd] **************",
                                "Saturday 04 April 2026  10:08:33 +0000 (0:00:00.009)       0:00:17.328 ********",
                                "ok: [instance]",
                                "",
                                "RUNNING HANDLER [vexxhost.containers.containerd : Restart containerd] **********",
                                "Saturday 04 April 2026  10:08:34 +0000 (0:00:00.904)       0:00:18.233 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Enable and start service] ***************",
                                "Saturday 04 April 2026  10:08:34 +0000 (0:00:00.445)       0:00:18.679 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:08:35 +0000 (0:00:00.584)       0:00:19.264 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:08:35 +0000 (0:00:00.204)       0:00:19.469 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://download.docker.com/linux/static/stable/x86_64/docker-24.0.9.tgz\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:08:35 +0000 (0:00:00.057)       0:00:19.526 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:08:36 +0000 (0:00:00.917)       0:00:20.444 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Install AppArmor packages] ******************",
                                "Saturday 04 April 2026  10:08:40 +0000 (0:00:04.359)       0:00:24.804 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Ensure group \"docker\" exists] ***************",
                                "Saturday 04 April 2026  10:08:41 +0000 (0:00:00.967)       0:00:25.771 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Create systemd service file for docker] *****",
                                "Saturday 04 April 2026  10:08:42 +0000 (0:00:00.368)       0:00:26.139 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Create folders for configuration] ***********",
                                "Saturday 04 April 2026  10:08:42 +0000 (0:00:00.449)       0:00:26.589 ********",
                                "changed: [instance] => (item={'path': '/etc/docker'})",
                                "changed: [instance] => (item={'path': '/var/lib/docker', 'mode': '0o710'})",
                                "changed: [instance] => (item={'path': '/run/docker', 'mode': '0o711'})",
                                "",
                                "TASK [vexxhost.containers.docker : Create systemd socket file for docker] ******",
                                "Saturday 04 April 2026  10:08:43 +0000 (0:00:00.541)       0:00:27.131 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Create docker daemon config file] ***********",
                                "Saturday 04 April 2026  10:08:43 +0000 (0:00:00.463)       0:00:27.594 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Force any restarts if necessary] ************",
                                "Saturday 04 April 2026  10:08:44 +0000 (0:00:00.487)       0:00:28.082 ********",
                                "",
                                "RUNNING HANDLER [vexxhost.containers.containerd : Reload systemd] **************",
                                "Saturday 04 April 2026  10:08:44 +0000 (0:00:00.008)       0:00:28.090 ********",
                                "ok: [instance]",
                                "",
                                "RUNNING HANDLER [vexxhost.containers.docker : Restart docker] ******************",
                                "Saturday 04 April 2026  10:08:44 +0000 (0:00:00.715)       0:00:28.806 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Enable and start service] *******************",
                                "Saturday 04 April 2026  10:08:46 +0000 (0:00:01.363)       0:00:30.170 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Gather variables for each operating system] ******",
                                "Saturday 04 April 2026  10:08:46 +0000 (0:00:00.539)       0:00:30.709 ********",
                                "ok: [instance] => (item=/home/zuul/.ansible/collections/ansible_collections/vexxhost/ceph/roles/cephadm/vars/ubuntu-22.04.yml)",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Install packages] ********************************",
                                "Saturday 04 April 2026  10:08:46 +0000 (0:00:00.062)       0:00:30.771 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Ensure services are started] *********************",
                                "Saturday 04 April 2026  10:08:52 +0000 (0:00:05.186)       0:00:35.958 ********",
                                "ok: [instance] => (item=chronyd)",
                                "ok: [instance] => (item=sshd)",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Download \"cephadm\"] ******************************",
                                "Saturday 04 April 2026  10:08:52 +0000 (0:00:00.635)       0:00:36.594 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Remove cephadm from old path] ********************",
                                "Saturday 04 April 2026  10:08:53 +0000 (0:00:00.794)       0:00:37.388 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Ensure \"cephadm\" user is present] ****************",
                                "Saturday 04 April 2026  10:08:53 +0000 (0:00:00.197)       0:00:37.586 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Allow \"cephadm\" user to have passwordless sudo] ***",
                                "Saturday 04 April 2026  10:08:54 +0000 (0:00:00.539)       0:00:38.125 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Get `cephadm ls` status] *****************************",
                                "Saturday 04 April 2026  10:08:54 +0000 (0:00:00.441)       0:00:38.567 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Parse the `cephadm ls` output] ***********************",
                                "Saturday 04 April 2026  10:08:56 +0000 (0:00:01.679)       0:00:40.247 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Assimilate existing configs in `ceph.conf`] **********",
                                "Saturday 04 April 2026  10:08:56 +0000 (0:00:00.054)       0:00:40.301 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Adopt monitor to cluster] ****************************",
                                "Saturday 04 April 2026  10:08:56 +0000 (0:00:00.042)       0:00:40.344 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Adopt manager to cluster] ****************************",
                                "Saturday 04 April 2026  10:08:56 +0000 (0:00:00.041)       0:00:40.386 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Enable \"cephadm\" mgr module] *************************",
                                "Saturday 04 April 2026  10:08:56 +0000 (0:00:00.041)       0:00:40.428 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Set orchestrator backend to \"cephadm\"] ***************",
                                "Saturday 04 April 2026  10:08:56 +0000 (0:00:00.038)       0:00:40.466 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Use `cephadm` user for cephadm] **********************",
                                "Saturday 04 April 2026  10:08:56 +0000 (0:00:00.039)       0:00:40.506 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Generate \"cephadm\" key] ******************************",
                                "Saturday 04 April 2026  10:08:56 +0000 (0:00:00.042)       0:00:40.548 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Set Ceph Monitor IP address] *************************",
                                "Saturday 04 April 2026  10:08:56 +0000 (0:00:00.040)       0:00:40.589 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Check if any node is bootstrapped] *******************",
                                "Saturday 04 April 2026  10:08:56 +0000 (0:00:00.115)       0:00:40.705 ********",
                                "ok: [instance] => (item=instance)",
                                "",
                                "TASK [vexxhost.ceph.mon : Select pre-existing bootstrap node if exists] ********",
                                "Saturday 04 April 2026  10:08:57 +0000 (0:00:00.229)       0:00:40.934 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Bootstrap cluster] ***********************************",
                                "Saturday 04 April 2026  10:08:57 +0000 (0:00:00.057)       0:00:40.991 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/ceph/roles/mon/tasks/bootstrap-ceph.yml for instance",
                                "",
                                "TASK [vexxhost.ceph.mon : Generate temporary file for \"ceph.conf\"] *************",
                                "Saturday 04 April 2026  10:08:57 +0000 (0:00:00.081)       0:00:41.072 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Include extra configuration values] ******************",
                                "Saturday 04 April 2026  10:08:57 +0000 (0:00:00.343)       0:00:41.416 ********",
                                "changed: [instance] => (item={'option': 'mon allow pool size one', 'section': 'global', 'value': True})",
                                "changed: [instance] => (item={'option': 'osd crush chooseleaf type', 'section': 'global', 'value': 0})",
                                "changed: [instance] => (item={'option': 'auth allow insecure global id reclaim', 'section': 'mon', 'value': False})",
                                "",
                                "TASK [vexxhost.ceph.mon : Run Bootstrap coomand] *******************************",
                                "Saturday 04 April 2026  10:08:58 +0000 (0:00:00.682)       0:00:42.098 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Remove temporary file for \"ceph.conf\"] ***************",
                                "Saturday 04 April 2026  10:11:16 +0000 (0:02:18.663)       0:03:00.762 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Set bootstrap node] **********************************",
                                "Saturday 04 April 2026  10:11:17 +0000 (0:00:00.233)       0:03:00.996 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Install Ceph host] *******************************************************",
                                "Saturday 04 April 2026  10:11:17 +0000 (0:00:00.052)       0:03:01.049 ********",
                                "included: vexxhost.ceph.cephadm_host for instance",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Get public SSH key for \"cephadm\" user] ******",
                                "Saturday 04 April 2026  10:11:17 +0000 (0:00:00.087)       0:03:01.136 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Set fact with public SSH key for \"cephadm\" user] ***",
                                "Saturday 04 April 2026  10:11:19 +0000 (0:00:01.717)       0:03:02.854 ********",
                                "ok: [instance] => (item=instance)",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Set authorized key for \"cephadm\"] ***********",
                                "Saturday 04 April 2026  10:11:19 +0000 (0:00:00.066)       0:03:02.920 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Add new host to Ceph] ***********************",
                                "Saturday 04 April 2026  10:11:19 +0000 (0:00:00.461)       0:03:03.381 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Configure \"mon\" label for monitors] ******************",
                                "Saturday 04 April 2026  10:11:21 +0000 (0:00:02.391)       0:03:05.773 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mon : Validate monitor exist] ******************************",
                                "Saturday 04 April 2026  10:11:23 +0000 (0:00:01.955)       0:03:07.729 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Install Ceph host] *******************************************************",
                                "Saturday 04 April 2026  10:11:35 +0000 (0:00:11.173)       0:03:18.903 ********",
                                "included: vexxhost.ceph.cephadm_host for instance",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Get public SSH key for \"cephadm\" user] ******",
                                "Saturday 04 April 2026  10:11:35 +0000 (0:00:00.109)       0:03:19.012 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Set fact with public SSH key for \"cephadm\" user] ***",
                                "Saturday 04 April 2026  10:11:35 +0000 (0:00:00.057)       0:03:19.070 ********",
                                "skipping: [instance] => (item=instance)",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Set authorized key for \"cephadm\"] ***********",
                                "Saturday 04 April 2026  10:11:35 +0000 (0:00:00.066)       0:03:19.137 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Add new host to Ceph] ***********************",
                                "Saturday 04 April 2026  10:11:35 +0000 (0:00:00.262)       0:03:19.400 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mgr : Configure \"mgr\" label for managers] ******************",
                                "Saturday 04 April 2026  10:11:38 +0000 (0:00:02.920)       0:03:22.320 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mgr : Validate manager exist] ******************************",
                                "Saturday 04 April 2026  10:11:41 +0000 (0:00:02.780)       0:03:25.101 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.mgr : Enable the Ceph Manager prometheus module] ***********",
                                "Saturday 04 April 2026  10:11:45 +0000 (0:00:04.581)       0:03:29.683 ********",
                                "ok: [instance]",
                                "",
                                "PLAY [Deploy Ceph OSDs] ********************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:11:48 +0000 (0:00:02.653)       0:03:32.336 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:11:49 +0000 (0:00:01.023)       0:03:33.359 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.package : Update state for tar] **********************",
                                "Saturday 04 April 2026  10:11:49 +0000 (0:00:00.219)       0:03:33.578 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.directory : Create directory (/var/lib/downloads)] ***",
                                "Saturday 04 April 2026  10:11:49 +0000 (0:00:00.048)       0:03:33.627 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:11:50 +0000 (0:00:00.218)       0:03:33.845 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://github.com/opencontainers/runc/releases/download/v1.4.0/runc.amd64\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:11:50 +0000 (0:00:00.060)       0:03:33.906 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:11:50 +0000 (0:00:00.304)       0:03:34.210 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.package : Update state for tar] **********************",
                                "Saturday 04 April 2026  10:11:50 +0000 (0:00:00.042)       0:03:34.253 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:11:50 +0000 (0:00:00.043)       0:03:34.296 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.package : Update state for tar] **********************",
                                "Saturday 04 April 2026  10:11:50 +0000 (0:00:00.206)       0:03:34.503 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:11:51 +0000 (0:00:01.120)       0:03:35.623 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://github.com/containerd/containerd/releases/download/v2.2.0/containerd-2.2.0-linux-amd64.tar.gz\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:11:51 +0000 (0:00:00.077)       0:03:35.700 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:11:52 +0000 (0:00:00.377)       0:03:36.078 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Install SELinux packages] ***************",
                                "Saturday 04 April 2026  10:11:54 +0000 (0:00:01.975)       0:03:38.053 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Set SELinux to permissive at runtime] ***",
                                "Saturday 04 April 2026  10:11:54 +0000 (0:00:00.027)       0:03:38.080 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Persist SELinux permissive mode] ********",
                                "Saturday 04 April 2026  10:11:54 +0000 (0:00:00.031)       0:03:38.112 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Install AppArmor packages] **************",
                                "Saturday 04 April 2026  10:11:54 +0000 (0:00:00.028)       0:03:38.140 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Create systemd service file for containerd] ***",
                                "Saturday 04 April 2026  10:11:55 +0000 (0:00:01.167)       0:03:39.307 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Create folders for configuration] *******",
                                "Saturday 04 April 2026  10:11:55 +0000 (0:00:00.470)       0:03:39.778 ********",
                                "ok: [instance] => (item={'path': '/etc/containerd'})",
                                "ok: [instance] => (item={'path': '/var/lib/containerd', 'mode': '0o700'})",
                                "ok: [instance] => (item={'path': '/run/containerd', 'mode': '0o711'})",
                                "ok: [instance] => (item={'path': '/run/containerd/io.containerd.grpc.v1.cri', 'mode': '0o700'})",
                                "ok: [instance] => (item={'path': '/run/containerd/io.containerd.sandbox.controller.v1.shim', 'mode': '0o700'})",
                                "",
                                "TASK [vexxhost.containers.containerd : Create containerd config file] **********",
                                "Saturday 04 April 2026  10:11:56 +0000 (0:00:00.945)       0:03:40.724 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Force any restarts if necessary] ********",
                                "Saturday 04 April 2026  10:11:57 +0000 (0:00:00.540)       0:03:41.265 ********",
                                "",
                                "TASK [vexxhost.containers.containerd : Enable and start service] ***************",
                                "Saturday 04 April 2026  10:11:57 +0000 (0:00:00.007)       0:03:41.272 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:11:57 +0000 (0:00:00.372)       0:03:41.644 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:11:58 +0000 (0:00:00.220)       0:03:41.864 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://download.docker.com/linux/static/stable/x86_64/docker-24.0.9.tgz\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:11:58 +0000 (0:00:00.062)       0:03:41.927 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:11:58 +0000 (0:00:00.374)       0:03:42.301 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Install AppArmor packages] ******************",
                                "Saturday 04 April 2026  10:12:01 +0000 (0:00:03.073)       0:03:45.375 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Ensure group \"docker\" exists] ***************",
                                "Saturday 04 April 2026  10:12:02 +0000 (0:00:01.174)       0:03:46.549 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Create systemd service file for docker] *****",
                                "Saturday 04 April 2026  10:12:02 +0000 (0:00:00.225)       0:03:46.774 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Create folders for configuration] ***********",
                                "Saturday 04 April 2026  10:12:03 +0000 (0:00:00.470)       0:03:47.244 ********",
                                "ok: [instance] => (item={'path': '/etc/docker'})",
                                "ok: [instance] => (item={'path': '/var/lib/docker', 'mode': '0o710'})",
                                "ok: [instance] => (item={'path': '/run/docker', 'mode': '0o711'})",
                                "",
                                "TASK [vexxhost.containers.docker : Create systemd socket file for docker] ******",
                                "Saturday 04 April 2026  10:12:04 +0000 (0:00:00.593)       0:03:47.838 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Create docker daemon config file] ***********",
                                "Saturday 04 April 2026  10:12:04 +0000 (0:00:00.433)       0:03:48.272 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.docker : Force any restarts if necessary] ************",
                                "Saturday 04 April 2026  10:12:04 +0000 (0:00:00.428)       0:03:48.700 ********",
                                "",
                                "TASK [vexxhost.containers.docker : Enable and start service] *******************",
                                "Saturday 04 April 2026  10:12:04 +0000 (0:00:00.007)       0:03:48.707 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Gather variables for each operating system] ******",
                                "Saturday 04 April 2026  10:12:05 +0000 (0:00:00.362)       0:03:49.070 ********",
                                "ok: [instance] => (item=/home/zuul/.ansible/collections/ansible_collections/vexxhost/ceph/roles/cephadm/vars/ubuntu-22.04.yml)",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Install packages] ********************************",
                                "Saturday 04 April 2026  10:12:05 +0000 (0:00:00.064)       0:03:49.135 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Ensure services are started] *********************",
                                "Saturday 04 April 2026  10:12:06 +0000 (0:00:01.120)       0:03:50.255 ********",
                                "ok: [instance] => (item=chronyd)",
                                "ok: [instance] => (item=sshd)",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Download \"cephadm\"] ******************************",
                                "Saturday 04 April 2026  10:12:07 +0000 (0:00:00.704)       0:03:50.960 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Remove cephadm from old path] ********************",
                                "Saturday 04 April 2026  10:12:07 +0000 (0:00:00.800)       0:03:51.761 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Ensure \"cephadm\" user is present] ****************",
                                "Saturday 04 April 2026  10:12:08 +0000 (0:00:00.215)       0:03:51.976 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm : Allow \"cephadm\" user to have passwordless sudo] ***",
                                "Saturday 04 April 2026  10:12:08 +0000 (0:00:00.284)       0:03:52.260 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Get monitor status] **********************************",
                                "Saturday 04 April 2026  10:12:08 +0000 (0:00:00.233)       0:03:52.493 ********",
                                "ok: [instance] => (item=instance)",
                                "",
                                "TASK [vexxhost.ceph.osd : Select admin host] ***********************************",
                                "Saturday 04 April 2026  10:12:08 +0000 (0:00:00.242)       0:03:52.736 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Get `cephadm ls` status] *****************************",
                                "Saturday 04 April 2026  10:12:08 +0000 (0:00:00.057)       0:03:52.794 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Parse the `cephadm ls` output] ***********************",
                                "Saturday 04 April 2026  10:12:14 +0000 (0:00:05.365)       0:03:58.159 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Install Ceph host] *******************************************************",
                                "Saturday 04 April 2026  10:12:14 +0000 (0:00:00.058)       0:03:58.218 ********",
                                "included: vexxhost.ceph.cephadm_host for instance",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Get public SSH key for \"cephadm\" user] ******",
                                "Saturday 04 April 2026  10:12:14 +0000 (0:00:00.067)       0:03:58.285 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Set fact with public SSH key for \"cephadm\" user] ***",
                                "Saturday 04 April 2026  10:12:14 +0000 (0:00:00.050)       0:03:58.335 ********",
                                "skipping: [instance] => (item=instance)",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Set authorized key for \"cephadm\"] ***********",
                                "Saturday 04 April 2026  10:12:14 +0000 (0:00:00.067)       0:03:58.403 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.cephadm_host : Add new host to Ceph] ***********************",
                                "Saturday 04 April 2026  10:12:14 +0000 (0:00:00.264)       0:03:58.668 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Adopt OSDs to cluster] *******************************",
                                "Saturday 04 April 2026  10:12:17 +0000 (0:00:02.205)       0:04:00.874 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Wait until OSD added to cephadm] *********************",
                                "Saturday 04 April 2026  10:12:17 +0000 (0:00:00.036)       0:04:00.911 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Ensure all OSDs are non-legacy] **********************",
                                "Saturday 04 April 2026  10:12:17 +0000 (0:00:00.041)       0:04:00.953 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Get `ceph-volume lvm list` status] *******************",
                                "Saturday 04 April 2026  10:12:22 +0000 (0:00:05.367)       0:04:06.320 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Install OSDs] ****************************************",
                                "Saturday 04 April 2026  10:12:35 +0000 (0:00:13.156)       0:04:19.477 ********",
                                "ok: [instance] => (item=/dev/ceph-instance-osd0/data)",
                                "ok: [instance] => (item=/dev/ceph-instance-osd1/data)",
                                "ok: [instance] => (item=/dev/ceph-instance-osd2/data)",
                                "",
                                "TASK [vexxhost.ceph.osd : Get mon dump] ****************************************",
                                "Saturday 04 April 2026  10:14:38 +0000 (0:02:03.319)       0:06:22.797 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Mark require osd release] ****************************",
                                "Saturday 04 April 2026  10:14:41 +0000 (0:00:02.775)       0:06:25.572 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Wait for all OSD to be running] **********************",
                                "Saturday 04 April 2026  10:14:43 +0000 (0:00:02.088)       0:06:27.661 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/ceph/roles/osd/tasks/check-osds.yml for instance",
                                "",
                                "TASK [vexxhost.ceph.osd : Set the retry count] *********************************",
                                "Saturday 04 April 2026  10:14:43 +0000 (0:00:00.094)       0:06:27.755 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Get `ceph orch ps`] **********************************",
                                "Saturday 04 April 2026  10:14:43 +0000 (0:00:00.052)       0:06:27.807 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : OSD daemon list] *************************************",
                                "Saturday 04 April 2026  10:14:45 +0000 (0:00:01.910)       0:06:29.718 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Fail if any OSD not running] *************************",
                                "Saturday 04 April 2026  10:14:45 +0000 (0:00:00.054)       0:06:29.772 ********",
                                "skipping: [instance] => (item=1)",
                                "skipping: [instance] => (item=1)",
                                "skipping: [instance] => (item=1)",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.ceph.osd : Fail if any duplicate OSD ID] ************************",
                                "Saturday 04 April 2026  10:14:46 +0000 (0:00:00.054)       0:06:29.827 ********",
                                "skipping: [instance]",
                                "",
                                "PLAY [all] *********************************************************************",
                                "",
                                "TASK [Ensure RBD kernel module is loaded] **************************************",
                                "Saturday 04 April 2026  10:14:46 +0000 (0:00:00.055)       0:06:29.882 ********",
                                "changed: [instance]",
                                "",
                                "PLAY [all] *********************************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:14:46 +0000 (0:00:00.402)       0:06:30.285 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.sysctl : Configure sysctl values] ********************",
                                "Saturday 04 April 2026  10:14:47 +0000 (0:00:01.018)       0:06:31.303 ********",
                                "changed: [instance] => (item={'name': 'fs.aio-max-nr', 'value': 1048576})",
                                "changed: [instance] => (item={'name': 'net.ipv4.tcp_timestamps', 'value': 0})",
                                "changed: [instance] => (item={'name': 'net.ipv4.tcp_sack', 'value': 1})",
                                "changed: [instance] => (item={'name': 'net.core.netdev_budget', 'value': 1000})",
                                "changed: [instance] => (item={'name': 'net.core.netdev_max_backlog', 'value': 250000})",
                                "changed: [instance] => (item={'name': 'net.core.rmem_max', 'value': 4194304})",
                                "changed: [instance] => (item={'name': 'net.core.wmem_max', 'value': 4194304})",
                                "changed: [instance] => (item={'name': 'net.core.rmem_default', 'value': 4194304})",
                                "changed: [instance] => (item={'name': 'net.core.wmem_default', 'value': 4194304})",
                                "changed: [instance] => (item={'name': 'net.core.optmem_max', 'value': 4194304})",
                                "changed: [instance] => (item={'name': 'net.ipv4.tcp_rmem', 'value': '4096 87380 4194304'})",
                                "changed: [instance] => (item={'name': 'net.ipv4.tcp_wmem', 'value': '4096 65536 4194304'})",
                                "changed: [instance] => (item={'name': 'net.ipv4.tcp_low_latency', 'value': 1})",
                                "changed: [instance] => (item={'name': 'net.ipv4.tcp_adv_win_scale', 'value': 1})",
                                "changed: [instance] => (item={'name': 'net.ipv4.neigh.default.gc_thresh1', 'value': 128})",
                                "changed: [instance] => (item={'name': 'net.ipv4.neigh.default.gc_thresh2', 'value': 28872})",
                                "changed: [instance] => (item={'name': 'net.ipv4.neigh.default.gc_thresh3', 'value': 32768})",
                                "changed: [instance] => (item={'name': 'net.ipv6.neigh.default.gc_thresh1', 'value': 128})",
                                "changed: [instance] => (item={'name': 'net.ipv6.neigh.default.gc_thresh2', 'value': 28872})",
                                "changed: [instance] => (item={'name': 'net.ipv6.neigh.default.gc_thresh3', 'value': 32768})",
                                "",
                                "TASK [vexxhost.atmosphere.ethtool : Create folder for persistent configuration] ***",
                                "Saturday 04 April 2026  10:14:53 +0000 (0:00:05.917)       0:06:37.221 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ethtool : Install persistent \"ethtool\" tuning] *******",
                                "Saturday 04 April 2026  10:14:53 +0000 (0:00:00.207)       0:06:37.429 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ethtool : Run \"ethtool\" tuning] **********************",
                                "Saturday 04 April 2026  10:14:54 +0000 (0:00:00.471)       0:06:37.900 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Set a fact with the \"atmosphere_images\" for other plays] *****************",
                                "Saturday 04 April 2026  10:14:54 +0000 (0:00:00.273)       0:06:38.173 ********",
                                "ok: [instance]",
                                "",
                                "PLAY [Configure Kubernetes VIP] ************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:14:54 +0000 (0:00:00.054)       0:06:38.228 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.directory : Create directory (/etc/kubernetes/manifests)] ***",
                                "Saturday 04 April 2026  10:14:55 +0000 (0:00:00.971)       0:06:39.199 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kube_vip : Uninstall legacy HA stack] ****************",
                                "Saturday 04 April 2026  10:14:55 +0000 (0:00:00.217)       0:06:39.416 ********",
                                "ok: [instance] => (item=/etc/keepalived/keepalived.conf)",
                                "ok: [instance] => (item=/etc/keepalived/check_apiserver.sh)",
                                "ok: [instance] => (item=/etc/kubernetes/manifests/keepalived.yaml)",
                                "ok: [instance] => (item=/etc/haproxy/haproxy.cfg)",
                                "ok: [instance] => (item=/etc/kubernetes/manifests/haproxy.yaml)",
                                "",
                                "TASK [vexxhost.kubernetes.kube_vip : Switch API server to run on port 6443] ****",
                                "Saturday 04 April 2026  10:14:56 +0000 (0:00:00.922)       0:06:40.339 ********",
                                "failed: [instance] (item=/etc/kubernetes/manifests/kube-apiserver.yaml) => {\"ansible_loop_var\": \"item\", \"changed\": false, \"item\": \"/etc/kubernetes/manifests/kube-apiserver.yaml\", \"msg\": \"Path /etc/kubernetes/manifests/kube-apiserver.yaml does not exist !\", \"rc\": 257}",
                                "failed: [instance] (item=/etc/kubernetes/controller-manager.conf) => {\"ansible_loop_var\": \"item\", \"changed\": false, \"item\": \"/etc/kubernetes/controller-manager.conf\", \"msg\": \"Path /etc/kubernetes/controller-manager.conf does not exist !\", \"rc\": 257}",
                                "failed: [instance] (item=/etc/kubernetes/scheduler.conf) => {\"ansible_loop_var\": \"item\", \"changed\": false, \"item\": \"/etc/kubernetes/scheduler.conf\", \"msg\": \"Path /etc/kubernetes/scheduler.conf does not exist !\", \"rc\": 257}",
                                "...ignoring",
                                "",
                                "TASK [vexxhost.kubernetes.kube_vip : Check if super-admin.conf exists] *********",
                                "Saturday 04 April 2026  10:14:57 +0000 (0:00:00.591)       0:06:40.930 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kube_vip : Check if kubeadm has already run] *********",
                                "Saturday 04 April 2026  10:14:57 +0000 (0:00:00.191)       0:06:41.122 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kube_vip : Set fact with KUBECONFIG path] ************",
                                "Saturday 04 April 2026  10:14:57 +0000 (0:00:00.191)       0:06:41.314 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kube_vip : Set fact with KUBECONFIG path (with super-admin.conf)] ***",
                                "Saturday 04 April 2026  10:14:57 +0000 (0:00:00.034)       0:06:41.348 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kube_vip : Upload Kubernetes manifest] ***************",
                                "Saturday 04 April 2026  10:14:57 +0000 (0:00:00.045)       0:06:41.394 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kube_vip : Ensure kube-vip configuration file] *******",
                                "Saturday 04 April 2026  10:14:58 +0000 (0:00:00.541)       0:06:41.935 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kube_vip : Flush handlers] ***************************",
                                "Saturday 04 April 2026  10:14:58 +0000 (0:00:00.198)       0:06:42.133 ********",
                                "",
                                "PLAY [Install Kubernetes] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:14:58 +0000 (0:00:00.061)       0:06:42.195 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:14:59 +0000 (0:00:01.144)       0:06:43.339 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.package : Update state for tar] **********************",
                                "Saturday 04 April 2026  10:14:59 +0000 (0:00:00.405)       0:06:43.744 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.directory : Create directory (/var/lib/downloads)] ***",
                                "Saturday 04 April 2026  10:14:59 +0000 (0:00:00.046)       0:06:43.791 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:15:00 +0000 (0:00:00.222)       0:06:44.013 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://github.com/opencontainers/runc/releases/download/v1.4.0/runc.amd64\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:15:00 +0000 (0:00:00.057)       0:06:44.071 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:15:00 +0000 (0:00:00.348)       0:06:44.419 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:15:00 +0000 (0:00:00.051)       0:06:44.471 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.package : Update state for tar] **********************",
                                "Saturday 04 April 2026  10:15:00 +0000 (0:00:00.216)       0:06:44.688 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:15:02 +0000 (0:00:01.162)       0:06:45.850 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://github.com/containerd/containerd/releases/download/v2.2.0/containerd-2.2.0-linux-amd64.tar.gz\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:15:02 +0000 (0:00:00.093)       0:06:45.943 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:15:02 +0000 (0:00:00.379)       0:06:46.323 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Install SELinux packages] ***************",
                                "Saturday 04 April 2026  10:15:04 +0000 (0:00:02.189)       0:06:48.513 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Set SELinux to permissive at runtime] ***",
                                "Saturday 04 April 2026  10:15:04 +0000 (0:00:00.037)       0:06:48.550 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Persist SELinux permissive mode] ********",
                                "Saturday 04 April 2026  10:15:04 +0000 (0:00:00.039)       0:06:48.589 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Install AppArmor packages] **************",
                                "Saturday 04 April 2026  10:15:04 +0000 (0:00:00.042)       0:06:48.632 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Create systemd service file for containerd] ***",
                                "Saturday 04 April 2026  10:15:05 +0000 (0:00:01.195)       0:06:49.827 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Create folders for configuration] *******",
                                "Saturday 04 April 2026  10:15:06 +0000 (0:00:00.451)       0:06:50.279 ********",
                                "ok: [instance] => (item={'path': '/etc/containerd'})",
                                "ok: [instance] => (item={'path': '/var/lib/containerd', 'mode': '0o700'})",
                                "ok: [instance] => (item={'path': '/run/containerd', 'mode': '0o711'})",
                                "ok: [instance] => (item={'path': '/run/containerd/io.containerd.grpc.v1.cri', 'mode': '0o700'})",
                                "ok: [instance] => (item={'path': '/run/containerd/io.containerd.sandbox.controller.v1.shim', 'mode': '0o700'})",
                                "",
                                "TASK [vexxhost.containers.containerd : Create containerd config file] **********",
                                "Saturday 04 April 2026  10:15:07 +0000 (0:00:00.946)       0:06:51.226 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Force any restarts if necessary] ********",
                                "Saturday 04 April 2026  10:15:07 +0000 (0:00:00.545)       0:06:51.771 ********",
                                "",
                                "TASK [vexxhost.containers.containerd : Enable and start service] ***************",
                                "Saturday 04 April 2026  10:15:07 +0000 (0:00:00.006)       0:06:51.778 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes_upgrade_check : Retrieve the \"kubeadm-config\" ConfigMap] ***",
                                "Saturday 04 April 2026  10:15:08 +0000 (0:00:00.376)       0:06:52.154 ********",
                                "fatal: [instance]: FAILED! => {\"changed\": false, \"msg\": \"Failed to import the required Python library (kubernetes) on instance's Python /usr/bin/python3.10. Please read the module documentation and install it in the appropriate location. If the required library is installed, but Ansible is using the wrong Python interpreter, please consult the documentation on ansible_python_interpreter\"}",
                                "...ignoring",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes_upgrade_check : Parse the ClusterConfiguration] ***",
                                "Saturday 04 April 2026  10:15:08 +0000 (0:00:00.670)       0:06:52.825 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes_upgrade_check : Retrieve the current Kubernetes version] ***",
                                "Saturday 04 April 2026  10:15:09 +0000 (0:00:00.036)       0:06:52.862 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes_upgrade_check : Extract major, minor, and patch versions] ***",
                                "Saturday 04 April 2026  10:15:09 +0000 (0:00:00.033)       0:06:52.896 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes_upgrade_check : Fail if we're jumping more than one minor version] ***",
                                "Saturday 04 April 2026  10:15:09 +0000 (0:00:00.039)       0:06:52.935 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes_upgrade_check : Set fact if we need to upgrade] ***",
                                "Saturday 04 April 2026  10:15:09 +0000 (0:00:00.044)       0:06:52.979 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:15:09 +0000 (0:00:00.058)       0:06:53.037 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:15:09 +0000 (0:00:00.219)       0:06:53.256 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://dl.k8s.io/release/v1.28.13/bin/linux/amd64/kubeadm\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:15:09 +0000 (0:00:00.043)       0:06:53.300 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:15:10 +0000 (0:00:00.743)       0:06:54.044 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:15:10 +0000 (0:00:00.049)       0:06:54.093 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:15:10 +0000 (0:00:00.213)       0:06:54.306 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://dl.k8s.io/release/v1.28.13/bin/linux/amd64/kubectl\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:15:10 +0000 (0:00:00.047)       0:06:54.354 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:15:11 +0000 (0:00:00.873)       0:06:55.228 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Install SELinux packages] ***************",
                                "Saturday 04 April 2026  10:15:11 +0000 (0:00:00.054)       0:06:55.282 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Set SELinux to permissive at runtime] ***",
                                "Saturday 04 April 2026  10:15:11 +0000 (0:00:00.038)       0:06:55.320 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Persist SELinux permissive mode] ********",
                                "Saturday 04 April 2026  10:15:11 +0000 (0:00:00.035)       0:06:55.356 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Install AppArmor packages] **************",
                                "Saturday 04 April 2026  10:15:11 +0000 (0:00:00.027)       0:06:55.383 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Create systemd service file for containerd] ***",
                                "Saturday 04 April 2026  10:15:12 +0000 (0:00:01.095)       0:06:56.479 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Create folders for configuration] *******",
                                "Saturday 04 April 2026  10:15:13 +0000 (0:00:00.458)       0:06:56.937 ********",
                                "ok: [instance] => (item={'path': '/etc/containerd'})",
                                "ok: [instance] => (item={'path': '/var/lib/containerd', 'mode': '0o700'})",
                                "ok: [instance] => (item={'path': '/run/containerd', 'mode': '0o711'})",
                                "ok: [instance] => (item={'path': '/run/containerd/io.containerd.grpc.v1.cri', 'mode': '0o700'})",
                                "ok: [instance] => (item={'path': '/run/containerd/io.containerd.sandbox.controller.v1.shim', 'mode': '0o700'})",
                                "",
                                "TASK [vexxhost.containers.containerd : Create containerd config file] **********",
                                "Saturday 04 April 2026  10:15:14 +0000 (0:00:00.949)       0:06:57.887 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.containerd : Force any restarts if necessary] ********",
                                "Saturday 04 April 2026  10:15:14 +0000 (0:00:00.504)       0:06:58.391 ********",
                                "",
                                "TASK [vexxhost.containers.containerd : Enable and start service] ***************",
                                "Saturday 04 April 2026  10:15:14 +0000 (0:00:00.008)       0:06:58.400 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:15:14 +0000 (0:00:00.377)       0:06:58.778 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:15:15 +0000 (0:00:00.214)       0:06:58.992 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://github.com/kubernetes-sigs/cri-tools/releases/download/v1.34.0/crictl-v1.34.0-linux-amd64.tar.gz\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:15:15 +0000 (0:00:00.061)       0:06:59.053 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:15:15 +0000 (0:00:00.574)       0:06:59.628 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:15:17 +0000 (0:00:01.425)       0:07:01.054 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://github.com/kubernetes-sigs/cri-tools/releases/download/v1.34.0/critest-v1.34.0-linux-amd64.tar.gz\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:15:17 +0000 (0:00:00.060)       0:07:01.115 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:15:17 +0000 (0:00:00.544)       0:07:01.659 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.cri_tools : Create crictl config] ********************",
                                "Saturday 04 April 2026  10:15:19 +0000 (0:00:01.646)       0:07:03.305 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.directory : Create directory (/opt/cni/bin)] *********",
                                "Saturday 04 April 2026  10:15:19 +0000 (0:00:00.468)       0:07:03.774 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:15:20 +0000 (0:00:00.222)       0:07:03.997 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:15:20 +0000 (0:00:00.274)       0:07:04.271 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://github.com/containernetworking/plugins/releases/download/v1.8.0/cni-plugins-linux-amd64-v1.8.0.tgz\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:15:20 +0000 (0:00:00.070)       0:07:04.342 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:15:21 +0000 (0:00:00.892)       0:07:05.234 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.cni_plugins : Gather variables for each operating system] ***",
                                "Saturday 04 April 2026  10:15:24 +0000 (0:00:02.917)       0:07:08.151 ********",
                                "ok: [instance] => (item=/home/zuul/.ansible/collections/ansible_collections/vexxhost/containers/roles/cni_plugins/vars/debian.yml)",
                                "",
                                "TASK [vexxhost.containers.cni_plugins : Install additional packages] ***********",
                                "Saturday 04 April 2026  10:15:24 +0000 (0:00:00.083)       0:07:08.235 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.cni_plugins : Ensure IPv6 is enabled] ****************",
                                "Saturday 04 April 2026  10:15:25 +0000 (0:00:01.062)       0:07:09.297 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.cni_plugins : Enable kernel modules on-boot] *********",
                                "Saturday 04 April 2026  10:15:25 +0000 (0:00:00.208)       0:07:09.506 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.cni_plugins : Enable kernel modules in runtime] ******",
                                "Saturday 04 April 2026  10:15:26 +0000 (0:00:00.476)       0:07:09.982 ********",
                                "changed: [instance] => (item=br_netfilter)",
                                "ok: [instance] => (item=ip_tables)",
                                "changed: [instance] => (item=ip6_tables)",
                                "ok: [instance] => (item=nf_conntrack)",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:15:26 +0000 (0:00:00.845)       0:07:10.828 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:15:27 +0000 (0:00:00.234)       0:07:11.062 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://dl.k8s.io/release/v1.28.13/bin/linux/amd64/kubelet\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:15:27 +0000 (0:00:00.055)       0:07:11.118 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:15:28 +0000 (0:00:01.131)       0:07:12.250 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Gather variables for each operating system] ***",
                                "Saturday 04 April 2026  10:15:28 +0000 (0:00:00.044)       0:07:12.294 ********",
                                "ok: [instance] => (item=/home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/kubelet/vars/debian.yml)",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Install coreutils] *************************",
                                "Saturday 04 April 2026  10:15:28 +0000 (0:00:00.065)       0:07:12.360 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Install additional packages] ***************",
                                "Saturday 04 April 2026  10:15:28 +0000 (0:00:00.050)       0:07:12.410 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Configure sysctl values] *******************",
                                "Saturday 04 April 2026  10:15:32 +0000 (0:00:04.314)       0:07:16.725 ********",
                                "changed: [instance] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})",
                                "changed: [instance] => (item={'name': 'net.bridge.bridge-nf-call-iptables', 'value': 1})",
                                "changed: [instance] => (item={'name': 'net.bridge.bridge-nf-call-ip6tables', 'value': 1})",
                                "changed: [instance] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 0})",
                                "changed: [instance] => (item={'name': 'fs.inotify.max_queued_events', 'value': 1048576})",
                                "changed: [instance] => (item={'name': 'fs.inotify.max_user_instances', 'value': 8192})",
                                "changed: [instance] => (item={'name': 'fs.inotify.max_user_watches', 'value': 1048576})",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Create folders for kubernetes configuration] ***",
                                "Saturday 04 April 2026  10:15:36 +0000 (0:00:03.344)       0:07:20.069 ********",
                                "changed: [instance] => (item=/etc/systemd/system/kubelet.service.d)",
                                "ok: [instance] => (item=/etc/kubernetes)",
                                "ok: [instance] => (item=/etc/kubernetes/manifests)",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Add kubelet systemd service config] ********",
                                "Saturday 04 April 2026  10:15:36 +0000 (0:00:00.555)       0:07:20.625 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Add kubeadm dropin for kubelet systemd service config] ***",
                                "Saturday 04 April 2026  10:15:37 +0000 (0:00:00.470)       0:07:21.096 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Check swap status] *************************",
                                "Saturday 04 April 2026  10:15:37 +0000 (0:00:00.463)       0:07:21.559 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Disable swap] ******************************",
                                "Saturday 04 April 2026  10:15:37 +0000 (0:00:00.204)       0:07:21.764 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Remove swapfile from /etc/fstab] ***********",
                                "Saturday 04 April 2026  10:15:37 +0000 (0:00:00.038)       0:07:21.803 ********",
                                "ok: [instance] => (item=swap)",
                                "ok: [instance] => (item=none)",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Create noswap systemd service config file] ***",
                                "Saturday 04 April 2026  10:15:38 +0000 (0:00:00.595)       0:07:22.398 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Enable noswap service] *********************",
                                "Saturday 04 April 2026  10:15:39 +0000 (0:00:00.481)       0:07:22.879 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Force any restarts if necessary] ***********",
                                "Saturday 04 April 2026  10:15:39 +0000 (0:00:00.645)       0:07:23.524 ********",
                                "",
                                "RUNNING HANDLER [vexxhost.kubernetes.kubelet : Reload systemd] *****************",
                                "Saturday 04 April 2026  10:15:39 +0000 (0:00:00.007)       0:07:23.532 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Enable and start kubelet service] **********",
                                "Saturday 04 April 2026  10:15:40 +0000 (0:00:00.758)       0:07:24.290 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Ensure availability of dbus on Debian] *****",
                                "Saturday 04 April 2026  10:15:41 +0000 (0:00:00.600)       0:07:24.890 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Configure short hostname] ******************",
                                "Saturday 04 April 2026  10:15:42 +0000 (0:00:01.010)       0:07:25.901 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubelet : Ensure hostname inside hosts file] *********",
                                "Saturday 04 April 2026  10:15:42 +0000 (0:00:00.766)       0:07:26.667 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Remove kubernetes repository] ***********",
                                "Saturday 04 April 2026  10:15:43 +0000 (0:00:00.246)       0:07:26.913 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Setup control plane] ********************",
                                "Saturday 04 April 2026  10:15:43 +0000 (0:00:00.544)       0:07:27.458 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/kubernetes/tasks/control-plane.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Bootstrap cluster] **********************",
                                "Saturday 04 April 2026  10:15:43 +0000 (0:00:00.079)       0:07:27.538 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/kubernetes/tasks/bootstrap-cluster.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Check if any control plane is bootstrapped] ***",
                                "Saturday 04 April 2026  10:15:43 +0000 (0:00:00.083)       0:07:27.622 ********",
                                "ok: [instance] => (item=instance)",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Pick node from pre-existing cluster] ****",
                                "Saturday 04 April 2026  10:15:44 +0000 (0:00:00.236)       0:07:27.858 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Select first node to initialize cluster] ***",
                                "Saturday 04 April 2026  10:15:44 +0000 (0:00:00.039)       0:07:27.898 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Print selected bootstrap node] **********",
                                "Saturday 04 April 2026  10:15:44 +0000 (0:00:00.051)       0:07:27.950 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"instance\"",
                                "}",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Upload cluster configuration for bootstrap node] ***",
                                "Saturday 04 April 2026  10:15:44 +0000 (0:00:00.049)       0:07:27.999 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Create pki folder] **********************",
                                "Saturday 04 April 2026  10:15:44 +0000 (0:00:00.566)       0:07:28.566 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Create kubernetes ca key] ***************",
                                "Saturday 04 April 2026  10:15:44 +0000 (0:00:00.037)       0:07:28.603 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Create kubernetes ca cert] **************",
                                "Saturday 04 April 2026  10:15:44 +0000 (0:00:00.032)       0:07:28.635 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Create etcd-ca key] *********************",
                                "Saturday 04 April 2026  10:15:44 +0000 (0:00:00.039)       0:07:28.675 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Create etcd-ca cert] ********************",
                                "Saturday 04 April 2026  10:15:44 +0000 (0:00:00.037)       0:07:28.712 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Create front-proxy-ca key] **************",
                                "Saturday 04 April 2026  10:15:44 +0000 (0:00:00.044)       0:07:28.757 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Create front-proxy-ca cert] *************",
                                "Saturday 04 April 2026  10:15:44 +0000 (0:00:00.041)       0:07:28.798 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Initialize cluster] *********************",
                                "Saturday 04 April 2026  10:15:45 +0000 (0:00:00.040)       0:07:28.839 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Check if the node is already part of the cluster] ***",
                                "Saturday 04 April 2026  10:16:14 +0000 (0:00:29.735)       0:07:58.574 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Join cluster] ***************************",
                                "Saturday 04 April 2026  10:16:14 +0000 (0:00:00.200)       0:07:58.775 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Create folder for admin configuration] ***",
                                "Saturday 04 April 2026  10:16:14 +0000 (0:00:00.042)       0:07:58.818 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Creating a symlink for admin configuration file] ***",
                                "Saturday 04 April 2026  10:16:15 +0000 (0:00:00.202)       0:07:59.020 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Add bash autocomplete for kubectl] ******",
                                "Saturday 04 April 2026  10:16:15 +0000 (0:00:00.220)       0:07:59.240 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Install PIP] ****************************",
                                "Saturday 04 April 2026  10:16:15 +0000 (0:00:00.196)       0:07:59.437 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Install Kubernetes Python package using pip on supported systems] ***",
                                "Saturday 04 April 2026  10:16:20 +0000 (0:00:05.003)       0:08:04.440 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Enable EPEL repository] *****************",
                                "Saturday 04 April 2026  10:16:25 +0000 (0:00:04.696)       0:08:09.137 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Install Kubernetes Python package using package manager on supported systems] ***",
                                "Saturday 04 April 2026  10:16:25 +0000 (0:00:00.046)       0:08:09.184 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Allow workload on control plane node] ***",
                                "Saturday 04 April 2026  10:16:25 +0000 (0:00:00.036)       0:08:09.220 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Remove kube-proxy resources] ************",
                                "Saturday 04 April 2026  10:16:26 +0000 (0:00:00.794)       0:08:10.015 ********",
                                "skipping: [instance] => (item=DaemonSet)",
                                "skipping: [instance] => (item=ConfigMap)",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Start an upgrade] ***********************",
                                "Saturday 04 April 2026  10:16:26 +0000 (0:00:00.043)       0:08:10.059 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Check if the Kubernetes API services is up to date] ***",
                                "Saturday 04 April 2026  10:16:26 +0000 (0:00:00.044)       0:08:10.104 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Trigger an upgrade of the Kubernetes API services] ***",
                                "Saturday 04 April 2026  10:16:26 +0000 (0:00:00.039)       0:08:10.143 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Setup nodes] ****************************",
                                "Saturday 04 April 2026  10:16:26 +0000 (0:00:00.046)       0:08:10.190 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Upgrade & restart Kubelet node for upgrade (if needed)] ***",
                                "Saturday 04 April 2026  10:16:26 +0000 (0:00:00.039)       0:08:10.229 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.kubernetes : Set node selector for CoreDNS components] ***",
                                "Saturday 04 April 2026  10:16:26 +0000 (0:00:00.521)       0:08:10.751 ********",
                                "changed: [instance]",
                                "",
                                "PLAY [Install control-plane components] ****************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:16:27 +0000 (0:00:00.951)       0:08:11.703 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.forget_package : Forget package] *********************",
                                "Saturday 04 April 2026  10:16:28 +0000 (0:00:00.969)       0:08:12.672 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.package : Update state for tar] **********************",
                                "Saturday 04 April 2026  10:16:29 +0000 (0:00:00.194)       0:08:12.867 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.directory : Create directory (/var/lib/downloads)] ***",
                                "Saturday 04 April 2026  10:16:30 +0000 (0:00:01.110)       0:08:13.977 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:16:30 +0000 (0:00:00.459)       0:08:14.436 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://get.helm.sh/helm-v3.11.2-linux-amd64.tar.gz\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:16:30 +0000 (0:00:00.042)       0:08:14.478 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:16:31 +0000 (0:00:00.514)       0:08:14.992 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.helm : Remove Helm repository] ***********************",
                                "Saturday 04 April 2026  10:16:32 +0000 (0:00:01.474)       0:08:16.467 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.helm : Add bash autocomplete for helm] ***************",
                                "Saturday 04 April 2026  10:16:32 +0000 (0:00:00.336)       0:08:16.803 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.helm : Get Helm plugins dir] *************************",
                                "Saturday 04 April 2026  10:16:33 +0000 (0:00:00.190)       0:08:16.994 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.helm : Create Helm plugins directory if it does not exist] ***",
                                "Saturday 04 April 2026  10:16:33 +0000 (0:00:00.245)       0:08:17.240 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.helm : Remove Helm diff plugin installed by kubernetes.core.helm_plugin] ***",
                                "Saturday 04 April 2026  10:16:33 +0000 (0:00:00.200)       0:08:17.440 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Install plugin] **********************************************************",
                                "Saturday 04 April 2026  10:16:33 +0000 (0:00:00.195)       0:08:17.636 ********",
                                "included: vexxhost.containers.download_artifact for instance",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Starting download of file] *******",
                                "Saturday 04 April 2026  10:16:33 +0000 (0:00:00.051)       0:08:17.688 ********",
                                "ok: [instance] => {",
                                "    \"msg\": \"https://github.com/databus23/helm-diff/releases/download/v3.8.1/helm-diff-linux-amd64.tgz\"",
                                "}",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Download item] *******************",
                                "Saturday 04 April 2026  10:16:33 +0000 (0:00:00.039)       0:08:17.728 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.containers.download_artifact : Extract archive] *****************",
                                "Saturday 04 April 2026  10:16:34 +0000 (0:00:00.968)       0:08:18.696 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:16:36 +0000 (0:00:02.026)       0:08:20.722 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:16:36 +0000 (0:00:00.042)       0:08:20.765 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.cilium : Get current Kubernetes version] *************",
                                "Saturday 04 April 2026  10:16:37 +0000 (0:00:00.634)       0:08:21.400 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.cilium : Deploy Helm chart] **************************",
                                "Saturday 04 April 2026  10:16:38 +0000 (0:00:00.870)       0:08:22.271 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:16:41 +0000 (0:00:02.598)       0:08:24.869 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:16:41 +0000 (0:00:00.032)       0:08:24.902 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.envoy_gateway : Deploy Helm chart] *******************",
                                "Saturday 04 April 2026  10:16:41 +0000 (0:00:00.481)       0:08:25.383 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.envoy_gateway : Create GatewayClass] *****************",
                                "Saturday 04 April 2026  10:17:23 +0000 (0:00:42.116)       0:09:07.500 ********",
                                "changed: [instance]",
                                "",
                                "PLAY [all] *********************************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:17:24 +0000 (0:00:00.803)       0:09:08.303 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kubernetes_node_labels : Add labels to node] *********",
                                "Saturday 04 April 2026  10:17:25 +0000 (0:00:01.127)       0:09:09.430 ********",
                                "changed: [instance]",
                                "",
                                "PLAY [all] *********************************************************************",
                                "",
                                "TASK [Uninstall unattended-upgrades] *******************************************",
                                "Saturday 04 April 2026  10:17:26 +0000 (0:00:00.688)       0:09:10.119 ********",
                                "ok: [instance]",
                                "",
                                "PLAY [controllers] *************************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:17:27 +0000 (0:00:00.915)       0:09:11.034 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:17:28 +0000 (0:00:01.222)       0:09:12.257 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:17:28 +0000 (0:00:00.045)       0:09:12.303 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.local_path_provisioner : Deploy Helm chart] **********",
                                "Saturday 04 April 2026  10:17:28 +0000 (0:00:00.465)       0:09:12.768 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:17:30 +0000 (0:00:01.905)       0:09:14.673 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ceph_csi_rbd : Collect \"ceph mon dump\" output from a monitor] ***",
                                "Saturday 04 April 2026  10:17:30 +0000 (0:00:00.051)       0:09:14.725 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ceph_csi_rbd : Generate fact with list of Ceph monitors] ***",
                                "Saturday 04 April 2026  10:17:30 +0000 (0:00:00.050)       0:09:14.776 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ceph_csi_rbd : Create Ceph pool] *********************",
                                "Saturday 04 April 2026  10:17:30 +0000 (0:00:00.036)       0:09:14.812 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ceph_csi_rbd : Create user client.kube] **************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.051)       0:09:14.863 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ceph_csi_rbd : Retrieve keyring for client.kube] *****",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.056)       0:09:14.920 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ceph_csi_rbd : Store keyring inside fact] ************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.057)       0:09:14.977 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ceph_csi_rbd : Suspend the existing HelmRelease] *****",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.043)       0:09:15.021 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ceph_csi_rbd : Remove the existing HelmRelease] ******",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.035)       0:09:15.057 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ceph_csi_rbd : Deploy Helm chart] ********************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.038)       0:09:15.095 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.powerstore_csi : Clone PowerStore CSI from GitHub] ***",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.049)       0:09:15.145 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.powerstore_csi : Create Secret] **********************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.041)       0:09:15.187 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.powerstore_csi : Create StorageClass] ****************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.039)       0:09:15.226 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.powerstore_csi : Deploy Helm chart] ******************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.036)       0:09:15.263 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.multipathd : Add backports PPA] **********************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.056)       0:09:15.319 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.multipathd : Install the multipathd package] *********",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.049)       0:09:15.369 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.multipathd : Install the configuration file] *********",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.042)       0:09:15.411 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.portworx : Detect if InitiatorName is set] ***********",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.047)       0:09:15.459 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.portworx : Generate a new InitiatorName] *************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.042)       0:09:15.501 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.portworx : Write the new InitiatorName] **************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.039)       0:09:15.540 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.portworx : Create namespace] *************************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.042)       0:09:15.582 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.portworx : Install Portworx] *************************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.040)       0:09:15.623 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.portworx : Wait till the CRDs are created] ***********",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.032)       0:09:15.655 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.portworx : Create Portworx Storage Cluster] **********",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.036)       0:09:15.691 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.storpool_csi : Deploy CSI RBAC] **********************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.057)       0:09:15.748 ********",
                                "skipping: [instance] => (item={'name': 'controllerplugin'})",
                                "skipping: [instance] => (item={'name': 'nodeplugin'})",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.storpool_csi : Deploy CSI] ***************************",
                                "Saturday 04 April 2026  10:17:31 +0000 (0:00:00.043)       0:09:15.792 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.storpool_csi : Create StorageClass] ******************",
                                "Saturday 04 April 2026  10:17:32 +0000 (0:00:00.038)       0:09:15.831 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ibm_block_csi_driver : Deploy CSI] *******************",
                                "Saturday 04 April 2026  10:17:32 +0000 (0:00:00.051)       0:09:15.882 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ibm_block_csi_driver : Create Secret] ****************",
                                "Saturday 04 April 2026  10:17:32 +0000 (0:00:00.033)       0:09:15.915 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ibm_block_csi_driver : Create StorageClass] **********",
                                "Saturday 04 April 2026  10:17:32 +0000 (0:00:00.033)       0:09:15.949 ********",
                                "skipping: [instance]",
                                "",
                                "PLAY [Deploy Infrastructure] ***************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:17:32 +0000 (0:00:00.143)       0:09:16.093 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.cert_manager : Suspend the existing HelmRelease] *****",
                                "Saturday 04 April 2026  10:17:33 +0000 (0:00:01.216)       0:09:17.309 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.cert_manager : Remove the existing HelmRelease] ******",
                                "Saturday 04 April 2026  10:17:34 +0000 (0:00:00.658)       0:09:17.968 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Deploy Helm chart] *******************************************************",
                                "Saturday 04 April 2026  10:17:34 +0000 (0:00:00.740)       0:09:18.709 ********",
                                "included: vexxhost.kubernetes.cert_manager for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:17:34 +0000 (0:00:00.063)       0:09:18.772 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:17:34 +0000 (0:00:00.054)       0:09:18.827 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.cert_manager : Deploy Helm chart] ********************",
                                "Saturday 04 April 2026  10:17:35 +0000 (0:00:00.501)       0:09:19.328 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.cluster_issuer : Create self-signed cluster issuer] ***",
                                "Saturday 04 April 2026  10:18:07 +0000 (0:00:31.837)       0:09:51.166 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.cluster_issuer : Import tasks for ClusterIssuer type] ***",
                                "Saturday 04 April 2026  10:18:08 +0000 (0:00:00.706)       0:09:51.872 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/atmosphere/roles/cluster_issuer/tasks/type/self-signed/main.yml for instance",
                                "",
                                "TASK [vexxhost.atmosphere.cluster_issuer : Create ClusterIssuer] ***************",
                                "Saturday 04 April 2026  10:18:08 +0000 (0:00:00.050)       0:09:51.922 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.cluster_issuer : Wait till the secret is created] ****",
                                "Saturday 04 April 2026  10:18:08 +0000 (0:00:00.775)       0:09:52.698 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.cluster_issuer : Copy CA certificate on host] ********",
                                "Saturday 04 April 2026  10:18:10 +0000 (0:00:01.868)       0:09:54.566 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.cluster_issuer : Flush all handlers] *****************",
                                "Saturday 04 April 2026  10:18:11 +0000 (0:00:00.530)       0:09:55.097 ********",
                                "",
                                "RUNNING HANDLER [vexxhost.atmosphere.cluster_issuer : Update CA certificates on host] ***",
                                "Saturday 04 April 2026  10:18:11 +0000 (0:00:00.012)       0:09:55.109 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:18:13 +0000 (0:00:01.738)       0:09:56.848 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:18:13 +0000 (0:00:00.046)       0:09:56.894 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ingress_nginx : Remove the existing HelmRelease] *****",
                                "Saturday 04 April 2026  10:18:13 +0000 (0:00:00.510)       0:09:57.404 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ingress_nginx : Deploy Helm chart] *******************",
                                "Saturday 04 April 2026  10:18:14 +0000 (0:00:00.786)       0:09:58.191 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:18:35 +0000 (0:00:21.074)       0:10:19.265 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:18:35 +0000 (0:00:00.044)       0:10:19.310 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.rabbitmq_cluster_operator : Suspend the existing HelmRelease] ***",
                                "Saturday 04 April 2026  10:18:35 +0000 (0:00:00.462)       0:10:19.772 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.rabbitmq_cluster_operator : Remove the existing HelmRelease] ***",
                                "Saturday 04 April 2026  10:18:36 +0000 (0:00:00.897)       0:10:20.670 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.rabbitmq_cluster_operator : Deploy Helm chart] *******",
                                "Saturday 04 April 2026  10:18:37 +0000 (0:00:00.859)       0:10:21.530 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:18:43 +0000 (0:00:05.317)       0:10:26.848 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:18:43 +0000 (0:00:00.052)       0:10:26.901 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster_operator : Suspend the existing HelmRelease] ***",
                                "Saturday 04 April 2026  10:18:43 +0000 (0:00:00.456)       0:10:27.357 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster_operator : Remove the existing HelmRelease] ***",
                                "Saturday 04 April 2026  10:18:44 +0000 (0:00:00.685)       0:10:28.042 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster_operator : Deploy Helm chart] ***",
                                "Saturday 04 April 2026  10:18:45 +0000 (0:00:00.799)       0:10:28.842 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Check if the Percona XtraDB cluster secret exists] ***",
                                "Saturday 04 April 2026  10:18:49 +0000 (0:00:04.973)       0:10:33.815 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Create a secret] ************",
                                "Saturday 04 April 2026  10:18:50 +0000 (0:00:00.755)       0:10:34.571 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Check if the Percona XtraDB cluster exists] ***",
                                "Saturday 04 April 2026  10:18:52 +0000 (0:00:01.280)       0:10:35.852 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Get current status of the cluster] ***",
                                "Saturday 04 April 2026  10:18:52 +0000 (0:00:00.717)       0:10:36.569 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Assert that the cluster is healthy before upgrade] ***",
                                "Saturday 04 April 2026  10:18:52 +0000 (0:00:00.039)       0:10:36.609 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Stop PXC-operator] **********",
                                "Saturday 04 April 2026  10:18:52 +0000 (0:00:00.044)       0:10:36.653 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Change the cluster Statefulset image to 8.0] ***",
                                "Saturday 04 April 2026  10:18:52 +0000 (0:00:00.041)       0:10:36.695 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Wait until the cluster Statefulset rollout] ***",
                                "Saturday 04 April 2026  10:18:52 +0000 (0:00:00.040)       0:10:36.736 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Update pxc cluster spec] ****",
                                "Saturday 04 April 2026  10:18:52 +0000 (0:00:00.047)       0:10:36.784 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Start PXC-operator] *********",
                                "Saturday 04 April 2026  10:18:52 +0000 (0:00:00.043)       0:10:36.827 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Apply Percona XtraDB cluster] ***",
                                "Saturday 04 April 2026  10:18:53 +0000 (0:00:00.042)       0:10:36.869 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.percona_xtradb_cluster : Create percona haproxy metric service] ***",
                                "Saturday 04 April 2026  10:20:12 +0000 (0:01:19.271)       0:11:56.140 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Retrieve list of all the needed endpoints] ***",
                                "Saturday 04 April 2026  10:20:14 +0000 (0:00:01.788)       0:11:57.929 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [Create RabbitMQ cluster] *************************************************",
                                "Saturday 04 April 2026  10:20:14 +0000 (0:00:00.046)       0:11:57.976 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Grab RabbitMQ cluster secret] ***",
                                "Saturday 04 April 2026  10:20:14 +0000 (0:00:00.039)       0:11:58.016 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Cache fact with RabbitMQ cluster credentials] ***",
                                "Saturday 04 April 2026  10:20:14 +0000 (0:00:00.035)       0:11:58.052 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Grab Percona XtraDB cluster secret] ***",
                                "Saturday 04 April 2026  10:20:14 +0000 (0:00:00.039)       0:11:58.091 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Cache fact with Percona XtraDB password] ***",
                                "Saturday 04 April 2026  10:20:15 +0000 (0:00:00.908)       0:11:58.999 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Reset value for OpenStack_Helm endpoints] ***",
                                "Saturday 04 April 2026  10:20:15 +0000 (0:00:00.041)       0:11:59.041 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Generate OpenStack-Helm endpoints] ***",
                                "Saturday 04 April 2026  10:20:15 +0000 (0:00:00.055)       0:11:59.097 ********",
                                "ok: [instance] => (item=oslo_db)",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Clean-up facts] ***********",
                                "Saturday 04 April 2026  10:20:15 +0000 (0:00:00.045)       0:11:59.142 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:20:15 +0000 (0:00:00.060)       0:11:59.203 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:20:15 +0000 (0:00:00.048)       0:11:59.251 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.keycloak : Get the Kuberentes service for Percona XtraDB Cluster] ***",
                                "Saturday 04 April 2026  10:20:15 +0000 (0:00:00.519)       0:11:59.771 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.keycloak : Install MySQL python package] *************",
                                "Saturday 04 April 2026  10:20:16 +0000 (0:00:00.725)       0:12:00.496 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.keycloak : Check MySQL ready] ************************",
                                "Saturday 04 April 2026  10:20:18 +0000 (0:00:02.032)       0:12:02.528 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.keycloak : Create Keycloak database] *****************",
                                "Saturday 04 April 2026  10:20:19 +0000 (0:00:00.469)       0:12:02.997 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.keycloak : Create a Keycloak user] *******************",
                                "Saturday 04 April 2026  10:20:19 +0000 (0:00:00.510)       0:12:03.508 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.keycloak : Disable pxc strict mode] ******************",
                                "Saturday 04 April 2026  10:20:20 +0000 (0:00:00.512)       0:12:04.020 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.keycloak : Deploy Helm chart] ************************",
                                "Saturday 04 April 2026  10:20:20 +0000 (0:00:00.539)       0:12:04.560 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.keycloak : Wait until keycloak ready] ****************",
                                "Saturday 04 April 2026  10:22:24 +0000 (0:02:03.477)       0:14:08.037 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Create Keycloak Ingress] *************************************************",
                                "Saturday 04 April 2026  10:22:24 +0000 (0:00:00.742)       0:14:08.780 ********",
                                "included: ingress for instance",
                                "",
                                "TASK [vexxhost.atmosphere.ingress : Create Ingress keycloak] *******************",
                                "Saturday 04 April 2026  10:22:25 +0000 (0:00:00.060)       0:14:08.841 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.keycloak : Enable pxc strict mode] *******************",
                                "Saturday 04 April 2026  10:22:25 +0000 (0:00:00.931)       0:14:09.772 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.keepalived : Deploy service] *************************",
                                "Saturday 04 April 2026  10:22:26 +0000 (0:00:00.284)       0:14:10.057 ********",
                                "changed: [instance]",
                                "",
                                "PLAY [Deploy Monitoring] *******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:22:27 +0000 (0:00:00.904)       0:14:10.962 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:22:28 +0000 (0:00:01.497)       0:14:12.459 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:22:28 +0000 (0:00:00.049)       0:14:12.508 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.node_feature_discovery : Suspend the existing HelmRelease] ***",
                                "Saturday 04 April 2026  10:22:29 +0000 (0:00:00.515)       0:14:13.024 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.node_feature_discovery : Remove the existing HelmRelease] ***",
                                "Saturday 04 April 2026  10:22:30 +0000 (0:00:00.870)       0:14:13.895 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.node_feature_discovery : Deploy Helm chart] **********",
                                "Saturday 04 April 2026  10:22:30 +0000 (0:00:00.883)       0:14:14.779 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:22:34 +0000 (0:00:03.662)       0:14:18.441 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:22:34 +0000 (0:00:00.055)       0:14:18.496 ********",
                                "changed: [instance]",
                                "",
                                "TASK [atmosphere.common.secretgen_controller : Deploy secretgen-controller] ****",
                                "Saturday 04 April 2026  10:22:35 +0000 (0:00:00.569)       0:14:19.065 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Uninstall the legacy \"ethtool-exporter\"] ***",
                                "Saturday 04 April 2026  10:22:36 +0000 (0:00:01.450)       0:14:20.516 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Suspend the existing HelmRelease] ***",
                                "Saturday 04 April 2026  10:22:37 +0000 (0:00:01.188)       0:14:21.704 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Remove the existing HelmRelease] ***",
                                "Saturday 04 April 2026  10:22:38 +0000 (0:00:01.088)       0:14:22.792 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Wait until Keycloak service is ready] ***",
                                "Saturday 04 April 2026  10:22:39 +0000 (0:00:00.814)       0:14:23.607 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Create Keycloak realm] *******",
                                "Saturday 04 April 2026  10:22:40 +0000 (0:00:00.709)       0:14:24.317 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Add client roles in \"id_token\"] ***",
                                "Saturday 04 April 2026  10:22:43 +0000 (0:00:03.141)       0:14:27.458 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Retrieve \"etcd\" CA certificate] ***",
                                "Saturday 04 April 2026  10:22:44 +0000 (0:00:00.807)       0:14:28.266 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Retrieve \"etcd\" client certificate] ***",
                                "Saturday 04 April 2026  10:22:44 +0000 (0:00:00.356)       0:14:28.622 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Retrieve \"etcd\" client key] ***",
                                "Saturday 04 April 2026  10:22:45 +0000 (0:00:00.206)       0:14:28.829 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Create secrets for monitoring] ***",
                                "Saturday 04 April 2026  10:22:45 +0000 (0:00:00.199)       0:14:29.029 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Generate client secret passwords] ***",
                                "Saturday 04 April 2026  10:22:46 +0000 (0:00:00.982)       0:14:30.012 ********",
                                "changed: [instance] => (item=alertmanager)",
                                "changed: [instance] => (item=grafana)",
                                "changed: [instance] => (item=prometheus)",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Collect all client secrets] ***",
                                "Saturday 04 April 2026  10:23:03 +0000 (0:00:17.453)       0:14:47.465 ********",
                                "ok: [instance] => (item=alertmanager)",
                                "ok: [instance] => (item=grafana)",
                                "ok: [instance] => (item=prometheus)",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Create Keycloak clients] *****",
                                "Saturday 04 April 2026  10:23:05 +0000 (0:00:01.995)       0:14:49.461 ********",
                                "changed: [instance] => (item=None)",
                                "changed: [instance] => (item=None)",
                                "changed: [instance] => (item=None)",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Create Keycloak roles] *******",
                                "Saturday 04 April 2026  10:23:07 +0000 (0:00:01.760)       0:14:51.222 ********",
                                "changed: [instance] => (item=None)",
                                "changed: [instance] => (item=None)",
                                "changed: [instance] => (item=None)",
                                "changed: [instance] => (item=None)",
                                "changed: [instance] => (item=None)",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Generate cookie secrets] *****",
                                "Saturday 04 April 2026  10:23:10 +0000 (0:00:02.847)       0:14:54.069 ********",
                                "changed: [instance] => (item=alertmanager)",
                                "changed: [instance] => (item=prometheus)",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Generate OAuth2 proxy configuration] ***",
                                "Saturday 04 April 2026  10:23:21 +0000 (0:00:11.624)       0:15:05.693 ********",
                                "changed: [instance] => (item=alertmanager)",
                                "changed: [instance] => (item=prometheus)",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Create certificate issuer] ***",
                                "Saturday 04 April 2026  10:23:33 +0000 (0:00:11.853)       0:15:17.547 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Install all CRDs] ************",
                                "Saturday 04 April 2026  10:23:34 +0000 (0:00:00.700)       0:15:18.247 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Deploy additional dashboards] ***",
                                "Saturday 04 April 2026  10:23:41 +0000 (0:00:07.042)       0:15:25.290 ********",
                                "changed: [instance] => (item={'name': 'haproxy', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'goldpinger', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'node-exporter-full', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'ceph-cluster', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'ceph-cluster-advanced', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'hosts-overview', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'host-details', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'pool-overview', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'pool-detail', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'osds-overview', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'osd-device-details', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'rbd-overview', 'state': 'present'})",
                                "changed: [instance] => (item={'name': 'rbd-details', 'state': 'present'})",
                                "",
                                "TASK [vexxhost.atmosphere.kube_prometheus_stack : Deploy Helm chart] ***********",
                                "Saturday 04 April 2026  10:23:57 +0000 (0:00:15.632)       0:15:40.923 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:24:30 +0000 (0:00:33.357)       0:16:14.280 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:24:30 +0000 (0:00:00.069)       0:16:14.350 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.loki : Deploy Helm chart] ****************************",
                                "Saturday 04 April 2026  10:24:31 +0000 (0:00:00.584)       0:16:14.935 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:24:35 +0000 (0:00:04.252)       0:16:19.187 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:24:35 +0000 (0:00:00.053)       0:16:19.240 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.vector : Deploy Helm chart] **************************",
                                "Saturday 04 April 2026  10:24:35 +0000 (0:00:00.530)       0:16:19.770 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:24:38 +0000 (0:00:02.076)       0:16:21.846 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:24:38 +0000 (0:00:00.062)       0:16:21.908 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.goldpinger : Deploy Helm chart] **********************",
                                "Saturday 04 April 2026  10:24:38 +0000 (0:00:00.584)       0:16:22.493 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.ipmi_exporter : Deploy service] **********************",
                                "Saturday 04 April 2026  10:24:43 +0000 (0:00:04.443)       0:16:26.936 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:24:43 +0000 (0:00:00.888)       0:16:27.825 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:24:44 +0000 (0:00:00.056)       0:16:27.881 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.prometheus_pushgateway : Suspend the existing HelmRelease] ***",
                                "Saturday 04 April 2026  10:24:44 +0000 (0:00:00.482)       0:16:28.363 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.prometheus_pushgateway : Remove the existing HelmRelease] ***",
                                "Saturday 04 April 2026  10:24:45 +0000 (0:00:00.760)       0:16:29.124 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.prometheus_pushgateway : Deploy Helm chart] **********",
                                "Saturday 04 April 2026  10:24:46 +0000 (0:00:00.808)       0:16:29.933 ********",
                                "changed: [instance]",
                                "",
                                "PLAY [Deploy OpenStack] ********************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Saturday 04 April 2026  10:24:48 +0000 (0:00:02.224)       0:16:32.158 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Retrieve list of all the needed endpoints] ***",
                                "Saturday 04 April 2026  10:24:50 +0000 (0:00:01.856)       0:16:34.014 ********",
                                "ok: [instance]",
                                "",
                                "TASK [Create RabbitMQ cluster] *************************************************",
                                "Saturday 04 April 2026  10:24:50 +0000 (0:00:00.087)       0:16:34.102 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Grab RabbitMQ cluster secret] ***",
                                "Saturday 04 April 2026  10:24:50 +0000 (0:00:00.047)       0:16:34.149 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Cache fact with RabbitMQ cluster credentials] ***",
                                "Saturday 04 April 2026  10:24:50 +0000 (0:00:00.053)       0:16:34.202 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Grab Percona XtraDB cluster secret] ***",
                                "Saturday 04 April 2026  10:24:50 +0000 (0:00:00.048)       0:16:34.251 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Cache fact with Percona XtraDB password] ***",
                                "Saturday 04 April 2026  10:24:50 +0000 (0:00:00.045)       0:16:34.296 ********",
                                "skipping: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Reset value for OpenStack_Helm endpoints] ***",
                                "Saturday 04 April 2026  10:24:50 +0000 (0:00:00.048)       0:16:34.344 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Generate OpenStack-Helm endpoints] ***",
                                "Saturday 04 April 2026  10:24:50 +0000 (0:00:00.059)       0:16:34.404 ********",
                                "ok: [instance] => (item=oslo_cache)",
                                "",
                                "TASK [vexxhost.atmosphere.openstack_helm_endpoints : Clean-up facts] ***********",
                                "Saturday 04 April 2026  10:24:50 +0000 (0:00:00.064)       0:16:34.469 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Include help chart upload method tasks] ***",
                                "Saturday 04 April 2026  10:24:50 +0000 (0:00:00.080)       0:16:34.549 ********",
                                "included: /home/zuul/.ansible/collections/ansible_collections/vexxhost/kubernetes/roles/upload_helm_chart/tasks/synchronize.yml for instance",
                                "",
                                "TASK [vexxhost.kubernetes.upload_helm_chart : Upload Helm chart] ***************",
                                "Saturday 04 April 2026  10:24:50 +0000 (0:00:00.065)       0:16:34.615 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.memcached : Suspend the existing HelmRelease] ********",
                                "Saturday 04 April 2026  10:24:51 +0000 (0:00:00.534)       0:16:35.150 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.memcached : Remove the existing HelmRelease] *********",
                                "Saturday 04 April 2026  10:24:52 +0000 (0:00:01.256)       0:16:36.406 ********",
                                "ok: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.memcached : Deploy Helm chart] ***********************",
                                "Saturday 04 April 2026  10:24:54 +0000 (0:00:01.903)       0:16:38.310 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.memcached : Apply manifests for monitoring] **********",
                                "Saturday 04 April 2026  10:24:56 +0000 (0:00:01.965)       0:16:40.276 ********",
                                "changed: [instance]",
                                "",
                                "TASK [vexxhost.atmosphere.openstacksdk : Install openstacksdk] *****************",
                                "Saturday 04 April 2026  10:24:57 +0000 (0:00:00.842)       0:16:41.118 ********",
                                "fatal: [instance]: FAILED! => {\"changed\": false, \"cmd\": [\"/usr/bin/python3.10\", \"-m\", \"pip.__main__\", \"install\", \"openstacksdk==0.61.0\"], \"msg\": \"stdout: Collecting openstacksdk==0.61.0\\n  Downloading openstacksdk-0.61.0-py3-none-any.whl (1.4 MB)\\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501                             0.4/1.4 MB 9.3 MB/s eta 0:00:01\\n\\n:stderr: ERROR: THESE PACKAGES DO NOT MATCH THE HASHES FROM THE REQUIREMENTS FILE. If you have updated the package versions, please update the hashes. Otherwise, examine the package contents carefully; someone may have tampered with them.\\n    openstacksdk==0.61.0 from https://files.pythonhosted.org/packages/c2/31/a3501e87efc4ad83ab998502d681dde5fefbfe7d75225cabe268f815dad6/openstacksdk-0.61.0-py3-none-any.whl#sha256=9894d3d510563dcfc50c4755287dbfbf98def1f37caf2cfc15e9d0e1fd5d9a41:\\n        Expected sha256 9894d3d510563dcfc50c4755287dbfbf98def1f37caf2cfc15e9d0e1fd5d9a41\\n             Got        062b9b99adeb179395c6cf10283baad7165cce13a5f6f2aebc05254c78edf341\\n\\n\"}",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "instance                   : ok=346  changed=123  unreachable=0    failed=1    skipped=107  rescued=0    ignored=2",
                                "",
                                "Saturday 04 April 2026  10:24:58 +0000 (0:00:01.201)       0:16:42.320 ********",
                                "===============================================================================",
                                "vexxhost.ceph.mon : Run Bootstrap coomand ----------------------------- 138.66s",
                                "vexxhost.atmosphere.keycloak : Deploy Helm chart ---------------------- 123.48s",
                                "vexxhost.ceph.osd : Install OSDs -------------------------------------- 123.32s",
                                "vexxhost.atmosphere.percona_xtradb_cluster : Apply Percona XtraDB cluster -- 79.27s",
                                "vexxhost.kubernetes.envoy_gateway : Deploy Helm chart ------------------ 42.12s",
                                "vexxhost.atmosphere.kube_prometheus_stack : Deploy Helm chart ---------- 33.36s",
                                "vexxhost.kubernetes.cert_manager : Deploy Helm chart ------------------- 31.84s",
                                "vexxhost.kubernetes.kubernetes : Initialize cluster -------------------- 29.74s",
                                "vexxhost.atmosphere.ingress_nginx : Deploy Helm chart ------------------ 21.07s",
                                "vexxhost.atmosphere.kube_prometheus_stack : Generate client secret passwords -- 17.45s",
                                "vexxhost.atmosphere.kube_prometheus_stack : Deploy additional dashboards -- 15.63s",
                                "vexxhost.ceph.osd : Get `ceph-volume lvm list` status ------------------ 13.16s",
                                "vexxhost.atmosphere.kube_prometheus_stack : Generate OAuth2 proxy configuration -- 11.85s",
                                "vexxhost.atmosphere.kube_prometheus_stack : Generate cookie secrets ---- 11.62s",
                                "vexxhost.ceph.mon : Validate monitor exist ----------------------------- 11.17s",
                                "vexxhost.atmosphere.kube_prometheus_stack : Install all CRDs ------------ 7.04s",
                                "vexxhost.atmosphere.sysctl : Configure sysctl values -------------------- 5.92s",
                                "vexxhost.containers.containerd : Install AppArmor packages -------------- 5.47s",
                                "vexxhost.ceph.osd : Ensure all OSDs are non-legacy ---------------------- 5.37s",
                                "vexxhost.ceph.osd : Get `cephadm ls` status ----------------------------- 5.37s",
                                "CRITICAL Ansible return code was 2, command was: ansible-playbook --inventory /home/zuul/.ansible/tmp/molecule.v9Wo.aio/inventory --skip-tags molecule-notest,notest --inventory=/home/zuul/src/github.com/vexxhost/atmosphere/inventory.yaml /home/zuul/src/github.com/vexxhost/atmosphere/molecule/aio/converge.yml",
                                "ERROR    [aio > converge] \u001b[31mExecuted: Failed\u001b[0m",
                                "ERROR    Ansible return code was 2, command was: ansible-playbook --inventory /home/zuul/.ansible/tmp/molecule.v9Wo.aio/inventory --skip-tags molecule-notest,notest --inventory=/home/zuul/src/github.com/vexxhost/atmosphere/inventory.yaml /home/zuul/src/github.com/vexxhost/atmosphere/molecule/aio/converge.yml"
                            ],
                            "zuul_log_id": "0242ac17-0011-9b67-3b0e-000000000006-1-instance"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:24:59.386986Z",
                            "start": "2026-04-04T10:06:38.729206Z"
                        },
                        "id": "0242ac17-0011-9b67-3b0e-000000000006",
                        "name": "Run Molecule scenario"
                    }
                }
            ]
        }
    ],
    "stats": {
        "instance": {
            "changed": 2,
            "failures": 1,
            "ignored": 0,
            "ok": 2,
            "rescued": 0,
            "skipped": 0,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "stable/zed",
    "index": "0",
    "phase": "post",
    "playbook": "github.com/vexxhost/atmosphere/test-playbooks/molecule/post.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-04-04T10:26:49.592210Z",
                    "start": "2026-04-04T10:25:01.135489Z"
                },
                "id": "0242ac17-0011-cc15-c3f6-000000000002",
                "name": "all"
            },
            "tasks": [
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/tmp/logs/system",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/tmp/logs/system",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": null,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/tmp/logs/system",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/tmp/logs/system",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-00000000000a",
                        "name": "gather-host-logs",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-host-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:25:01.535920Z",
                            "start": "2026-04-04T10:25:01.154203Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-00000000000c",
                        "name": "creating directory for system status"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "shell",
                            "changed": true,
                            "cmd": "set -x\nsystemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt\nip addr > /tmp/logs/system/ip-addr.txt\nip route > /tmp/logs/system/ip-route.txt\nlsblk > /tmp/logs/system/lsblk.txt\nmount > /tmp/logs/system/mount.txt\ndocker images > /tmp/logs/system/docker-images.txt\nbrctl show > /tmp/logs/system/brctl-show.txt\nps aux --sort=-%mem > /tmp/logs/system/ps.txt\ndpkg -l > /tmp/logs/system/packages.txt\nCONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul))\nif [ ! -z \"$CONTAINERS\" ]; then\n  mkdir -p \"/tmp/logs/system/containers\"\n  for CONTAINER in ${CONTAINERS}; do\n    docker logs \"${CONTAINER}\" > \"/tmp/logs/system/containers/${CONTAINER}.txt\"\n  done\nfi",
                            "delta": "0:00:00.142325",
                            "end": "2026-04-04 10:25:02.073326",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -x\nsystemd-cgls --full --all --no-pager > /tmp/logs/system/systemd-cgls.txt\nip addr > /tmp/logs/system/ip-addr.txt\nip route > /tmp/logs/system/ip-route.txt\nlsblk > /tmp/logs/system/lsblk.txt\nmount > /tmp/logs/system/mount.txt\ndocker images > /tmp/logs/system/docker-images.txt\nbrctl show > /tmp/logs/system/brctl-show.txt\nps aux --sort=-%mem > /tmp/logs/system/ps.txt\ndpkg -l > /tmp/logs/system/packages.txt\nCONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul))\nif [ ! -z \"$CONTAINERS\" ]; then\n  mkdir -p \"/tmp/logs/system/containers\"\n  for CONTAINER in ${CONTAINERS}; do\n    docker logs \"${CONTAINER}\" > \"/tmp/logs/system/containers/${CONTAINER}.txt\"\n  done\nfi",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-cc15-c3f6-00000000000d-1-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-04-04 10:25:01.931001",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ systemd-cgls --full --all --no-pager\n+ ip addr\n+ ip route\n+ lsblk\n+ mount\n+ docker images\n+ brctl show\n/bin/bash: line 8: brctl: command not found\n+ ps aux --sort=-%mem\n+ dpkg -l\n+ CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul))\n++ docker ps -a --format '{{ .Names }}' --filter label=zuul\n+ '[' '!' -z '' ']'",
                            "stdout_lines": [
                                "+ systemd-cgls --full --all --no-pager",
                                "+ ip addr",
                                "+ ip route",
                                "+ lsblk",
                                "+ mount",
                                "+ docker images",
                                "+ brctl show",
                                "/bin/bash: line 8: brctl: command not found",
                                "+ ps aux --sort=-%mem",
                                "+ dpkg -l",
                                "+ CONTAINERS=($(docker ps -a --format '{{ .Names }}' --filter label=zuul))",
                                "++ docker ps -a --format '{{ .Names }}' --filter label=zuul",
                                "+ '[' '!' -z '' ']'"
                            ],
                            "zuul_log_id": "0242ac17-0011-cc15-c3f6-00000000000d-1-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-00000000000a",
                        "name": "gather-host-logs",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-host-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:25:02.615329Z",
                            "start": "2026-04-04T10:25:01.565680Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-00000000000d",
                        "name": "Get logs for each host"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "synchronize",
                            "changed": true,
                            "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --rsh='/usr/bin/ssh -S none -o Port=22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null' --rsync-path='sudo -u root rsync' --out-format='<<CHANGED>>%i %n%L' zuul@199.204.45.153:/tmp/logs/system /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                            "invocation": {
                                "module_args": {
                                    "_local_rsync_password": null,
                                    "_local_rsync_path": "rsync",
                                    "_ssh_args": null,
                                    "_substitute_controller": false,
                                    "archive": true,
                                    "checksum": false,
                                    "compress": true,
                                    "copy_links": false,
                                    "delay_updates": true,
                                    "delete": false,
                                    "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                                    "dest_port": 22,
                                    "dirs": false,
                                    "existing_only": false,
                                    "group": null,
                                    "link_dest": null,
                                    "links": null,
                                    "mode": "pull",
                                    "owner": null,
                                    "partial": false,
                                    "perms": null,
                                    "private_key": null,
                                    "recursive": null,
                                    "rsync_opts": [],
                                    "rsync_path": "sudo -u root rsync",
                                    "rsync_timeout": 0,
                                    "set_remote_user": true,
                                    "src": "zuul@199.204.45.153:/tmp/logs/system",
                                    "ssh_connection_multiplexing": false,
                                    "times": null,
                                    "use_ssh_args": false,
                                    "verify_host": false
                                }
                            },
                            "msg": "created directory /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance\ncd+++++++++ system/\n>f+++++++++ system/brctl-show.txt\n>f+++++++++ system/docker-images.txt\n>f+++++++++ system/ip-addr.txt\n>f+++++++++ system/ip-route.txt\n>f+++++++++ system/lsblk.txt\n>f+++++++++ system/mount.txt\n>f+++++++++ system/packages.txt\n>f+++++++++ system/ps.txt\n>f+++++++++ system/systemd-cgls.txt\n",
                            "rc": 0,
                            "stdout_lines": [
                                "created directory /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                                "cd+++++++++ system/",
                                ">f+++++++++ system/brctl-show.txt",
                                ">f+++++++++ system/docker-images.txt",
                                ">f+++++++++ system/ip-addr.txt",
                                ">f+++++++++ system/ip-route.txt",
                                ">f+++++++++ system/lsblk.txt",
                                ">f+++++++++ system/mount.txt",
                                ">f+++++++++ system/packages.txt",
                                ">f+++++++++ system/ps.txt",
                                ">f+++++++++ system/systemd-cgls.txt"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-00000000000a",
                        "name": "gather-host-logs",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-host-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:25:03.273817Z",
                            "start": "2026-04-04T10:25:02.621321Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-00000000000e",
                        "name": "Downloads logs to executor"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "directory",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/tmp/logs/helm/values",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/tmp/logs/helm/values",
                                            "state": "absent"
                                        }
                                    },
                                    "directory": "values",
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": null,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/tmp/logs/helm/values",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/tmp/logs/helm/values",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0
                                },
                                {
                                    "ansible_loop_var": "directory",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/tmp/logs/helm/releases",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/tmp/logs/helm/releases",
                                            "state": "absent"
                                        }
                                    },
                                    "directory": "releases",
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": null,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/tmp/logs/helm/releases",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/tmp/logs/helm/releases",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000010",
                        "name": "helm-release-status",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/helm-release-status"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:25:03.837665Z",
                            "start": "2026-04-04T10:25:03.283511Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000012",
                        "name": "creating directory for helm release status"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "shell",
                            "changed": true,
                            "cmd": "set -e\n\nfor namespace in $(kubectl get namespaces --no-headers --output custom-columns=\":metadata.name\"); do\n      # get all Helm releases including pending and failed releases\n      for release in $(helm list --all --short --namespace $namespace); do\n              # Make respective directories only when a Helm release actually exists in the namespace\n              # to prevent uploading a bunch of empty directories for namespaces without a Helm release.\n              mkdir -p /tmp/logs/helm/releases/$namespace\n              mkdir -p /tmp/logs/helm/values/$namespace\n\n              helm status $release --namespace $namespace >> /tmp/logs/helm/releases/$namespace/$release.txt\n              helm get values $release --namespace $namespace --all >> /tmp/logs/helm/values/$namespace/$release.yaml\n      done\ndone",
                            "delta": "0:00:04.535874",
                            "end": "2026-04-04 10:25:08.610344",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -e\n\nfor namespace in $(kubectl get namespaces --no-headers --output custom-columns=\":metadata.name\"); do\n      # get all Helm releases including pending and failed releases\n      for release in $(helm list --all --short --namespace $namespace); do\n              # Make respective directories only when a Helm release actually exists in the namespace\n              # to prevent uploading a bunch of empty directories for namespaces without a Helm release.\n              mkdir -p /tmp/logs/helm/releases/$namespace\n              mkdir -p /tmp/logs/helm/values/$namespace\n\n              helm status $release --namespace $namespace >> /tmp/logs/helm/releases/$namespace/$release.txt\n              helm get values $release --namespace $namespace --all >> /tmp/logs/helm/values/$namespace/$release.yaml\n      done\ndone",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-cc15-c3f6-000000000014-1-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-04-04 10:25:04.074470",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac17-0011-cc15-c3f6-000000000014-1-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000010",
                        "name": "helm-release-status",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/helm-release-status"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:25:08.894710Z",
                            "start": "2026-04-04T10:25:03.870149Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000014",
                        "name": "Gather get release status for helm charts"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "synchronize",
                            "changed": true,
                            "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --rsh='/usr/bin/ssh -S none -o Port=22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null' --rsync-path='sudo -u root rsync' --out-format='<<CHANGED>>%i %n%L' zuul@199.204.45.153:/tmp/logs/helm /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                            "invocation": {
                                "module_args": {
                                    "_local_rsync_password": null,
                                    "_local_rsync_path": "rsync",
                                    "_ssh_args": null,
                                    "_substitute_controller": false,
                                    "archive": true,
                                    "checksum": false,
                                    "compress": true,
                                    "copy_links": false,
                                    "delay_updates": true,
                                    "delete": false,
                                    "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                                    "dest_port": 22,
                                    "dirs": false,
                                    "existing_only": false,
                                    "group": null,
                                    "link_dest": null,
                                    "links": null,
                                    "mode": "pull",
                                    "owner": null,
                                    "partial": false,
                                    "perms": null,
                                    "private_key": null,
                                    "recursive": null,
                                    "rsync_opts": [],
                                    "rsync_path": "sudo -u root rsync",
                                    "rsync_timeout": 0,
                                    "set_remote_user": true,
                                    "src": "zuul@199.204.45.153:/tmp/logs/helm",
                                    "ssh_connection_multiplexing": false,
                                    "times": null,
                                    "use_ssh_args": false,
                                    "verify_host": false
                                }
                            },
                            "msg": "cd+++++++++ helm/\ncd+++++++++ helm/releases/\ncd+++++++++ helm/releases/auth-system/\n>f+++++++++ helm/releases/auth-system/keycloak.txt\ncd+++++++++ helm/releases/cert-manager/\n>f+++++++++ helm/releases/cert-manager/cert-manager.txt\ncd+++++++++ helm/releases/envoy-gateway-system/\n>f+++++++++ helm/releases/envoy-gateway-system/envoy-gateway.txt\ncd+++++++++ helm/releases/ingress-nginx/\n>f+++++++++ helm/releases/ingress-nginx/ingress-nginx.txt\ncd+++++++++ helm/releases/kube-system/\n>f+++++++++ helm/releases/kube-system/cilium.txt\ncd+++++++++ helm/releases/local-path-storage/\n>f+++++++++ helm/releases/local-path-storage/local-path-provisioner.txt\ncd+++++++++ helm/releases/monitoring/\n>f+++++++++ helm/releases/monitoring/goldpinger.txt\n>f+++++++++ helm/releases/monitoring/kube-prometheus-stack.txt\n>f+++++++++ helm/releases/monitoring/loki.txt\n>f+++++++++ helm/releases/monitoring/node-feature-discovery.txt\n>f+++++++++ helm/releases/monitoring/prometheus-pushgateway.txt\n>f+++++++++ helm/releases/monitoring/vector.txt\ncd+++++++++ helm/releases/openstack/\n>f+++++++++ helm/releases/openstack/memcached.txt\n>f+++++++++ helm/releases/openstack/pxc-operator.txt\n>f+++++++++ helm/releases/openstack/rabbitmq-cluster-operator.txt\ncd+++++++++ helm/values/\ncd+++++++++ helm/values/auth-system/\n>f+++++++++ helm/values/auth-system/keycloak.yaml\ncd+++++++++ helm/values/cert-manager/\n>f+++++++++ helm/values/cert-manager/cert-manager.yaml\ncd+++++++++ helm/values/envoy-gateway-system/\n>f+++++++++ helm/values/envoy-gateway-system/envoy-gateway.yaml\ncd+++++++++ helm/values/ingress-nginx/\n>f+++++++++ helm/values/ingress-nginx/ingress-nginx.yaml\ncd+++++++++ helm/values/kube-system/\n>f+++++++++ helm/values/kube-system/cilium.yaml\ncd+++++++++ helm/values/local-path-storage/\n>f+++++++++ helm/values/local-path-storage/local-path-provisioner.yaml\ncd+++++++++ helm/values/monitoring/\n>f+++++++++ helm/values/monitoring/goldpinger.yaml\n>f+++++++++ helm/values/monitoring/kube-prometheus-stack.yaml\n>f+++++++++ helm/values/monitoring/loki.yaml\n>f+++++++++ helm/values/monitoring/node-feature-discovery.yaml\n>f+++++++++ helm/values/monitoring/prometheus-pushgateway.yaml\n>f+++++++++ helm/values/monitoring/vector.yaml\ncd+++++++++ helm/values/openstack/\n>f+++++++++ helm/values/openstack/memcached.yaml\n>f+++++++++ helm/values/openstack/pxc-operator.yaml\n>f+++++++++ helm/values/openstack/rabbitmq-cluster-operator.yaml\n",
                            "rc": 0,
                            "stdout_lines": [
                                "cd+++++++++ helm/",
                                "cd+++++++++ helm/releases/",
                                "cd+++++++++ helm/releases/auth-system/",
                                ">f+++++++++ helm/releases/auth-system/keycloak.txt",
                                "cd+++++++++ helm/releases/cert-manager/",
                                ">f+++++++++ helm/releases/cert-manager/cert-manager.txt",
                                "cd+++++++++ helm/releases/envoy-gateway-system/",
                                ">f+++++++++ helm/releases/envoy-gateway-system/envoy-gateway.txt",
                                "cd+++++++++ helm/releases/ingress-nginx/",
                                ">f+++++++++ helm/releases/ingress-nginx/ingress-nginx.txt",
                                "cd+++++++++ helm/releases/kube-system/",
                                ">f+++++++++ helm/releases/kube-system/cilium.txt",
                                "cd+++++++++ helm/releases/local-path-storage/",
                                ">f+++++++++ helm/releases/local-path-storage/local-path-provisioner.txt",
                                "cd+++++++++ helm/releases/monitoring/",
                                ">f+++++++++ helm/releases/monitoring/goldpinger.txt",
                                ">f+++++++++ helm/releases/monitoring/kube-prometheus-stack.txt",
                                ">f+++++++++ helm/releases/monitoring/loki.txt",
                                ">f+++++++++ helm/releases/monitoring/node-feature-discovery.txt",
                                ">f+++++++++ helm/releases/monitoring/prometheus-pushgateway.txt",
                                ">f+++++++++ helm/releases/monitoring/vector.txt",
                                "cd+++++++++ helm/releases/openstack/",
                                ">f+++++++++ helm/releases/openstack/memcached.txt",
                                ">f+++++++++ helm/releases/openstack/pxc-operator.txt",
                                ">f+++++++++ helm/releases/openstack/rabbitmq-cluster-operator.txt",
                                "cd+++++++++ helm/values/",
                                "cd+++++++++ helm/values/auth-system/",
                                ">f+++++++++ helm/values/auth-system/keycloak.yaml",
                                "cd+++++++++ helm/values/cert-manager/",
                                ">f+++++++++ helm/values/cert-manager/cert-manager.yaml",
                                "cd+++++++++ helm/values/envoy-gateway-system/",
                                ">f+++++++++ helm/values/envoy-gateway-system/envoy-gateway.yaml",
                                "cd+++++++++ helm/values/ingress-nginx/",
                                ">f+++++++++ helm/values/ingress-nginx/ingress-nginx.yaml",
                                "cd+++++++++ helm/values/kube-system/",
                                ">f+++++++++ helm/values/kube-system/cilium.yaml",
                                "cd+++++++++ helm/values/local-path-storage/",
                                ">f+++++++++ helm/values/local-path-storage/local-path-provisioner.yaml",
                                "cd+++++++++ helm/values/monitoring/",
                                ">f+++++++++ helm/values/monitoring/goldpinger.yaml",
                                ">f+++++++++ helm/values/monitoring/kube-prometheus-stack.yaml",
                                ">f+++++++++ helm/values/monitoring/loki.yaml",
                                ">f+++++++++ helm/values/monitoring/node-feature-discovery.yaml",
                                ">f+++++++++ helm/values/monitoring/prometheus-pushgateway.yaml",
                                ">f+++++++++ helm/values/monitoring/vector.yaml",
                                "cd+++++++++ helm/values/openstack/",
                                ">f+++++++++ helm/values/openstack/memcached.yaml",
                                ">f+++++++++ helm/values/openstack/pxc-operator.yaml",
                                ">f+++++++++ helm/values/openstack/rabbitmq-cluster-operator.yaml"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000010",
                        "name": "helm-release-status",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/helm-release-status"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:25:09.421829Z",
                            "start": "2026-04-04T10:25:08.901399Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000015",
                        "name": "Downloads logs to executor"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/tmp/logs/objects/cluster",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/tmp/logs/objects/cluster",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": null,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/tmp/logs/objects/cluster",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/tmp/logs/objects/cluster",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000017",
                        "name": "describe-kubernetes-objects",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/describe-kubernetes-objects"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:25:09.723588Z",
                            "start": "2026-04-04T10:25:09.485871Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000019",
                        "name": "creating directory for cluster scoped objects"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "shell",
                            "changed": true,
                            "cmd": "set -e\nexport OBJECT_TYPE=node,clusterrole,clusterrolebinding,storageclass,namespace\nexport PARALLELISM_FACTOR=2\n\nfunction list_objects () {\n  printf ${OBJECT_TYPE} | xargs -d ',' -I {} -P1 -n1 bash -c 'echo \"$@\"' _ {}\n}\nexport -f list_objects\n\nfunction name_objects () {\n  export OBJECT=$1\n  kubectl get ${OBJECT} -o name | xargs -L1 -I {} -P1 -n1 bash -c 'echo \"${OBJECT} ${1#*/}\"' _ {}\n}\nexport -f name_objects\n\nfunction get_objects () {\n  input=($1)\n  export OBJECT=${input[0]}\n  export NAME=${input[1]#*/}\n  echo \"${OBJECT}/${NAME}\"\n  DIR=\"/tmp/logs/objects/cluster/${OBJECT}\"\n  mkdir -p ${DIR}\n  kubectl get ${OBJECT} ${NAME} -o yaml > \"${DIR}/${NAME}.yaml\"\n  kubectl describe ${OBJECT} ${NAME} > \"${DIR}/${NAME}.txt\"\n}\nexport -f get_objects\n\nlist_objects |  xargs -r -n 1 -P ${PARALLELISM_FACTOR} -I {} bash -c 'name_objects \"$@\"' _ {} |  xargs -r -n 1 -P ${PARALLELISM_FACTOR} -I {} bash -c 'get_objects \"$@\"' _ {}",
                            "delta": "0:00:17.903279",
                            "end": "2026-04-04 10:25:27.857978",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -e\nexport OBJECT_TYPE=node,clusterrole,clusterrolebinding,storageclass,namespace\nexport PARALLELISM_FACTOR=2\n\nfunction list_objects () {\n  printf ${OBJECT_TYPE} | xargs -d ',' -I {} -P1 -n1 bash -c 'echo \"$@\"' _ {}\n}\nexport -f list_objects\n\nfunction name_objects () {\n  export OBJECT=$1\n  kubectl get ${OBJECT} -o name | xargs -L1 -I {} -P1 -n1 bash -c 'echo \"${OBJECT} ${1#*/}\"' _ {}\n}\nexport -f name_objects\n\nfunction get_objects () {\n  input=($1)\n  export OBJECT=${input[0]}\n  export NAME=${input[1]#*/}\n  echo \"${OBJECT}/${NAME}\"\n  DIR=\"/tmp/logs/objects/cluster/${OBJECT}\"\n  mkdir -p ${DIR}\n  kubectl get ${OBJECT} ${NAME} -o yaml > \"${DIR}/${NAME}.yaml\"\n  kubectl describe ${OBJECT} ${NAME} > \"${DIR}/${NAME}.txt\"\n}\nexport -f get_objects\n\nlist_objects |  xargs -r -n 1 -P ${PARALLELISM_FACTOR} -I {} bash -c 'name_objects \"$@\"' _ {} |  xargs -r -n 1 -P ${PARALLELISM_FACTOR} -I {} bash -c 'get_objects \"$@\"' _ {}",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-cc15-c3f6-00000000001a-1-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-04-04 10:25:09.954699",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value\nxargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nnode/instance\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nclusterrole/admin\nclusterrole/cert-manager-cainjector\nclusterrole/cert-manager-controller-approve:cert-manager-io\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nclusterrole/cert-manager-controller-certificates\nclusterrole/cert-manager-controller-certificatesigningrequests\nclusterrole/cert-manager-controller-challenges\nclusterrole/cert-manager-controller-clusterissuers\nclusterrole/cert-manager-controller-ingress-shim\nclusterrole/cert-manager-controller-issuers\nclusterrole/cert-manager-controller-orders\nclusterrole/cert-manager-edit\nclusterrole/cert-manager-view\nclusterrole/cert-manager-webhook:subjectaccessreviews\nclusterrole/cilium\nclusterrole/cilium-operator\nclusterrole/cluster-admin\nclusterrole/edit\nclusterrole/envoy-gateway-gateway-helm-certgen:envoy-gateway-system\nclusterrole/envoy-gateway-gateway-helm-envoy-gateway-role\nclusterrole/goldpinger-clusterrole\nclusterrole/ingress-nginx\nclusterrole/kube-prometheus-stack-grafana-clusterrole\nclusterrole/kube-prometheus-stack-kube-state-metrics\nclusterrole/kube-prometheus-stack-operator\nclusterrole/kube-prometheus-stack-prometheus\nclusterrole/kubeadm:get-nodes\nclusterrole/local-path-provisioner\nclusterrolebinding/cert-manager-cainjector\nclusterrole/loki-clusterrole\nclusterrolebinding/cert-manager-controller-approve:cert-manager-io\nclusterrole/node-feature-discovery\nclusterrolebinding/cert-manager-controller-certificates\nclusterrole/rabbitmq-cluster-operator\nclusterrole/rabbitmq-messaging-topology-operator\nclusterrolebinding/cert-manager-controller-certificatesigningrequests\nclusterrolebinding/cert-manager-controller-challenges\nclusterrole/secretgen-controller-cluster-role\nclusterrolebinding/cert-manager-controller-clusterissuers\nclusterrole/system:aggregate-to-admin\nclusterrolebinding/cert-manager-controller-ingress-shim\nclusterrole/system:aggregate-to-edit\nclusterrolebinding/cert-manager-controller-issuers\nclusterrole/system:aggregate-to-view\nclusterrolebinding/cert-manager-controller-orders\nclusterrole/system:auth-delegator\nclusterrolebinding/cert-manager-webhook:subjectaccessreviews\nclusterrole/system:basic-user\nclusterrolebinding/cilium\nclusterrole/system:certificates.k8s.io:certificatesigningrequests:nodeclient\nclusterrole/system:certificates.k8s.io:certificatesigningrequests:selfnodeclient\nclusterrolebinding/cilium-operator\nclusterrole/system:certificates.k8s.io:kube-apiserver-client-approver\nclusterrolebinding/cluster-admin\nclusterrole/system:certificates.k8s.io:kube-apiserver-client-kubelet-approver\nclusterrolebinding/envoy-gateway-gateway-helm-certgen:envoy-gateway-system\nclusterrole/system:certificates.k8s.io:kubelet-serving-approver\nclusterrolebinding/envoy-gateway-gateway-helm-envoy-gateway-rolebinding\nclusterrolebinding/goldpinger-clusterrolebinding\nclusterrole/system:certificates.k8s.io:legacy-unknown-approver\nclusterrolebinding/ingress-nginx\nclusterrolebinding/kube-prometheus-stack-grafana-clusterrolebinding\nclusterrolebinding/kube-prometheus-stack-kube-state-metrics\nclusterrole/system:controller:attachdetach-controller\nclusterrolebinding/kube-prometheus-stack-operator\nclusterrole/system:controller:certificate-controller\nclusterrolebinding/kube-prometheus-stack-prometheus\nclusterrole/system:controller:clusterrole-aggregation-controller\nclusterrolebinding/kubeadm:get-nodes\nclusterrole/system:controller:cronjob-controller\nclusterrolebinding/kubeadm:kubelet-bootstrap\nclusterrole/system:controller:daemon-set-controller\nclusterrolebinding/kubeadm:node-autoapprove-bootstrap\nclusterrole/system:controller:deployment-controller\nclusterrolebinding/kubeadm:node-autoapprove-certificate-rotation\nclusterrole/system:controller:disruption-controller\nclusterrolebinding/kubeadm:node-proxier\nclusterrole/system:controller:endpoint-controller\nclusterrolebinding/local-path-provisioner\nclusterrolebinding/loki-clusterrolebinding\nclusterrole/system:controller:endpointslice-controller\nclusterrole/system:controller:endpointslicemirroring-controller\nclusterrolebinding/node-feature-discovery\nclusterrole/system:controller:ephemeral-volume-controller\nclusterrolebinding/rabbitmq-cluster-operator\nclusterrolebinding/rabbitmq-messaging-topology-operator\nclusterrole/system:controller:expand-controller\nclusterrolebinding/secretgen-controller-cluster-role-binding\nclusterrole/system:controller:generic-garbage-collector\nclusterrole/system:controller:horizontal-pod-autoscaler\nclusterrole/system:controller:job-controller\nclusterrole/system:controller:namespace-controller\nclusterrolebinding/system:basic-user\nclusterrole/system:controller:node-controller\nclusterrolebinding/system:controller:attachdetach-controller\nclusterrolebinding/system:controller:certificate-controller\nclusterrole/system:controller:persistent-volume-binder\nclusterrolebinding/system:controller:clusterrole-aggregation-controller\nclusterrole/system:controller:pod-garbage-collector\nclusterrolebinding/system:controller:cronjob-controller\nclusterrole/system:controller:pv-protection-controller\nclusterrolebinding/system:controller:daemon-set-controller\nclusterrole/system:controller:pvc-protection-controller\nclusterrolebinding/system:controller:deployment-controller\nclusterrole/system:controller:replicaset-controller\nclusterrolebinding/system:controller:disruption-controller\nclusterrolebinding/system:controller:endpoint-controller\nclusterrole/system:controller:replication-controller\nclusterrolebinding/system:controller:endpointslice-controller\nclusterrole/system:controller:resourcequota-controller\nclusterrolebinding/system:controller:endpointslicemirroring-controller\nclusterrole/system:controller:root-ca-cert-publisher\nclusterrolebinding/system:controller:ephemeral-volume-controller\nclusterrole/system:controller:route-controller\nclusterrolebinding/system:controller:expand-controller\nclusterrole/system:controller:service-account-controller\nclusterrolebinding/system:controller:generic-garbage-collector\nclusterrole/system:controller:service-controller\nclusterrolebinding/system:controller:horizontal-pod-autoscaler\nclusterrole/system:controller:statefulset-controller\nclusterrolebinding/system:controller:job-controller\nclusterrole/system:controller:ttl-after-finished-controller\nclusterrolebinding/system:controller:namespace-controller\nclusterrolebinding/system:controller:node-controller\nclusterrole/system:controller:ttl-controller\nclusterrolebinding/system:controller:persistent-volume-binder\nclusterrole/system:coredns\nclusterrole/system:discovery\nclusterrole/system:heapster\nclusterrolebinding/system:controller:pod-garbage-collector\nclusterrole/system:kube-aggregator\nclusterrole/system:kube-controller-manager\nclusterrolebinding/system:controller:pv-protection-controller\nclusterrole/system:kube-dns\nclusterrolebinding/system:controller:pvc-protection-controller\nclusterrole/system:kube-scheduler\nclusterrole/system:kubelet-api-admin\nclusterrolebinding/system:controller:replicaset-controller\nclusterrole/system:monitoring\nclusterrolebinding/system:controller:replication-controller\nclusterrole/system:node\nclusterrolebinding/system:controller:resourcequota-controller\nclusterrole/system:node-bootstrapper\nclusterrolebinding/system:controller:root-ca-cert-publisher\nclusterrole/system:node-problem-detector\nclusterrolebinding/system:controller:route-controller\nclusterrole/system:node-proxier\nclusterrolebinding/system:controller:service-account-controller\nclusterrole/system:persistent-volume-provisioner\nclusterrole/system:public-info-viewer\nclusterrolebinding/system:controller:service-controller\nclusterrole/system:service-account-issuer-discovery\nclusterrolebinding/system:controller:statefulset-controller\nclusterrole/system:volume-scheduler\nclusterrolebinding/system:controller:ttl-after-finished-controller\nclusterrole/vector\nclusterrole/view\nclusterrolebinding/system:controller:ttl-controller\nclusterrolebinding/system:coredns\nclusterrolebinding/system:discovery\nclusterrolebinding/system:kube-controller-manager\nclusterrolebinding/system:kube-dns\nclusterrolebinding/system:kube-scheduler\nclusterrolebinding/system:monitoring\nclusterrolebinding/system:node\nclusterrolebinding/system:node-proxier\nclusterrolebinding/system:public-info-viewer\nclusterrolebinding/system:service-account-issuer-discovery\nclusterrolebinding/system:volume-scheduler\nclusterrolebinding/vector\nstorageclass/general\nnamespace/auth-system\nnamespace/cert-manager\nnamespace/default\nnamespace/envoy-gateway-system\nnamespace/ingress-nginx\nnamespace/kube-node-lease\nnamespace/kube-public\nnamespace/kube-system\nnamespace/local-path-storage\nnamespace/monitoring\nnamespace/openstack\nnamespace/secretgen-controller",
                            "stdout_lines": [
                                "xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value",
                                "xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "node/instance",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "clusterrole/admin",
                                "clusterrole/cert-manager-cainjector",
                                "clusterrole/cert-manager-controller-approve:cert-manager-io",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "clusterrole/cert-manager-controller-certificates",
                                "clusterrole/cert-manager-controller-certificatesigningrequests",
                                "clusterrole/cert-manager-controller-challenges",
                                "clusterrole/cert-manager-controller-clusterissuers",
                                "clusterrole/cert-manager-controller-ingress-shim",
                                "clusterrole/cert-manager-controller-issuers",
                                "clusterrole/cert-manager-controller-orders",
                                "clusterrole/cert-manager-edit",
                                "clusterrole/cert-manager-view",
                                "clusterrole/cert-manager-webhook:subjectaccessreviews",
                                "clusterrole/cilium",
                                "clusterrole/cilium-operator",
                                "clusterrole/cluster-admin",
                                "clusterrole/edit",
                                "clusterrole/envoy-gateway-gateway-helm-certgen:envoy-gateway-system",
                                "clusterrole/envoy-gateway-gateway-helm-envoy-gateway-role",
                                "clusterrole/goldpinger-clusterrole",
                                "clusterrole/ingress-nginx",
                                "clusterrole/kube-prometheus-stack-grafana-clusterrole",
                                "clusterrole/kube-prometheus-stack-kube-state-metrics",
                                "clusterrole/kube-prometheus-stack-operator",
                                "clusterrole/kube-prometheus-stack-prometheus",
                                "clusterrole/kubeadm:get-nodes",
                                "clusterrole/local-path-provisioner",
                                "clusterrolebinding/cert-manager-cainjector",
                                "clusterrole/loki-clusterrole",
                                "clusterrolebinding/cert-manager-controller-approve:cert-manager-io",
                                "clusterrole/node-feature-discovery",
                                "clusterrolebinding/cert-manager-controller-certificates",
                                "clusterrole/rabbitmq-cluster-operator",
                                "clusterrole/rabbitmq-messaging-topology-operator",
                                "clusterrolebinding/cert-manager-controller-certificatesigningrequests",
                                "clusterrolebinding/cert-manager-controller-challenges",
                                "clusterrole/secretgen-controller-cluster-role",
                                "clusterrolebinding/cert-manager-controller-clusterissuers",
                                "clusterrole/system:aggregate-to-admin",
                                "clusterrolebinding/cert-manager-controller-ingress-shim",
                                "clusterrole/system:aggregate-to-edit",
                                "clusterrolebinding/cert-manager-controller-issuers",
                                "clusterrole/system:aggregate-to-view",
                                "clusterrolebinding/cert-manager-controller-orders",
                                "clusterrole/system:auth-delegator",
                                "clusterrolebinding/cert-manager-webhook:subjectaccessreviews",
                                "clusterrole/system:basic-user",
                                "clusterrolebinding/cilium",
                                "clusterrole/system:certificates.k8s.io:certificatesigningrequests:nodeclient",
                                "clusterrole/system:certificates.k8s.io:certificatesigningrequests:selfnodeclient",
                                "clusterrolebinding/cilium-operator",
                                "clusterrole/system:certificates.k8s.io:kube-apiserver-client-approver",
                                "clusterrolebinding/cluster-admin",
                                "clusterrole/system:certificates.k8s.io:kube-apiserver-client-kubelet-approver",
                                "clusterrolebinding/envoy-gateway-gateway-helm-certgen:envoy-gateway-system",
                                "clusterrole/system:certificates.k8s.io:kubelet-serving-approver",
                                "clusterrolebinding/envoy-gateway-gateway-helm-envoy-gateway-rolebinding",
                                "clusterrolebinding/goldpinger-clusterrolebinding",
                                "clusterrole/system:certificates.k8s.io:legacy-unknown-approver",
                                "clusterrolebinding/ingress-nginx",
                                "clusterrolebinding/kube-prometheus-stack-grafana-clusterrolebinding",
                                "clusterrolebinding/kube-prometheus-stack-kube-state-metrics",
                                "clusterrole/system:controller:attachdetach-controller",
                                "clusterrolebinding/kube-prometheus-stack-operator",
                                "clusterrole/system:controller:certificate-controller",
                                "clusterrolebinding/kube-prometheus-stack-prometheus",
                                "clusterrole/system:controller:clusterrole-aggregation-controller",
                                "clusterrolebinding/kubeadm:get-nodes",
                                "clusterrole/system:controller:cronjob-controller",
                                "clusterrolebinding/kubeadm:kubelet-bootstrap",
                                "clusterrole/system:controller:daemon-set-controller",
                                "clusterrolebinding/kubeadm:node-autoapprove-bootstrap",
                                "clusterrole/system:controller:deployment-controller",
                                "clusterrolebinding/kubeadm:node-autoapprove-certificate-rotation",
                                "clusterrole/system:controller:disruption-controller",
                                "clusterrolebinding/kubeadm:node-proxier",
                                "clusterrole/system:controller:endpoint-controller",
                                "clusterrolebinding/local-path-provisioner",
                                "clusterrolebinding/loki-clusterrolebinding",
                                "clusterrole/system:controller:endpointslice-controller",
                                "clusterrole/system:controller:endpointslicemirroring-controller",
                                "clusterrolebinding/node-feature-discovery",
                                "clusterrole/system:controller:ephemeral-volume-controller",
                                "clusterrolebinding/rabbitmq-cluster-operator",
                                "clusterrolebinding/rabbitmq-messaging-topology-operator",
                                "clusterrole/system:controller:expand-controller",
                                "clusterrolebinding/secretgen-controller-cluster-role-binding",
                                "clusterrole/system:controller:generic-garbage-collector",
                                "clusterrole/system:controller:horizontal-pod-autoscaler",
                                "clusterrole/system:controller:job-controller",
                                "clusterrole/system:controller:namespace-controller",
                                "clusterrolebinding/system:basic-user",
                                "clusterrole/system:controller:node-controller",
                                "clusterrolebinding/system:controller:attachdetach-controller",
                                "clusterrolebinding/system:controller:certificate-controller",
                                "clusterrole/system:controller:persistent-volume-binder",
                                "clusterrolebinding/system:controller:clusterrole-aggregation-controller",
                                "clusterrole/system:controller:pod-garbage-collector",
                                "clusterrolebinding/system:controller:cronjob-controller",
                                "clusterrole/system:controller:pv-protection-controller",
                                "clusterrolebinding/system:controller:daemon-set-controller",
                                "clusterrole/system:controller:pvc-protection-controller",
                                "clusterrolebinding/system:controller:deployment-controller",
                                "clusterrole/system:controller:replicaset-controller",
                                "clusterrolebinding/system:controller:disruption-controller",
                                "clusterrolebinding/system:controller:endpoint-controller",
                                "clusterrole/system:controller:replication-controller",
                                "clusterrolebinding/system:controller:endpointslice-controller",
                                "clusterrole/system:controller:resourcequota-controller",
                                "clusterrolebinding/system:controller:endpointslicemirroring-controller",
                                "clusterrole/system:controller:root-ca-cert-publisher",
                                "clusterrolebinding/system:controller:ephemeral-volume-controller",
                                "clusterrole/system:controller:route-controller",
                                "clusterrolebinding/system:controller:expand-controller",
                                "clusterrole/system:controller:service-account-controller",
                                "clusterrolebinding/system:controller:generic-garbage-collector",
                                "clusterrole/system:controller:service-controller",
                                "clusterrolebinding/system:controller:horizontal-pod-autoscaler",
                                "clusterrole/system:controller:statefulset-controller",
                                "clusterrolebinding/system:controller:job-controller",
                                "clusterrole/system:controller:ttl-after-finished-controller",
                                "clusterrolebinding/system:controller:namespace-controller",
                                "clusterrolebinding/system:controller:node-controller",
                                "clusterrole/system:controller:ttl-controller",
                                "clusterrolebinding/system:controller:persistent-volume-binder",
                                "clusterrole/system:coredns",
                                "clusterrole/system:discovery",
                                "clusterrole/system:heapster",
                                "clusterrolebinding/system:controller:pod-garbage-collector",
                                "clusterrole/system:kube-aggregator",
                                "clusterrole/system:kube-controller-manager",
                                "clusterrolebinding/system:controller:pv-protection-controller",
                                "clusterrole/system:kube-dns",
                                "clusterrolebinding/system:controller:pvc-protection-controller",
                                "clusterrole/system:kube-scheduler",
                                "clusterrole/system:kubelet-api-admin",
                                "clusterrolebinding/system:controller:replicaset-controller",
                                "clusterrole/system:monitoring",
                                "clusterrolebinding/system:controller:replication-controller",
                                "clusterrole/system:node",
                                "clusterrolebinding/system:controller:resourcequota-controller",
                                "clusterrole/system:node-bootstrapper",
                                "clusterrolebinding/system:controller:root-ca-cert-publisher",
                                "clusterrole/system:node-problem-detector",
                                "clusterrolebinding/system:controller:route-controller",
                                "clusterrole/system:node-proxier",
                                "clusterrolebinding/system:controller:service-account-controller",
                                "clusterrole/system:persistent-volume-provisioner",
                                "clusterrole/system:public-info-viewer",
                                "clusterrolebinding/system:controller:service-controller",
                                "clusterrole/system:service-account-issuer-discovery",
                                "clusterrolebinding/system:controller:statefulset-controller",
                                "clusterrole/system:volume-scheduler",
                                "clusterrolebinding/system:controller:ttl-after-finished-controller",
                                "clusterrole/vector",
                                "clusterrole/view",
                                "clusterrolebinding/system:controller:ttl-controller",
                                "clusterrolebinding/system:coredns",
                                "clusterrolebinding/system:discovery",
                                "clusterrolebinding/system:kube-controller-manager",
                                "clusterrolebinding/system:kube-dns",
                                "clusterrolebinding/system:kube-scheduler",
                                "clusterrolebinding/system:monitoring",
                                "clusterrolebinding/system:node",
                                "clusterrolebinding/system:node-proxier",
                                "clusterrolebinding/system:public-info-viewer",
                                "clusterrolebinding/system:service-account-issuer-discovery",
                                "clusterrolebinding/system:volume-scheduler",
                                "clusterrolebinding/vector",
                                "storageclass/general",
                                "namespace/auth-system",
                                "namespace/cert-manager",
                                "namespace/default",
                                "namespace/envoy-gateway-system",
                                "namespace/ingress-nginx",
                                "namespace/kube-node-lease",
                                "namespace/kube-public",
                                "namespace/kube-system",
                                "namespace/local-path-storage",
                                "namespace/monitoring",
                                "namespace/openstack",
                                "namespace/secretgen-controller"
                            ],
                            "zuul_log_id": "0242ac17-0011-cc15-c3f6-00000000001a-1-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000017",
                        "name": "describe-kubernetes-objects",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/describe-kubernetes-objects"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:25:28.292351Z",
                            "start": "2026-04-04T10:25:09.749774Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-00000000001a",
                        "name": "Gathering descriptions for cluster scoped objects"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/tmp/logs/objects/namespaced",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/tmp/logs/objects/namespaced",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": null,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/tmp/logs/objects/namespaced",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/tmp/logs/objects/namespaced",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000017",
                        "name": "describe-kubernetes-objects",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/describe-kubernetes-objects"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:25:29.302699Z",
                            "start": "2026-04-04T10:25:28.299574Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-00000000001b",
                        "name": "creating directory for namespace scoped objects"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "shell",
                            "changed": true,
                            "cmd": "set -e\nexport OBJECT_TYPE=configmaps,cronjobs,daemonsets,deployment,endpoints,ingresses,jobs,networkpolicies,pods,podsecuritypolicies,persistentvolumeclaims,rolebindings,roles,secrets,serviceaccounts,services,statefulsets\nexport PARALLELISM_FACTOR=2\nfunction get_namespaces () {\n  kubectl get namespaces -o name | awk -F '/' '{ print $NF }'\n}\n\nfunction list_namespaced_objects () {\n  export NAMESPACE=$1\n  printf ${OBJECT_TYPE} | xargs -d ',' -I {} -P1 -n1 bash -c 'echo \"${NAMESPACE} $@\"' _ {}\n}\nexport -f list_namespaced_objects\n\nfunction name_objects () {\n  input=($1)\n  export NAMESPACE=${input[0]}\n  export OBJECT=${input[1]}\n  kubectl get -n ${NAMESPACE} ${OBJECT} -o name | xargs -L1 -I {} -P1 -n1 bash -c 'echo \"${NAMESPACE} ${OBJECT} $@\"' _ {}\n}\nexport -f name_objects\n\nfunction get_objects () {\n  input=($1)\n  export NAMESPACE=${input[0]}\n  export OBJECT=${input[1]}\n  export NAME=${input[2]#*/}\n  echo \"${NAMESPACE}/${OBJECT}/${NAME}\"\n  DIR=\"/tmp/logs/objects/namespaced/${NAMESPACE}/${OBJECT}\"\n  mkdir -p ${DIR}\n  kubectl get -n ${NAMESPACE} ${OBJECT} ${NAME} -o yaml > \"${DIR}/${NAME}.yaml\"\n  kubectl describe -n ${NAMESPACE} ${OBJECT} ${NAME} > \"${DIR}/${NAME}.txt\"\n}\nexport -f get_objects\n\nget_namespaces |  xargs -r -n 1 -P ${PARALLELISM_FACTOR} -I {} bash -c 'list_namespaced_objects \"$@\"' _ {} |  xargs -r -n 1 -P ${PARALLELISM_FACTOR} -I {} bash -c 'name_objects \"$@\"' _ {} |  xargs -r -n 1 -P ${PARALLELISM_FACTOR} -I {} bash -c 'get_objects \"$@\"' _ {}",
                            "delta": "0:00:51.626810",
                            "end": "2026-04-04 10:26:21.149286",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -e\nexport OBJECT_TYPE=configmaps,cronjobs,daemonsets,deployment,endpoints,ingresses,jobs,networkpolicies,pods,podsecuritypolicies,persistentvolumeclaims,rolebindings,roles,secrets,serviceaccounts,services,statefulsets\nexport PARALLELISM_FACTOR=2\nfunction get_namespaces () {\n  kubectl get namespaces -o name | awk -F '/' '{ print $NF }'\n}\n\nfunction list_namespaced_objects () {\n  export NAMESPACE=$1\n  printf ${OBJECT_TYPE} | xargs -d ',' -I {} -P1 -n1 bash -c 'echo \"${NAMESPACE} $@\"' _ {}\n}\nexport -f list_namespaced_objects\n\nfunction name_objects () {\n  input=($1)\n  export NAMESPACE=${input[0]}\n  export OBJECT=${input[1]}\n  kubectl get -n ${NAMESPACE} ${OBJECT} -o name | xargs -L1 -I {} -P1 -n1 bash -c 'echo \"${NAMESPACE} ${OBJECT} $@\"' _ {}\n}\nexport -f name_objects\n\nfunction get_objects () {\n  input=($1)\n  export NAMESPACE=${input[0]}\n  export OBJECT=${input[1]}\n  export NAME=${input[2]#*/}\n  echo \"${NAMESPACE}/${OBJECT}/${NAME}\"\n  DIR=\"/tmp/logs/objects/namespaced/${NAMESPACE}/${OBJECT}\"\n  mkdir -p ${DIR}\n  kubectl get -n ${NAMESPACE} ${OBJECT} ${NAME} -o yaml > \"${DIR}/${NAME}.yaml\"\n  kubectl describe -n ${NAMESPACE} ${OBJECT} ${NAME} > \"${DIR}/${NAME}.txt\"\n}\nexport -f get_objects\n\nget_namespaces |  xargs -r -n 1 -P ${PARALLELISM_FACTOR} -I {} bash -c 'list_namespaced_objects \"$@\"' _ {} |  xargs -r -n 1 -P ${PARALLELISM_FACTOR} -I {} bash -c 'name_objects \"$@\"' _ {} |  xargs -r -n 1 -P ${PARALLELISM_FACTOR} -I {} bash -c 'get_objects \"$@\"' _ {}",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-cc15-c3f6-00000000001c-1-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-04-04 10:25:29.522476",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value\nxargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value\nxargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nauth-system/configmaps/keycloak-env-vars\ncert-manager/configmaps/cert-manager-webhook\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nauth-system/configmaps/kube-root-ca.crt\ncert-manager/configmaps/kube-root-ca.crt\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/deployment/cert-manager\ncert-manager/deployment/cert-manager-cainjector\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/deployment/cert-manager-webhook\nauth-system/endpoints/keycloak\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nauth-system/endpoints/keycloak-headless\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nauth-system/endpoints/keycloak-metrics\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/endpoints/cert-manager\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/endpoints/cert-manager-webhook\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nauth-system/ingresses/keycloak\nauth-system/pods/keycloak-0\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/pods/cert-manager-75969b45cf-d5c4v\ncert-manager/pods/cert-manager-cainjector-84f45b698d-v64n5\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/pods/cert-manager-webhook-9997c64fc-8w8sz\ncert-manager/rolebindings/cert-manager-cainjector:leaderelection\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/rolebindings/cert-manager-webhook:dynamic-serving\ncert-manager/rolebindings/cert-manager:leaderelection\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/roles/cert-manager-cainjector:leaderelection\ncert-manager/roles/cert-manager-webhook:dynamic-serving\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/roles/cert-manager:leaderelection\nauth-system/secrets/keycloak\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nauth-system/secrets/keycloak-externaldb\nauth-system/secrets/keycloak.199-204-45-153.nip.io-tls\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nauth-system/secrets/sh.helm.release.v1.keycloak.v1\ncert-manager/secrets/cert-manager-selfsigned-ca\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/secrets/cert-manager-webhook-ca\ncert-manager/secrets/kube-prometheus-stack-ca\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/secrets/sh.helm.release.v1.cert-manager.v1\ncert-manager/serviceaccounts/cert-manager\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/serviceaccounts/cert-manager-cainjector\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/serviceaccounts/cert-manager-webhook\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/serviceaccounts/default\nauth-system/serviceaccounts/default\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nauth-system/serviceaccounts/keycloak\ncert-manager/services/cert-manager\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ncert-manager/services/cert-manager-webhook\nauth-system/services/keycloak\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nauth-system/services/keycloak-headless\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nauth-system/services/keycloak-metrics\nauth-system/statefulsets/keycloak\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\ndefault/configmaps/kube-root-ca.crt\nenvoy-gateway-system/configmaps/envoy-gateway-config\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/configmaps/kube-root-ca.crt\ndefault/endpoints/kubernetes\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/deployment/envoy-gateway\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/endpoints/envoy-gateway\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/pods/envoy-gateway-78446f96c9-7zvgb\nenvoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-certgen\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-infra-manager\nenvoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-leader-election-rolebinding\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/roles/envoy-gateway-gateway-helm-certgen\nenvoy-gateway-system/roles/envoy-gateway-gateway-helm-infra-manager\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/roles/envoy-gateway-gateway-helm-leader-election-role\ndefault/serviceaccounts/default\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ndefault/services/kubernetes\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/secrets/envoy\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/secrets/envoy-gateway\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/secrets/envoy-oidc-hmac\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\nenvoy-gateway-system/secrets/envoy-rate-limit\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/secrets/sh.helm.release.v1.envoy-gateway.v1\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/serviceaccounts/default\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/serviceaccounts/envoy-gateway\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/serviceaccounts/envoy-gateway-gateway-helm-certgen\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nenvoy-gateway-system/services/envoy-gateway\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/configmaps/ingress-nginx-controller\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/configmaps/ingress-nginx-tcp\ningress-nginx/configmaps/ingress-nginx-udp\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/configmaps/kube-root-ca.crt\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-node-lease/configmaps/kube-root-ca.crt\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/daemonsets/ingress-nginx-controller\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/deployment/ingress-nginx-defaultbackend\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/endpoints/ingress-nginx-controller\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/endpoints/ingress-nginx-controller-admission\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/endpoints/ingress-nginx-controller-metrics\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/endpoints/ingress-nginx-defaultbackend\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/pods/ingress-nginx-controller-dmjt4\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/pods/ingress-nginx-defaultbackend-6987ff55cf-plsk9\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/rolebindings/ingress-nginx\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/roles/ingress-nginx\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/serviceaccounts/default\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/serviceaccounts/ingress-nginx\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/serviceaccounts/ingress-nginx-backend\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/secrets/ingress-nginx-admission\ningress-nginx/secrets/sh.helm.release.v1.ingress-nginx.v1\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/services/ingress-nginx-controller\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/services/ingress-nginx-controller-admission\ningress-nginx/services/ingress-nginx-controller-metrics\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\ningress-nginx/services/ingress-nginx-defaultbackend\nkube-node-lease/serviceaccounts/default\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-system/configmaps/cilium-config\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-system/configmaps/coredns\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-system/configmaps/extension-apiserver-authentication\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-system/configmaps/kube-apiserver-legacy-service-account-token-tracking\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-system/configmaps/kube-proxy\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-system/configmaps/kube-root-ca.crt\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-public/configmaps/cluster-info\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-system/configmaps/kubeadm-config\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nerror: the server doesn't have a resource type \"podsecuritypolicies\"\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-public/configmaps/kube-root-ca.crt\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-system/configmaps/kubelet-config\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-system/daemonsets/cilium\nkube-system/daemonsets/kube-proxy\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nxargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value\nkube-system/deployment/cilium-operator\nkube-system/deployment/coredns\nkube-system/endpoints/kube-dns\nkube-system/endpoints/kube-prometheus-stack-coredns\nkube-system/endpoints/kube-prometheus-stack-kube-controller-manager\nkube-system/endpoints/kube-prometheus-stack-kube-etcd\nkube-system/endpoints/kube-prometheus-stack-kube-proxy\nkube-system/endpoints/kube-prometheus-stack-kube-scheduler\nkube-system/endpoints/kube-prometheus-stack-kubelet\nkube-system/pods/cilium-operator-869df985b8-nqvtb\nkube-system/pods/cilium-sk6n5\nkube-system/pods/coredns-67659f764b-mdrt6\nkube-system/pods/coredns-67659f764b-xb6lv\nkube-system/pods/etcd-instance\nkube-system/pods/kube-apiserver-instance\nkube-system/pods/kube-controller-manager-instance\nkube-system/pods/kube-proxy-cxlxv\nkube-system/pods/kube-scheduler-instance\nkube-system/pods/kube-vip-instance\nkube-public/rolebindings/kubeadm:bootstrap-signer-clusterinfo\nkube-public/rolebindings/system:controller:bootstrap-signer\nkube-public/roles/kubeadm:bootstrap-signer-clusterinfo\nkube-public/roles/system:controller:bootstrap-signer\nkube-system/rolebindings/cilium-config-agent\nkube-system/rolebindings/kube-proxy\nkube-system/rolebindings/kubeadm:kubeadm-certs\nkube-system/rolebindings/kubeadm:kubelet-config\nkube-system/rolebindings/kubeadm:nodes-kubeadm-config\nkube-system/rolebindings/system::extension-apiserver-authentication-reader\nkube-system/rolebindings/system::leader-locking-kube-controller-manager\nkube-system/rolebindings/system::leader-locking-kube-scheduler\nkube-system/rolebindings/system:controller:bootstrap-signer\nkube-system/rolebindings/system:controller:cloud-provider\nkube-system/rolebindings/system:controller:token-cleaner\nkube-public/serviceaccounts/default\nkube-system/roles/cilium-config-agent\nkube-system/roles/extension-apiserver-authentication-reader\nkube-system/roles/kube-proxy\nkube-system/roles/kubeadm:kubeadm-certs\nkube-system/roles/kubeadm:kubelet-config\nkube-system/roles/kubeadm:nodes-kubeadm-config\nkube-system/roles/system::leader-locking-kube-controller-manager\nkube-system/roles/system::leader-locking-kube-scheduler\nkube-system/roles/system:controller:bootstrap-signer\nkube-system/roles/system:controller:cloud-provider\nkube-system/roles/system:controller:token-cleaner\nkube-system/secrets/bootstrap-token-3n8qak\nkube-system/secrets/bootstrap-token-w4o225\nkube-system/secrets/kubeadm-certs\nkube-system/secrets/sh.helm.release.v1.cilium.v1\nkube-system/serviceaccounts/attachdetach-controller\nkube-system/serviceaccounts/bootstrap-signer\nkube-system/serviceaccounts/certificate-controller\nkube-system/serviceaccounts/cilium\nkube-system/serviceaccounts/cilium-operator\nkube-system/serviceaccounts/clusterrole-aggregation-controller\nkube-system/serviceaccounts/coredns\nkube-system/serviceaccounts/cronjob-controller\nkube-system/serviceaccounts/daemon-set-controller\nkube-system/serviceaccounts/default\nkube-system/serviceaccounts/deployment-controller\nkube-system/serviceaccounts/disruption-controller\nkube-system/serviceaccounts/endpoint-controller\nkube-system/serviceaccounts/endpointslice-controller\nkube-system/serviceaccounts/endpointslicemirroring-controller\nkube-system/serviceaccounts/ephemeral-volume-controller\nkube-system/serviceaccounts/expand-controller\nkube-system/serviceaccounts/generic-garbage-collector\nkube-system/serviceaccounts/horizontal-pod-autoscaler\nkube-system/serviceaccounts/job-controller\nkube-system/serviceaccounts/kube-proxy\nkube-system/serviceaccounts/namespace-controller\nkube-system/serviceaccounts/node-controller\nkube-system/serviceaccounts/persistent-volume-binder\nkube-system/serviceaccounts/pod-garbage-collector\nkube-system/serviceaccounts/pv-protection-controller\nkube-system/serviceaccounts/pvc-protection-controller\nkube-system/serviceaccounts/replicaset-controller\nkube-system/serviceaccounts/replication-controller\nkube-system/serviceaccounts/resourcequota-controller\nkube-system/serviceaccounts/root-ca-cert-publisher\nkube-system/services/kube-dns\nkube-system/serviceaccounts/service-account-controller\nkube-system/services/kube-prometheus-stack-coredns\nkube-system/serviceaccounts/service-controller\nkube-system/services/kube-prometheus-stack-kube-controller-manager\nkube-system/serviceaccounts/statefulset-controller\nkube-system/services/kube-prometheus-stack-kube-etcd\nkube-system/serviceaccounts/token-cleaner\nkube-system/services/kube-prometheus-stack-kube-proxy\nkube-system/services/kube-prometheus-stack-kube-scheduler\nkube-system/serviceaccounts/ttl-after-finished-controller\nkube-system/services/kube-prometheus-stack-kubelet\nkube-system/serviceaccounts/ttl-controller\nlocal-path-storage/configmaps/kube-root-ca.crt\nlocal-path-storage/configmaps/local-path-config\nlocal-path-storage/deployment/local-path-provisioner\nmonitoring/configmaps/goldpinger-zap\nmonitoring/configmaps/ipmi-exporter\nmonitoring/configmaps/kube-prometheus-stack-alertmanager-overview\nmonitoring/configmaps/kube-prometheus-stack-apiserver\nmonitoring/configmaps/kube-prometheus-stack-cluster-total\nmonitoring/configmaps/kube-prometheus-stack-controller-manager\nmonitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster\nmonitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster-advanced\nmonitoring/configmaps/kube-prometheus-stack-dashboard-goldpinger\nmonitoring/configmaps/kube-prometheus-stack-dashboard-haproxy\nmonitoring/configmaps/kube-prometheus-stack-dashboard-host-details\nmonitoring/configmaps/kube-prometheus-stack-dashboard-hosts-overview\nmonitoring/configmaps/kube-prometheus-stack-dashboard-node-exporter-full\nmonitoring/configmaps/kube-prometheus-stack-dashboard-osd-device-details\nmonitoring/configmaps/kube-prometheus-stack-dashboard-osds-overview\nmonitoring/configmaps/kube-prometheus-stack-dashboard-pool-detail\nmonitoring/configmaps/kube-prometheus-stack-dashboard-pool-overview\nmonitoring/configmaps/kube-prometheus-stack-dashboard-rbd-details\nmonitoring/configmaps/kube-prometheus-stack-dashboard-rbd-overview\nmonitoring/configmaps/kube-prometheus-stack-etcd\nmonitoring/configmaps/kube-prometheus-stack-grafana\nmonitoring/configmaps/kube-prometheus-stack-grafana-config-dashboards\nmonitoring/configmaps/kube-prometheus-stack-grafana-datasource\nmonitoring/configmaps/kube-prometheus-stack-grafana-overview\nmonitoring/configmaps/kube-prometheus-stack-k8s-coredns\nmonitoring/configmaps/kube-prometheus-stack-k8s-resources-cluster\nmonitoring/configmaps/kube-prometheus-stack-k8s-resources-multicluster\nmonitoring/configmaps/kube-prometheus-stack-k8s-resources-namespace\nmonitoring/configmaps/kube-prometheus-stack-k8s-resources-node\nmonitoring/configmaps/kube-prometheus-stack-k8s-resources-pod\nmonitoring/configmaps/kube-prometheus-stack-k8s-resources-workload\nmonitoring/configmaps/kube-prometheus-stack-k8s-resources-workloads-namespace\nmonitoring/configmaps/kube-prometheus-stack-kubelet\nmonitoring/configmaps/kube-prometheus-stack-namespace-by-pod\nmonitoring/configmaps/kube-prometheus-stack-namespace-by-workload\nmonitoring/configmaps/kube-prometheus-stack-node-cluster-rsrc-use\nmonitoring/configmaps/kube-prometheus-stack-node-exporter\nmonitoring/daemonsets/goldpinger\nmonitoring/configmaps/kube-prometheus-stack-node-rsrc-use\nmonitoring/daemonsets/ipmi-exporter\nmonitoring/configmaps/kube-prometheus-stack-nodes\nmonitoring/daemonsets/kube-prometheus-stack-prometheus-node-exporter\nmonitoring/configmaps/kube-prometheus-stack-nodes-darwin\nmonitoring/daemonsets/node-feature-discovery-worker\nmonitoring/configmaps/kube-prometheus-stack-persistentvolumesusage\nmonitoring/daemonsets/vector\nmonitoring/configmaps/kube-prometheus-stack-pod-total\nmonitoring/configmaps/kube-prometheus-stack-prometheus\nmonitoring/configmaps/kube-prometheus-stack-prometheus-tls\nmonitoring/configmaps/kube-prometheus-stack-proxy\nmonitoring/configmaps/kube-prometheus-stack-scheduler\nmonitoring/configmaps/kube-prometheus-stack-workload-total\nmonitoring/configmaps/kube-root-ca.crt\nmonitoring/configmaps/loki\nmonitoring/configmaps/loki-alerting-rules\nmonitoring/configmaps/loki-dashboards-1\nmonitoring/configmaps/loki-dashboards-2\nmonitoring/configmaps/loki-gateway\nmonitoring/configmaps/loki-runtime\nmonitoring/configmaps/node-feature-discovery-worker-conf\nmonitoring/configmaps/prometheus-kube-prometheus-stack-prometheus-rulefiles-0\nmonitoring/configmaps/vector\nmonitoring/deployment/kube-prometheus-stack-grafana\nmonitoring/deployment/kube-prometheus-stack-kube-state-metrics\nmonitoring/deployment/kube-prometheus-stack-operator\nmonitoring/deployment/loki-gateway\nmonitoring/deployment/node-feature-discovery-master\nmonitoring/deployment/prometheus-pushgateway\nmonitoring/endpoints/alertmanager-operated\nmonitoring/endpoints/goldpinger\nlocal-path-storage/pods/local-path-provisioner-679c578f5-m52kp\nmonitoring/endpoints/kube-prometheus-stack-alertmanager\nmonitoring/endpoints/kube-prometheus-stack-grafana\nmonitoring/endpoints/kube-prometheus-stack-kube-state-metrics\nmonitoring/endpoints/kube-prometheus-stack-operator\nmonitoring/endpoints/kube-prometheus-stack-prometheus\nmonitoring/endpoints/kube-prometheus-stack-prometheus-node-exporter\nmonitoring/endpoints/loki\nmonitoring/endpoints/loki-gateway\nmonitoring/endpoints/loki-headless\nmonitoring/endpoints/loki-memberlist\nmonitoring/endpoints/node-feature-discovery-master\nmonitoring/endpoints/prometheus-operated\nmonitoring/endpoints/prometheus-pushgateway\nmonitoring/endpoints/vector-headless\nmonitoring/ingresses/kube-prometheus-stack-alertmanager\nmonitoring/ingresses/kube-prometheus-stack-grafana\nmonitoring/ingresses/kube-prometheus-stack-prometheus\nlocal-path-storage/secrets/sh.helm.release.v1.local-path-provisioner.v1\nlocal-path-storage/serviceaccounts/default\nlocal-path-storage/serviceaccounts/local-path-provisioner\nmonitoring/pods/alertmanager-kube-prometheus-stack-alertmanager-0\nmonitoring/pods/goldpinger-gz8xd\nmonitoring/pods/kube-prometheus-stack-grafana-69fc56495c-5h4gd\nmonitoring/pods/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj\nmonitoring/pods/kube-prometheus-stack-operator-cc88b86d6-xspgk\nmonitoring/pods/kube-prometheus-stack-prometheus-node-exporter-jmj6n\nmonitoring/pods/loki-0\nmonitoring/pods/loki-gateway-5cc45946cd-wp7ks\nmonitoring/pods/node-feature-discovery-master-68cf466d5f-qknzk\nmonitoring/pods/node-feature-discovery-worker-7l5z7\nmonitoring/pods/prometheus-kube-prometheus-stack-prometheus-0\nmonitoring/pods/prometheus-pushgateway-5c694d99cc-jnjt7\nmonitoring/pods/vector-75jgf\nmonitoring/persistentvolumeclaims/alertmanager-kube-prometheus-stack-alertmanager-db-alertmanager-kube-prometheus-stack-alertmanager-0\nmonitoring/persistentvolumeclaims/prometheus-kube-prometheus-stack-prometheus-db-prometheus-kube-prometheus-stack-prometheus-0\nmonitoring/persistentvolumeclaims/storage-loki-0\nmonitoring/rolebindings/kube-prometheus-stack-grafana\nmonitoring/rolebindings/kube-prometheus-stack-pod-tls-sidecar\nmonitoring/roles/kube-prometheus-stack-grafana\nmonitoring/roles/kube-prometheus-stack-pod-tls-sidecar\nmonitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager\nmonitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-generated\nmonitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-tls-assets-0\nmonitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-web-config\nmonitoring/secrets/alertmanager-tls\nmonitoring/secrets/grafana-tls\nmonitoring/secrets/kube-prometheus-stack-admission\nmonitoring/secrets/kube-prometheus-stack-alertmanager-client-secret\nmonitoring/secrets/kube-prometheus-stack-alertmanager-cookie-secret\nmonitoring/secrets/kube-prometheus-stack-alertmanager-oauth2-proxy\nmonitoring/secrets/kube-prometheus-stack-etcd-client-cert\nmonitoring/secrets/kube-prometheus-stack-grafana\nmonitoring/secrets/kube-prometheus-stack-grafana-client-secret\nmonitoring/secrets/kube-prometheus-stack-prometheus-client-secret\nmonitoring/secrets/kube-prometheus-stack-prometheus-cookie-secret\nmonitoring/secrets/kube-prometheus-stack-prometheus-node-exporter-jmj6n-tls\nmonitoring/secrets/kube-prometheus-stack-prometheus-oauth2-proxy\nmonitoring/secrets/prometheus-kube-prometheus-stack-prometheus\nmonitoring/secrets/prometheus-kube-prometheus-stack-prometheus-0-tls\nmonitoring/secrets/prometheus-kube-prometheus-stack-prometheus-tls-assets-0\nmonitoring/secrets/prometheus-kube-prometheus-stack-prometheus-web-config\nmonitoring/secrets/prometheus-tls\nmonitoring/secrets/sh.helm.release.v1.goldpinger.v1\nmonitoring/secrets/sh.helm.release.v1.kube-prometheus-stack.v1\nmonitoring/secrets/sh.helm.release.v1.loki.v1\nmonitoring/secrets/sh.helm.release.v1.node-feature-discovery.v1\nmonitoring/secrets/sh.helm.release.v1.prometheus-pushgateway.v1\nmonitoring/secrets/sh.helm.release.v1.vector.v1\nopenstack/configmaps/keepalived-bin\nopenstack/configmaps/kube-root-ca.crt\nopenstack/configmaps/memcached-memcached-bin\nopenstack/configmaps/messaging-topology-operator-leader-election\nopenstack/configmaps/percona-xtradb-haproxy\nopenstack/configmaps/percona-xtradb-pxc\nopenstack/configmaps/rabbitmq-cluster-operator-leader-election\nmonitoring/serviceaccounts/default\nmonitoring/serviceaccounts/goldpinger\nmonitoring/serviceaccounts/kube-prometheus-stack-alertmanager\nmonitoring/serviceaccounts/kube-prometheus-stack-grafana\nmonitoring/serviceaccounts/kube-prometheus-stack-kube-state-metrics\nmonitoring/serviceaccounts/kube-prometheus-stack-operator\nmonitoring/serviceaccounts/kube-prometheus-stack-prometheus\nmonitoring/serviceaccounts/kube-prometheus-stack-prometheus-node-exporter\nmonitoring/serviceaccounts/loki\nmonitoring/serviceaccounts/node-feature-discovery\nmonitoring/serviceaccounts/node-feature-discovery-worker\nmonitoring/serviceaccounts/prometheus-pushgateway\nmonitoring/serviceaccounts/vector\nmonitoring/services/alertmanager-operated\nmonitoring/services/goldpinger\nmonitoring/services/kube-prometheus-stack-alertmanager\nmonitoring/services/kube-prometheus-stack-grafana\nmonitoring/services/kube-prometheus-stack-kube-state-metrics\nmonitoring/services/kube-prometheus-stack-operator\nmonitoring/services/kube-prometheus-stack-prometheus\nmonitoring/services/kube-prometheus-stack-prometheus-node-exporter\nmonitoring/services/loki\nmonitoring/services/loki-gateway\nmonitoring/services/loki-headless\nmonitoring/services/loki-memberlist\nmonitoring/services/node-feature-discovery-master\nmonitoring/services/prometheus-operated\nmonitoring/services/prometheus-pushgateway\nmonitoring/services/vector-headless\nmonitoring/statefulsets/alertmanager-kube-prometheus-stack-alertmanager\nmonitoring/statefulsets/loki\nmonitoring/statefulsets/prometheus-kube-prometheus-stack-prometheus\nsecretgen-controller/configmaps/kube-root-ca.crt\nopenstack/daemonsets/keepalived\nopenstack/deployment/memcached-memcached\nopenstack/deployment/pxc-operator\nopenstack/deployment/rabbitmq-cluster-operator\nopenstack/deployment/rabbitmq-messaging-topology-operator\nopenstack/endpoints/memcached\nopenstack/endpoints/memcached-metrics\nopenstack/endpoints/percona-xtradb-cluster-operator\nopenstack/endpoints/percona-xtradb-haproxy\nopenstack/endpoints/percona-xtradb-haproxy-metrics\nopenstack/endpoints/percona-xtradb-haproxy-replicas\nopenstack/endpoints/percona-xtradb-pxc\nopenstack/endpoints/percona-xtradb-pxc-unready\nopenstack/endpoints/rabbitmq-messaging-topology-operator-webhook\nsecretgen-controller/deployment/secretgen-controller\nopenstack/pods/keepalived-wktjf\nopenstack/pods/memcached-memcached-c47d949f5-vc5cm\nopenstack/pods/percona-xtradb-haproxy-0\nopenstack/pods/percona-xtradb-pxc-0\nopenstack/pods/pxc-operator-59d66cbc78-clxvf\nopenstack/pods/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf\nopenstack/pods/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc\nopenstack/persistentvolumeclaims/datadir-percona-xtradb-pxc-0\nopenstack/rolebindings/keepalived\nopenstack/rolebindings/pxc-operator\nopenstack/rolebindings/rabbitmq-cluster-operator\nopenstack/rolebindings/rabbitmq-messaging-topology-operator\nsecretgen-controller/pods/secretgen-controller-5cf976ccc7-vwxz4\nopenstack/roles/keepalived\nopenstack/roles/pxc-operator\nopenstack/roles/rabbitmq-cluster-operator\nopenstack/roles/rabbitmq-messaging-topology-operator\nopenstack/secrets/internal-percona-xtradb\nopenstack/secrets/keepalived-etc\nopenstack/secrets/percona-xtradb\nopenstack/secrets/rabbitmq-messaging-topology-operator-webhook\nopenstack/secrets/sh.helm.release.v1.memcached.v1\nopenstack/secrets/sh.helm.release.v1.pxc-operator.v1\nopenstack/secrets/sh.helm.release.v1.rabbitmq-cluster-operator.v1\nopenstack/serviceaccounts/default\nopenstack/serviceaccounts/keepalived\nopenstack/serviceaccounts/memcached-memcached\nopenstack/serviceaccounts/pxc-operator\nopenstack/serviceaccounts/rabbitmq-cluster-operator\nopenstack/serviceaccounts/rabbitmq-messaging-topology-operator\nopenstack/statefulsets/percona-xtradb-haproxy\nopenstack/statefulsets/percona-xtradb-pxc\nopenstack/services/memcached\nopenstack/services/memcached-metrics\nopenstack/services/percona-xtradb-cluster-operator\nopenstack/services/percona-xtradb-haproxy\nopenstack/services/percona-xtradb-haproxy-metrics\nopenstack/services/percona-xtradb-haproxy-replicas\nopenstack/services/percona-xtradb-pxc\nopenstack/services/percona-xtradb-pxc-unready\nopenstack/services/rabbitmq-messaging-topology-operator-webhook\nsecretgen-controller/serviceaccounts/default\nsecretgen-controller/serviceaccounts/secretgen-controller-sa",
                            "stdout_lines": [
                                "xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value",
                                "xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value",
                                "xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "auth-system/configmaps/keycloak-env-vars",
                                "cert-manager/configmaps/cert-manager-webhook",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "auth-system/configmaps/kube-root-ca.crt",
                                "cert-manager/configmaps/kube-root-ca.crt",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/deployment/cert-manager",
                                "cert-manager/deployment/cert-manager-cainjector",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/deployment/cert-manager-webhook",
                                "auth-system/endpoints/keycloak",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "auth-system/endpoints/keycloak-headless",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "auth-system/endpoints/keycloak-metrics",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/endpoints/cert-manager",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/endpoints/cert-manager-webhook",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "auth-system/ingresses/keycloak",
                                "auth-system/pods/keycloak-0",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/pods/cert-manager-75969b45cf-d5c4v",
                                "cert-manager/pods/cert-manager-cainjector-84f45b698d-v64n5",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/pods/cert-manager-webhook-9997c64fc-8w8sz",
                                "cert-manager/rolebindings/cert-manager-cainjector:leaderelection",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/rolebindings/cert-manager-webhook:dynamic-serving",
                                "cert-manager/rolebindings/cert-manager:leaderelection",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/roles/cert-manager-cainjector:leaderelection",
                                "cert-manager/roles/cert-manager-webhook:dynamic-serving",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/roles/cert-manager:leaderelection",
                                "auth-system/secrets/keycloak",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "auth-system/secrets/keycloak-externaldb",
                                "auth-system/secrets/keycloak.199-204-45-153.nip.io-tls",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "auth-system/secrets/sh.helm.release.v1.keycloak.v1",
                                "cert-manager/secrets/cert-manager-selfsigned-ca",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/secrets/cert-manager-webhook-ca",
                                "cert-manager/secrets/kube-prometheus-stack-ca",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/secrets/sh.helm.release.v1.cert-manager.v1",
                                "cert-manager/serviceaccounts/cert-manager",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/serviceaccounts/cert-manager-cainjector",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/serviceaccounts/cert-manager-webhook",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/serviceaccounts/default",
                                "auth-system/serviceaccounts/default",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "auth-system/serviceaccounts/keycloak",
                                "cert-manager/services/cert-manager",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "cert-manager/services/cert-manager-webhook",
                                "auth-system/services/keycloak",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "auth-system/services/keycloak-headless",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "auth-system/services/keycloak-metrics",
                                "auth-system/statefulsets/keycloak",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "default/configmaps/kube-root-ca.crt",
                                "envoy-gateway-system/configmaps/envoy-gateway-config",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/configmaps/kube-root-ca.crt",
                                "default/endpoints/kubernetes",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/deployment/envoy-gateway",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/endpoints/envoy-gateway",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/pods/envoy-gateway-78446f96c9-7zvgb",
                                "envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-certgen",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-infra-manager",
                                "envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-leader-election-rolebinding",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/roles/envoy-gateway-gateway-helm-certgen",
                                "envoy-gateway-system/roles/envoy-gateway-gateway-helm-infra-manager",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/roles/envoy-gateway-gateway-helm-leader-election-role",
                                "default/serviceaccounts/default",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "default/services/kubernetes",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/secrets/envoy",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/secrets/envoy-gateway",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/secrets/envoy-oidc-hmac",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "envoy-gateway-system/secrets/envoy-rate-limit",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/secrets/sh.helm.release.v1.envoy-gateway.v1",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/serviceaccounts/default",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/serviceaccounts/envoy-gateway",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/serviceaccounts/envoy-gateway-gateway-helm-certgen",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "envoy-gateway-system/services/envoy-gateway",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/configmaps/ingress-nginx-controller",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/configmaps/ingress-nginx-tcp",
                                "ingress-nginx/configmaps/ingress-nginx-udp",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/configmaps/kube-root-ca.crt",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-node-lease/configmaps/kube-root-ca.crt",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/daemonsets/ingress-nginx-controller",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/deployment/ingress-nginx-defaultbackend",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/endpoints/ingress-nginx-controller",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/endpoints/ingress-nginx-controller-admission",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/endpoints/ingress-nginx-controller-metrics",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/endpoints/ingress-nginx-defaultbackend",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/pods/ingress-nginx-controller-dmjt4",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/pods/ingress-nginx-defaultbackend-6987ff55cf-plsk9",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/rolebindings/ingress-nginx",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/roles/ingress-nginx",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/serviceaccounts/default",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/serviceaccounts/ingress-nginx",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/serviceaccounts/ingress-nginx-backend",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/secrets/ingress-nginx-admission",
                                "ingress-nginx/secrets/sh.helm.release.v1.ingress-nginx.v1",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/services/ingress-nginx-controller",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/services/ingress-nginx-controller-admission",
                                "ingress-nginx/services/ingress-nginx-controller-metrics",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "ingress-nginx/services/ingress-nginx-defaultbackend",
                                "kube-node-lease/serviceaccounts/default",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-system/configmaps/cilium-config",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-system/configmaps/coredns",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-system/configmaps/extension-apiserver-authentication",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-system/configmaps/kube-apiserver-legacy-service-account-token-tracking",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-system/configmaps/kube-proxy",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-system/configmaps/kube-root-ca.crt",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-public/configmaps/cluster-info",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-system/configmaps/kubeadm-config",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "error: the server doesn't have a resource type \"podsecuritypolicies\"",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-public/configmaps/kube-root-ca.crt",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-system/configmaps/kubelet-config",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-system/daemonsets/cilium",
                                "kube-system/daemonsets/kube-proxy",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "xargs: warning: options --max-lines and --replace/-I/-i are mutually exclusive, ignoring previous --max-lines value",
                                "kube-system/deployment/cilium-operator",
                                "kube-system/deployment/coredns",
                                "kube-system/endpoints/kube-dns",
                                "kube-system/endpoints/kube-prometheus-stack-coredns",
                                "kube-system/endpoints/kube-prometheus-stack-kube-controller-manager",
                                "kube-system/endpoints/kube-prometheus-stack-kube-etcd",
                                "kube-system/endpoints/kube-prometheus-stack-kube-proxy",
                                "kube-system/endpoints/kube-prometheus-stack-kube-scheduler",
                                "kube-system/endpoints/kube-prometheus-stack-kubelet",
                                "kube-system/pods/cilium-operator-869df985b8-nqvtb",
                                "kube-system/pods/cilium-sk6n5",
                                "kube-system/pods/coredns-67659f764b-mdrt6",
                                "kube-system/pods/coredns-67659f764b-xb6lv",
                                "kube-system/pods/etcd-instance",
                                "kube-system/pods/kube-apiserver-instance",
                                "kube-system/pods/kube-controller-manager-instance",
                                "kube-system/pods/kube-proxy-cxlxv",
                                "kube-system/pods/kube-scheduler-instance",
                                "kube-system/pods/kube-vip-instance",
                                "kube-public/rolebindings/kubeadm:bootstrap-signer-clusterinfo",
                                "kube-public/rolebindings/system:controller:bootstrap-signer",
                                "kube-public/roles/kubeadm:bootstrap-signer-clusterinfo",
                                "kube-public/roles/system:controller:bootstrap-signer",
                                "kube-system/rolebindings/cilium-config-agent",
                                "kube-system/rolebindings/kube-proxy",
                                "kube-system/rolebindings/kubeadm:kubeadm-certs",
                                "kube-system/rolebindings/kubeadm:kubelet-config",
                                "kube-system/rolebindings/kubeadm:nodes-kubeadm-config",
                                "kube-system/rolebindings/system::extension-apiserver-authentication-reader",
                                "kube-system/rolebindings/system::leader-locking-kube-controller-manager",
                                "kube-system/rolebindings/system::leader-locking-kube-scheduler",
                                "kube-system/rolebindings/system:controller:bootstrap-signer",
                                "kube-system/rolebindings/system:controller:cloud-provider",
                                "kube-system/rolebindings/system:controller:token-cleaner",
                                "kube-public/serviceaccounts/default",
                                "kube-system/roles/cilium-config-agent",
                                "kube-system/roles/extension-apiserver-authentication-reader",
                                "kube-system/roles/kube-proxy",
                                "kube-system/roles/kubeadm:kubeadm-certs",
                                "kube-system/roles/kubeadm:kubelet-config",
                                "kube-system/roles/kubeadm:nodes-kubeadm-config",
                                "kube-system/roles/system::leader-locking-kube-controller-manager",
                                "kube-system/roles/system::leader-locking-kube-scheduler",
                                "kube-system/roles/system:controller:bootstrap-signer",
                                "kube-system/roles/system:controller:cloud-provider",
                                "kube-system/roles/system:controller:token-cleaner",
                                "kube-system/secrets/bootstrap-token-3n8qak",
                                "kube-system/secrets/bootstrap-token-w4o225",
                                "kube-system/secrets/kubeadm-certs",
                                "kube-system/secrets/sh.helm.release.v1.cilium.v1",
                                "kube-system/serviceaccounts/attachdetach-controller",
                                "kube-system/serviceaccounts/bootstrap-signer",
                                "kube-system/serviceaccounts/certificate-controller",
                                "kube-system/serviceaccounts/cilium",
                                "kube-system/serviceaccounts/cilium-operator",
                                "kube-system/serviceaccounts/clusterrole-aggregation-controller",
                                "kube-system/serviceaccounts/coredns",
                                "kube-system/serviceaccounts/cronjob-controller",
                                "kube-system/serviceaccounts/daemon-set-controller",
                                "kube-system/serviceaccounts/default",
                                "kube-system/serviceaccounts/deployment-controller",
                                "kube-system/serviceaccounts/disruption-controller",
                                "kube-system/serviceaccounts/endpoint-controller",
                                "kube-system/serviceaccounts/endpointslice-controller",
                                "kube-system/serviceaccounts/endpointslicemirroring-controller",
                                "kube-system/serviceaccounts/ephemeral-volume-controller",
                                "kube-system/serviceaccounts/expand-controller",
                                "kube-system/serviceaccounts/generic-garbage-collector",
                                "kube-system/serviceaccounts/horizontal-pod-autoscaler",
                                "kube-system/serviceaccounts/job-controller",
                                "kube-system/serviceaccounts/kube-proxy",
                                "kube-system/serviceaccounts/namespace-controller",
                                "kube-system/serviceaccounts/node-controller",
                                "kube-system/serviceaccounts/persistent-volume-binder",
                                "kube-system/serviceaccounts/pod-garbage-collector",
                                "kube-system/serviceaccounts/pv-protection-controller",
                                "kube-system/serviceaccounts/pvc-protection-controller",
                                "kube-system/serviceaccounts/replicaset-controller",
                                "kube-system/serviceaccounts/replication-controller",
                                "kube-system/serviceaccounts/resourcequota-controller",
                                "kube-system/serviceaccounts/root-ca-cert-publisher",
                                "kube-system/services/kube-dns",
                                "kube-system/serviceaccounts/service-account-controller",
                                "kube-system/services/kube-prometheus-stack-coredns",
                                "kube-system/serviceaccounts/service-controller",
                                "kube-system/services/kube-prometheus-stack-kube-controller-manager",
                                "kube-system/serviceaccounts/statefulset-controller",
                                "kube-system/services/kube-prometheus-stack-kube-etcd",
                                "kube-system/serviceaccounts/token-cleaner",
                                "kube-system/services/kube-prometheus-stack-kube-proxy",
                                "kube-system/services/kube-prometheus-stack-kube-scheduler",
                                "kube-system/serviceaccounts/ttl-after-finished-controller",
                                "kube-system/services/kube-prometheus-stack-kubelet",
                                "kube-system/serviceaccounts/ttl-controller",
                                "local-path-storage/configmaps/kube-root-ca.crt",
                                "local-path-storage/configmaps/local-path-config",
                                "local-path-storage/deployment/local-path-provisioner",
                                "monitoring/configmaps/goldpinger-zap",
                                "monitoring/configmaps/ipmi-exporter",
                                "monitoring/configmaps/kube-prometheus-stack-alertmanager-overview",
                                "monitoring/configmaps/kube-prometheus-stack-apiserver",
                                "monitoring/configmaps/kube-prometheus-stack-cluster-total",
                                "monitoring/configmaps/kube-prometheus-stack-controller-manager",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster-advanced",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-goldpinger",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-haproxy",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-host-details",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-hosts-overview",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-node-exporter-full",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-osd-device-details",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-osds-overview",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-pool-detail",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-pool-overview",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-rbd-details",
                                "monitoring/configmaps/kube-prometheus-stack-dashboard-rbd-overview",
                                "monitoring/configmaps/kube-prometheus-stack-etcd",
                                "monitoring/configmaps/kube-prometheus-stack-grafana",
                                "monitoring/configmaps/kube-prometheus-stack-grafana-config-dashboards",
                                "monitoring/configmaps/kube-prometheus-stack-grafana-datasource",
                                "monitoring/configmaps/kube-prometheus-stack-grafana-overview",
                                "monitoring/configmaps/kube-prometheus-stack-k8s-coredns",
                                "monitoring/configmaps/kube-prometheus-stack-k8s-resources-cluster",
                                "monitoring/configmaps/kube-prometheus-stack-k8s-resources-multicluster",
                                "monitoring/configmaps/kube-prometheus-stack-k8s-resources-namespace",
                                "monitoring/configmaps/kube-prometheus-stack-k8s-resources-node",
                                "monitoring/configmaps/kube-prometheus-stack-k8s-resources-pod",
                                "monitoring/configmaps/kube-prometheus-stack-k8s-resources-workload",
                                "monitoring/configmaps/kube-prometheus-stack-k8s-resources-workloads-namespace",
                                "monitoring/configmaps/kube-prometheus-stack-kubelet",
                                "monitoring/configmaps/kube-prometheus-stack-namespace-by-pod",
                                "monitoring/configmaps/kube-prometheus-stack-namespace-by-workload",
                                "monitoring/configmaps/kube-prometheus-stack-node-cluster-rsrc-use",
                                "monitoring/configmaps/kube-prometheus-stack-node-exporter",
                                "monitoring/daemonsets/goldpinger",
                                "monitoring/configmaps/kube-prometheus-stack-node-rsrc-use",
                                "monitoring/daemonsets/ipmi-exporter",
                                "monitoring/configmaps/kube-prometheus-stack-nodes",
                                "monitoring/daemonsets/kube-prometheus-stack-prometheus-node-exporter",
                                "monitoring/configmaps/kube-prometheus-stack-nodes-darwin",
                                "monitoring/daemonsets/node-feature-discovery-worker",
                                "monitoring/configmaps/kube-prometheus-stack-persistentvolumesusage",
                                "monitoring/daemonsets/vector",
                                "monitoring/configmaps/kube-prometheus-stack-pod-total",
                                "monitoring/configmaps/kube-prometheus-stack-prometheus",
                                "monitoring/configmaps/kube-prometheus-stack-prometheus-tls",
                                "monitoring/configmaps/kube-prometheus-stack-proxy",
                                "monitoring/configmaps/kube-prometheus-stack-scheduler",
                                "monitoring/configmaps/kube-prometheus-stack-workload-total",
                                "monitoring/configmaps/kube-root-ca.crt",
                                "monitoring/configmaps/loki",
                                "monitoring/configmaps/loki-alerting-rules",
                                "monitoring/configmaps/loki-dashboards-1",
                                "monitoring/configmaps/loki-dashboards-2",
                                "monitoring/configmaps/loki-gateway",
                                "monitoring/configmaps/loki-runtime",
                                "monitoring/configmaps/node-feature-discovery-worker-conf",
                                "monitoring/configmaps/prometheus-kube-prometheus-stack-prometheus-rulefiles-0",
                                "monitoring/configmaps/vector",
                                "monitoring/deployment/kube-prometheus-stack-grafana",
                                "monitoring/deployment/kube-prometheus-stack-kube-state-metrics",
                                "monitoring/deployment/kube-prometheus-stack-operator",
                                "monitoring/deployment/loki-gateway",
                                "monitoring/deployment/node-feature-discovery-master",
                                "monitoring/deployment/prometheus-pushgateway",
                                "monitoring/endpoints/alertmanager-operated",
                                "monitoring/endpoints/goldpinger",
                                "local-path-storage/pods/local-path-provisioner-679c578f5-m52kp",
                                "monitoring/endpoints/kube-prometheus-stack-alertmanager",
                                "monitoring/endpoints/kube-prometheus-stack-grafana",
                                "monitoring/endpoints/kube-prometheus-stack-kube-state-metrics",
                                "monitoring/endpoints/kube-prometheus-stack-operator",
                                "monitoring/endpoints/kube-prometheus-stack-prometheus",
                                "monitoring/endpoints/kube-prometheus-stack-prometheus-node-exporter",
                                "monitoring/endpoints/loki",
                                "monitoring/endpoints/loki-gateway",
                                "monitoring/endpoints/loki-headless",
                                "monitoring/endpoints/loki-memberlist",
                                "monitoring/endpoints/node-feature-discovery-master",
                                "monitoring/endpoints/prometheus-operated",
                                "monitoring/endpoints/prometheus-pushgateway",
                                "monitoring/endpoints/vector-headless",
                                "monitoring/ingresses/kube-prometheus-stack-alertmanager",
                                "monitoring/ingresses/kube-prometheus-stack-grafana",
                                "monitoring/ingresses/kube-prometheus-stack-prometheus",
                                "local-path-storage/secrets/sh.helm.release.v1.local-path-provisioner.v1",
                                "local-path-storage/serviceaccounts/default",
                                "local-path-storage/serviceaccounts/local-path-provisioner",
                                "monitoring/pods/alertmanager-kube-prometheus-stack-alertmanager-0",
                                "monitoring/pods/goldpinger-gz8xd",
                                "monitoring/pods/kube-prometheus-stack-grafana-69fc56495c-5h4gd",
                                "monitoring/pods/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj",
                                "monitoring/pods/kube-prometheus-stack-operator-cc88b86d6-xspgk",
                                "monitoring/pods/kube-prometheus-stack-prometheus-node-exporter-jmj6n",
                                "monitoring/pods/loki-0",
                                "monitoring/pods/loki-gateway-5cc45946cd-wp7ks",
                                "monitoring/pods/node-feature-discovery-master-68cf466d5f-qknzk",
                                "monitoring/pods/node-feature-discovery-worker-7l5z7",
                                "monitoring/pods/prometheus-kube-prometheus-stack-prometheus-0",
                                "monitoring/pods/prometheus-pushgateway-5c694d99cc-jnjt7",
                                "monitoring/pods/vector-75jgf",
                                "monitoring/persistentvolumeclaims/alertmanager-kube-prometheus-stack-alertmanager-db-alertmanager-kube-prometheus-stack-alertmanager-0",
                                "monitoring/persistentvolumeclaims/prometheus-kube-prometheus-stack-prometheus-db-prometheus-kube-prometheus-stack-prometheus-0",
                                "monitoring/persistentvolumeclaims/storage-loki-0",
                                "monitoring/rolebindings/kube-prometheus-stack-grafana",
                                "monitoring/rolebindings/kube-prometheus-stack-pod-tls-sidecar",
                                "monitoring/roles/kube-prometheus-stack-grafana",
                                "monitoring/roles/kube-prometheus-stack-pod-tls-sidecar",
                                "monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager",
                                "monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-generated",
                                "monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-tls-assets-0",
                                "monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-web-config",
                                "monitoring/secrets/alertmanager-tls",
                                "monitoring/secrets/grafana-tls",
                                "monitoring/secrets/kube-prometheus-stack-admission",
                                "monitoring/secrets/kube-prometheus-stack-alertmanager-client-secret",
                                "monitoring/secrets/kube-prometheus-stack-alertmanager-cookie-secret",
                                "monitoring/secrets/kube-prometheus-stack-alertmanager-oauth2-proxy",
                                "monitoring/secrets/kube-prometheus-stack-etcd-client-cert",
                                "monitoring/secrets/kube-prometheus-stack-grafana",
                                "monitoring/secrets/kube-prometheus-stack-grafana-client-secret",
                                "monitoring/secrets/kube-prometheus-stack-prometheus-client-secret",
                                "monitoring/secrets/kube-prometheus-stack-prometheus-cookie-secret",
                                "monitoring/secrets/kube-prometheus-stack-prometheus-node-exporter-jmj6n-tls",
                                "monitoring/secrets/kube-prometheus-stack-prometheus-oauth2-proxy",
                                "monitoring/secrets/prometheus-kube-prometheus-stack-prometheus",
                                "monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-0-tls",
                                "monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-tls-assets-0",
                                "monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-web-config",
                                "monitoring/secrets/prometheus-tls",
                                "monitoring/secrets/sh.helm.release.v1.goldpinger.v1",
                                "monitoring/secrets/sh.helm.release.v1.kube-prometheus-stack.v1",
                                "monitoring/secrets/sh.helm.release.v1.loki.v1",
                                "monitoring/secrets/sh.helm.release.v1.node-feature-discovery.v1",
                                "monitoring/secrets/sh.helm.release.v1.prometheus-pushgateway.v1",
                                "monitoring/secrets/sh.helm.release.v1.vector.v1",
                                "openstack/configmaps/keepalived-bin",
                                "openstack/configmaps/kube-root-ca.crt",
                                "openstack/configmaps/memcached-memcached-bin",
                                "openstack/configmaps/messaging-topology-operator-leader-election",
                                "openstack/configmaps/percona-xtradb-haproxy",
                                "openstack/configmaps/percona-xtradb-pxc",
                                "openstack/configmaps/rabbitmq-cluster-operator-leader-election",
                                "monitoring/serviceaccounts/default",
                                "monitoring/serviceaccounts/goldpinger",
                                "monitoring/serviceaccounts/kube-prometheus-stack-alertmanager",
                                "monitoring/serviceaccounts/kube-prometheus-stack-grafana",
                                "monitoring/serviceaccounts/kube-prometheus-stack-kube-state-metrics",
                                "monitoring/serviceaccounts/kube-prometheus-stack-operator",
                                "monitoring/serviceaccounts/kube-prometheus-stack-prometheus",
                                "monitoring/serviceaccounts/kube-prometheus-stack-prometheus-node-exporter",
                                "monitoring/serviceaccounts/loki",
                                "monitoring/serviceaccounts/node-feature-discovery",
                                "monitoring/serviceaccounts/node-feature-discovery-worker",
                                "monitoring/serviceaccounts/prometheus-pushgateway",
                                "monitoring/serviceaccounts/vector",
                                "monitoring/services/alertmanager-operated",
                                "monitoring/services/goldpinger",
                                "monitoring/services/kube-prometheus-stack-alertmanager",
                                "monitoring/services/kube-prometheus-stack-grafana",
                                "monitoring/services/kube-prometheus-stack-kube-state-metrics",
                                "monitoring/services/kube-prometheus-stack-operator",
                                "monitoring/services/kube-prometheus-stack-prometheus",
                                "monitoring/services/kube-prometheus-stack-prometheus-node-exporter",
                                "monitoring/services/loki",
                                "monitoring/services/loki-gateway",
                                "monitoring/services/loki-headless",
                                "monitoring/services/loki-memberlist",
                                "monitoring/services/node-feature-discovery-master",
                                "monitoring/services/prometheus-operated",
                                "monitoring/services/prometheus-pushgateway",
                                "monitoring/services/vector-headless",
                                "monitoring/statefulsets/alertmanager-kube-prometheus-stack-alertmanager",
                                "monitoring/statefulsets/loki",
                                "monitoring/statefulsets/prometheus-kube-prometheus-stack-prometheus",
                                "secretgen-controller/configmaps/kube-root-ca.crt",
                                "openstack/daemonsets/keepalived",
                                "openstack/deployment/memcached-memcached",
                                "openstack/deployment/pxc-operator",
                                "openstack/deployment/rabbitmq-cluster-operator",
                                "openstack/deployment/rabbitmq-messaging-topology-operator",
                                "openstack/endpoints/memcached",
                                "openstack/endpoints/memcached-metrics",
                                "openstack/endpoints/percona-xtradb-cluster-operator",
                                "openstack/endpoints/percona-xtradb-haproxy",
                                "openstack/endpoints/percona-xtradb-haproxy-metrics",
                                "openstack/endpoints/percona-xtradb-haproxy-replicas",
                                "openstack/endpoints/percona-xtradb-pxc",
                                "openstack/endpoints/percona-xtradb-pxc-unready",
                                "openstack/endpoints/rabbitmq-messaging-topology-operator-webhook",
                                "secretgen-controller/deployment/secretgen-controller",
                                "openstack/pods/keepalived-wktjf",
                                "openstack/pods/memcached-memcached-c47d949f5-vc5cm",
                                "openstack/pods/percona-xtradb-haproxy-0",
                                "openstack/pods/percona-xtradb-pxc-0",
                                "openstack/pods/pxc-operator-59d66cbc78-clxvf",
                                "openstack/pods/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf",
                                "openstack/pods/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc",
                                "openstack/persistentvolumeclaims/datadir-percona-xtradb-pxc-0",
                                "openstack/rolebindings/keepalived",
                                "openstack/rolebindings/pxc-operator",
                                "openstack/rolebindings/rabbitmq-cluster-operator",
                                "openstack/rolebindings/rabbitmq-messaging-topology-operator",
                                "secretgen-controller/pods/secretgen-controller-5cf976ccc7-vwxz4",
                                "openstack/roles/keepalived",
                                "openstack/roles/pxc-operator",
                                "openstack/roles/rabbitmq-cluster-operator",
                                "openstack/roles/rabbitmq-messaging-topology-operator",
                                "openstack/secrets/internal-percona-xtradb",
                                "openstack/secrets/keepalived-etc",
                                "openstack/secrets/percona-xtradb",
                                "openstack/secrets/rabbitmq-messaging-topology-operator-webhook",
                                "openstack/secrets/sh.helm.release.v1.memcached.v1",
                                "openstack/secrets/sh.helm.release.v1.pxc-operator.v1",
                                "openstack/secrets/sh.helm.release.v1.rabbitmq-cluster-operator.v1",
                                "openstack/serviceaccounts/default",
                                "openstack/serviceaccounts/keepalived",
                                "openstack/serviceaccounts/memcached-memcached",
                                "openstack/serviceaccounts/pxc-operator",
                                "openstack/serviceaccounts/rabbitmq-cluster-operator",
                                "openstack/serviceaccounts/rabbitmq-messaging-topology-operator",
                                "openstack/statefulsets/percona-xtradb-haproxy",
                                "openstack/statefulsets/percona-xtradb-pxc",
                                "openstack/services/memcached",
                                "openstack/services/memcached-metrics",
                                "openstack/services/percona-xtradb-cluster-operator",
                                "openstack/services/percona-xtradb-haproxy",
                                "openstack/services/percona-xtradb-haproxy-metrics",
                                "openstack/services/percona-xtradb-haproxy-replicas",
                                "openstack/services/percona-xtradb-pxc",
                                "openstack/services/percona-xtradb-pxc-unready",
                                "openstack/services/rabbitmq-messaging-topology-operator-webhook",
                                "secretgen-controller/serviceaccounts/default",
                                "secretgen-controller/serviceaccounts/secretgen-controller-sa"
                            ],
                            "zuul_log_id": "0242ac17-0011-cc15-c3f6-00000000001c-1-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000017",
                        "name": "describe-kubernetes-objects",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/describe-kubernetes-objects"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:21.590541Z",
                            "start": "2026-04-04T10:25:29.329115Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-00000000001c",
                        "name": "Gathering descriptions for namespace scoped objects"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "synchronize",
                            "changed": true,
                            "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --rsh='/usr/bin/ssh -S none -o Port=22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null' --rsync-path='sudo -u root rsync' --out-format='<<CHANGED>>%i %n%L' zuul@199.204.45.153:/tmp/logs/objects /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                            "invocation": {
                                "module_args": {
                                    "_local_rsync_password": null,
                                    "_local_rsync_path": "rsync",
                                    "_ssh_args": null,
                                    "_substitute_controller": false,
                                    "archive": true,
                                    "checksum": false,
                                    "compress": true,
                                    "copy_links": false,
                                    "delay_updates": true,
                                    "delete": false,
                                    "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                                    "dest_port": 22,
                                    "dirs": false,
                                    "existing_only": false,
                                    "group": null,
                                    "link_dest": null,
                                    "links": null,
                                    "mode": "pull",
                                    "owner": null,
                                    "partial": false,
                                    "perms": null,
                                    "private_key": null,
                                    "recursive": null,
                                    "rsync_opts": [],
                                    "rsync_path": "sudo -u root rsync",
                                    "rsync_timeout": 0,
                                    "set_remote_user": true,
                                    "src": "zuul@199.204.45.153:/tmp/logs/objects",
                                    "ssh_connection_multiplexing": false,
                                    "times": null,
                                    "use_ssh_args": false,
                                    "verify_host": false
                                }
                            },
                            "msg": "cd+++++++++ objects/\ncd+++++++++ objects/cluster/\ncd+++++++++ objects/cluster/clusterrole/\n>f+++++++++ objects/cluster/clusterrole/admin.txt\n>f+++++++++ objects/cluster/clusterrole/admin.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-cainjector.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-cainjector.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-approve:cert-manager-io.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-approve:cert-manager-io.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-certificates.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-certificates.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-certificatesigningrequests.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-certificatesigningrequests.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-challenges.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-challenges.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-clusterissuers.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-clusterissuers.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-ingress-shim.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-ingress-shim.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-issuers.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-issuers.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-orders.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-controller-orders.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-edit.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-edit.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-view.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-view.yaml\n>f+++++++++ objects/cluster/clusterrole/cert-manager-webhook:subjectaccessreviews.txt\n>f+++++++++ objects/cluster/clusterrole/cert-manager-webhook:subjectaccessreviews.yaml\n>f+++++++++ objects/cluster/clusterrole/cilium-operator.txt\n>f+++++++++ objects/cluster/clusterrole/cilium-operator.yaml\n>f+++++++++ objects/cluster/clusterrole/cilium.txt\n>f+++++++++ objects/cluster/clusterrole/cilium.yaml\n>f+++++++++ objects/cluster/clusterrole/cluster-admin.txt\n>f+++++++++ objects/cluster/clusterrole/cluster-admin.yaml\n>f+++++++++ objects/cluster/clusterrole/edit.txt\n>f+++++++++ objects/cluster/clusterrole/edit.yaml\n>f+++++++++ objects/cluster/clusterrole/envoy-gateway-gateway-helm-certgen:envoy-gateway-system.txt\n>f+++++++++ objects/cluster/clusterrole/envoy-gateway-gateway-helm-certgen:envoy-gateway-system.yaml\n>f+++++++++ objects/cluster/clusterrole/envoy-gateway-gateway-helm-envoy-gateway-role.txt\n>f+++++++++ objects/cluster/clusterrole/envoy-gateway-gateway-helm-envoy-gateway-role.yaml\n>f+++++++++ objects/cluster/clusterrole/goldpinger-clusterrole.txt\n>f+++++++++ objects/cluster/clusterrole/goldpinger-clusterrole.yaml\n>f+++++++++ objects/cluster/clusterrole/ingress-nginx.txt\n>f+++++++++ objects/cluster/clusterrole/ingress-nginx.yaml\n>f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-grafana-clusterrole.txt\n>f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-grafana-clusterrole.yaml\n>f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-kube-state-metrics.txt\n>f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-kube-state-metrics.yaml\n>f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-operator.txt\n>f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-operator.yaml\n>f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-prometheus.txt\n>f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-prometheus.yaml\n>f+++++++++ objects/cluster/clusterrole/kubeadm:get-nodes.txt\n>f+++++++++ objects/cluster/clusterrole/kubeadm:get-nodes.yaml\n>f+++++++++ objects/cluster/clusterrole/local-path-provisioner.txt\n>f+++++++++ objects/cluster/clusterrole/local-path-provisioner.yaml\n>f+++++++++ objects/cluster/clusterrole/loki-clusterrole.txt\n>f+++++++++ objects/cluster/clusterrole/loki-clusterrole.yaml\n>f+++++++++ objects/cluster/clusterrole/node-feature-discovery.txt\n>f+++++++++ objects/cluster/clusterrole/node-feature-discovery.yaml\n>f+++++++++ objects/cluster/clusterrole/rabbitmq-cluster-operator.txt\n>f+++++++++ objects/cluster/clusterrole/rabbitmq-cluster-operator.yaml\n>f+++++++++ objects/cluster/clusterrole/rabbitmq-messaging-topology-operator.txt\n>f+++++++++ objects/cluster/clusterrole/rabbitmq-messaging-topology-operator.yaml\n>f+++++++++ objects/cluster/clusterrole/secretgen-controller-cluster-role.txt\n>f+++++++++ objects/cluster/clusterrole/secretgen-controller-cluster-role.yaml\n>f+++++++++ objects/cluster/clusterrole/system:aggregate-to-admin.txt\n>f+++++++++ objects/cluster/clusterrole/system:aggregate-to-admin.yaml\n>f+++++++++ objects/cluster/clusterrole/system:aggregate-to-edit.txt\n>f+++++++++ objects/cluster/clusterrole/system:aggregate-to-edit.yaml\n>f+++++++++ objects/cluster/clusterrole/system:aggregate-to-view.txt\n>f+++++++++ objects/cluster/clusterrole/system:aggregate-to-view.yaml\n>f+++++++++ objects/cluster/clusterrole/system:auth-delegator.txt\n>f+++++++++ objects/cluster/clusterrole/system:auth-delegator.yaml\n>f+++++++++ objects/cluster/clusterrole/system:basic-user.txt\n>f+++++++++ objects/cluster/clusterrole/system:basic-user.yaml\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:certificatesigningrequests:nodeclient.txt\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:certificatesigningrequests:nodeclient.yaml\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:certificatesigningrequests:selfnodeclient.txt\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:certificatesigningrequests:selfnodeclient.yaml\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kube-apiserver-client-approver.txt\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kube-apiserver-client-approver.yaml\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kube-apiserver-client-kubelet-approver.txt\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kube-apiserver-client-kubelet-approver.yaml\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kubelet-serving-approver.txt\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kubelet-serving-approver.yaml\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:legacy-unknown-approver.txt\n>f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:legacy-unknown-approver.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:attachdetach-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:attachdetach-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:certificate-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:certificate-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:clusterrole-aggregation-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:clusterrole-aggregation-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:cronjob-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:cronjob-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:daemon-set-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:daemon-set-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:deployment-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:deployment-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:disruption-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:disruption-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:endpoint-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:endpoint-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:endpointslice-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:endpointslice-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:endpointslicemirroring-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:endpointslicemirroring-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:ephemeral-volume-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:ephemeral-volume-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:expand-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:expand-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:generic-garbage-collector.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:generic-garbage-collector.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:horizontal-pod-autoscaler.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:horizontal-pod-autoscaler.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:job-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:job-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:namespace-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:namespace-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:node-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:node-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:persistent-volume-binder.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:persistent-volume-binder.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:pod-garbage-collector.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:pod-garbage-collector.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:pv-protection-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:pv-protection-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:pvc-protection-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:pvc-protection-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:replicaset-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:replicaset-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:replication-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:replication-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:resourcequota-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:resourcequota-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:root-ca-cert-publisher.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:root-ca-cert-publisher.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:route-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:route-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:service-account-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:service-account-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:service-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:service-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:statefulset-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:statefulset-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:ttl-after-finished-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:ttl-after-finished-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:controller:ttl-controller.txt\n>f+++++++++ objects/cluster/clusterrole/system:controller:ttl-controller.yaml\n>f+++++++++ objects/cluster/clusterrole/system:coredns.txt\n>f+++++++++ objects/cluster/clusterrole/system:coredns.yaml\n>f+++++++++ objects/cluster/clusterrole/system:discovery.txt\n>f+++++++++ objects/cluster/clusterrole/system:discovery.yaml\n>f+++++++++ objects/cluster/clusterrole/system:heapster.txt\n>f+++++++++ objects/cluster/clusterrole/system:heapster.yaml\n>f+++++++++ objects/cluster/clusterrole/system:kube-aggregator.txt\n>f+++++++++ objects/cluster/clusterrole/system:kube-aggregator.yaml\n>f+++++++++ objects/cluster/clusterrole/system:kube-controller-manager.txt\n>f+++++++++ objects/cluster/clusterrole/system:kube-controller-manager.yaml\n>f+++++++++ objects/cluster/clusterrole/system:kube-dns.txt\n>f+++++++++ objects/cluster/clusterrole/system:kube-dns.yaml\n>f+++++++++ objects/cluster/clusterrole/system:kube-scheduler.txt\n>f+++++++++ objects/cluster/clusterrole/system:kube-scheduler.yaml\n>f+++++++++ objects/cluster/clusterrole/system:kubelet-api-admin.txt\n>f+++++++++ objects/cluster/clusterrole/system:kubelet-api-admin.yaml\n>f+++++++++ objects/cluster/clusterrole/system:monitoring.txt\n>f+++++++++ objects/cluster/clusterrole/system:monitoring.yaml\n>f+++++++++ objects/cluster/clusterrole/system:node-bootstrapper.txt\n>f+++++++++ objects/cluster/clusterrole/system:node-bootstrapper.yaml\n>f+++++++++ objects/cluster/clusterrole/system:node-problem-detector.txt\n>f+++++++++ objects/cluster/clusterrole/system:node-problem-detector.yaml\n>f+++++++++ objects/cluster/clusterrole/system:node-proxier.txt\n>f+++++++++ objects/cluster/clusterrole/system:node-proxier.yaml\n>f+++++++++ objects/cluster/clusterrole/system:node.txt\n>f+++++++++ objects/cluster/clusterrole/system:node.yaml\n>f+++++++++ objects/cluster/clusterrole/system:persistent-volume-provisioner.txt\n>f+++++++++ objects/cluster/clusterrole/system:persistent-volume-provisioner.yaml\n>f+++++++++ objects/cluster/clusterrole/system:public-info-viewer.txt\n>f+++++++++ objects/cluster/clusterrole/system:public-info-viewer.yaml\n>f+++++++++ objects/cluster/clusterrole/system:service-account-issuer-discovery.txt\n>f+++++++++ objects/cluster/clusterrole/system:service-account-issuer-discovery.yaml\n>f+++++++++ objects/cluster/clusterrole/system:volume-scheduler.txt\n>f+++++++++ objects/cluster/clusterrole/system:volume-scheduler.yaml\n>f+++++++++ objects/cluster/clusterrole/vector.txt\n>f+++++++++ objects/cluster/clusterrole/vector.yaml\n>f+++++++++ objects/cluster/clusterrole/view.txt\n>f+++++++++ objects/cluster/clusterrole/view.yaml\ncd+++++++++ objects/cluster/clusterrolebinding/\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-cainjector.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-cainjector.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-approve:cert-manager-io.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-approve:cert-manager-io.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-certificates.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-certificates.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-certificatesigningrequests.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-certificatesigningrequests.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-challenges.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-challenges.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-clusterissuers.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-clusterissuers.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-ingress-shim.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-ingress-shim.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-issuers.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-issuers.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-orders.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-orders.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-webhook:subjectaccessreviews.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cert-manager-webhook:subjectaccessreviews.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cilium-operator.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cilium-operator.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cilium.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cilium.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/cluster-admin.txt\n>f+++++++++ objects/cluster/clusterrolebinding/cluster-admin.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/envoy-gateway-gateway-helm-certgen:envoy-gateway-system.txt\n>f+++++++++ objects/cluster/clusterrolebinding/envoy-gateway-gateway-helm-certgen:envoy-gateway-system.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/envoy-gateway-gateway-helm-envoy-gateway-rolebinding.txt\n>f+++++++++ objects/cluster/clusterrolebinding/envoy-gateway-gateway-helm-envoy-gateway-rolebinding.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/goldpinger-clusterrolebinding.txt\n>f+++++++++ objects/cluster/clusterrolebinding/goldpinger-clusterrolebinding.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/ingress-nginx.txt\n>f+++++++++ objects/cluster/clusterrolebinding/ingress-nginx.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-grafana-clusterrolebinding.txt\n>f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-grafana-clusterrolebinding.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-kube-state-metrics.txt\n>f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-kube-state-metrics.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-operator.txt\n>f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-operator.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-prometheus.txt\n>f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-prometheus.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/kubeadm:get-nodes.txt\n>f+++++++++ objects/cluster/clusterrolebinding/kubeadm:get-nodes.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/kubeadm:kubelet-bootstrap.txt\n>f+++++++++ objects/cluster/clusterrolebinding/kubeadm:kubelet-bootstrap.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-autoapprove-bootstrap.txt\n>f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-autoapprove-bootstrap.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-autoapprove-certificate-rotation.txt\n>f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-autoapprove-certificate-rotation.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-proxier.txt\n>f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-proxier.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/local-path-provisioner.txt\n>f+++++++++ objects/cluster/clusterrolebinding/local-path-provisioner.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/loki-clusterrolebinding.txt\n>f+++++++++ objects/cluster/clusterrolebinding/loki-clusterrolebinding.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/node-feature-discovery.txt\n>f+++++++++ objects/cluster/clusterrolebinding/node-feature-discovery.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/rabbitmq-cluster-operator.txt\n>f+++++++++ objects/cluster/clusterrolebinding/rabbitmq-cluster-operator.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/rabbitmq-messaging-topology-operator.txt\n>f+++++++++ objects/cluster/clusterrolebinding/rabbitmq-messaging-topology-operator.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/secretgen-controller-cluster-role-binding.txt\n>f+++++++++ objects/cluster/clusterrolebinding/secretgen-controller-cluster-role-binding.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:basic-user.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:basic-user.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:attachdetach-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:attachdetach-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:certificate-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:certificate-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:clusterrole-aggregation-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:clusterrole-aggregation-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:cronjob-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:cronjob-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:daemon-set-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:daemon-set-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:deployment-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:deployment-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:disruption-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:disruption-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpoint-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpoint-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpointslice-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpointslice-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpointslicemirroring-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpointslicemirroring-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:ephemeral-volume-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:ephemeral-volume-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:expand-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:expand-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:generic-garbage-collector.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:generic-garbage-collector.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:horizontal-pod-autoscaler.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:horizontal-pod-autoscaler.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:job-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:job-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:namespace-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:namespace-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:node-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:node-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:persistent-volume-binder.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:persistent-volume-binder.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:pod-garbage-collector.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:pod-garbage-collector.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:pv-protection-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:pv-protection-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:pvc-protection-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:pvc-protection-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:replicaset-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:replicaset-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:replication-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:replication-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:resourcequota-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:resourcequota-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:root-ca-cert-publisher.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:root-ca-cert-publisher.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:route-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:route-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:service-account-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:service-account-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:service-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:service-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:statefulset-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:statefulset-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:ttl-after-finished-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:ttl-after-finished-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:ttl-controller.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:controller:ttl-controller.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:coredns.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:coredns.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:discovery.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:discovery.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:kube-controller-manager.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:kube-controller-manager.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:kube-dns.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:kube-dns.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:kube-scheduler.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:kube-scheduler.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:monitoring.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:monitoring.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:node-proxier.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:node-proxier.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:node.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:node.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:public-info-viewer.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:public-info-viewer.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:service-account-issuer-discovery.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:service-account-issuer-discovery.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/system:volume-scheduler.txt\n>f+++++++++ objects/cluster/clusterrolebinding/system:volume-scheduler.yaml\n>f+++++++++ objects/cluster/clusterrolebinding/vector.txt\n>f+++++++++ objects/cluster/clusterrolebinding/vector.yaml\ncd+++++++++ objects/cluster/namespace/\n>f+++++++++ objects/cluster/namespace/auth-system.txt\n>f+++++++++ objects/cluster/namespace/auth-system.yaml\n>f+++++++++ objects/cluster/namespace/cert-manager.txt\n>f+++++++++ objects/cluster/namespace/cert-manager.yaml\n>f+++++++++ objects/cluster/namespace/default.txt\n>f+++++++++ objects/cluster/namespace/default.yaml\n>f+++++++++ objects/cluster/namespace/envoy-gateway-system.txt\n>f+++++++++ objects/cluster/namespace/envoy-gateway-system.yaml\n>f+++++++++ objects/cluster/namespace/ingress-nginx.txt\n>f+++++++++ objects/cluster/namespace/ingress-nginx.yaml\n>f+++++++++ objects/cluster/namespace/kube-node-lease.txt\n>f+++++++++ objects/cluster/namespace/kube-node-lease.yaml\n>f+++++++++ objects/cluster/namespace/kube-public.txt\n>f+++++++++ objects/cluster/namespace/kube-public.yaml\n>f+++++++++ objects/cluster/namespace/kube-system.txt\n>f+++++++++ objects/cluster/namespace/kube-system.yaml\n>f+++++++++ objects/cluster/namespace/local-path-storage.txt\n>f+++++++++ objects/cluster/namespace/local-path-storage.yaml\n>f+++++++++ objects/cluster/namespace/monitoring.txt\n>f+++++++++ objects/cluster/namespace/monitoring.yaml\n>f+++++++++ objects/cluster/namespace/openstack.txt\n>f+++++++++ objects/cluster/namespace/openstack.yaml\n>f+++++++++ objects/cluster/namespace/secretgen-controller.txt\n>f+++++++++ objects/cluster/namespace/secretgen-controller.yaml\ncd+++++++++ objects/cluster/node/\n>f+++++++++ objects/cluster/node/instance.txt\n>f+++++++++ objects/cluster/node/instance.yaml\ncd+++++++++ objects/cluster/storageclass/\n>f+++++++++ objects/cluster/storageclass/general.txt\n>f+++++++++ objects/cluster/storageclass/general.yaml\ncd+++++++++ objects/namespaced/\ncd+++++++++ objects/namespaced/auth-system/\ncd+++++++++ objects/namespaced/auth-system/configmaps/\n>f+++++++++ objects/namespaced/auth-system/configmaps/keycloak-env-vars.txt\n>f+++++++++ objects/namespaced/auth-system/configmaps/keycloak-env-vars.yaml\n>f+++++++++ objects/namespaced/auth-system/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/auth-system/configmaps/kube-root-ca.crt.yaml\ncd+++++++++ objects/namespaced/auth-system/endpoints/\n>f+++++++++ objects/namespaced/auth-system/endpoints/keycloak-headless.txt\n>f+++++++++ objects/namespaced/auth-system/endpoints/keycloak-headless.yaml\n>f+++++++++ objects/namespaced/auth-system/endpoints/keycloak-metrics.txt\n>f+++++++++ objects/namespaced/auth-system/endpoints/keycloak-metrics.yaml\n>f+++++++++ objects/namespaced/auth-system/endpoints/keycloak.txt\n>f+++++++++ objects/namespaced/auth-system/endpoints/keycloak.yaml\ncd+++++++++ objects/namespaced/auth-system/ingresses/\n>f+++++++++ objects/namespaced/auth-system/ingresses/keycloak.txt\n>f+++++++++ objects/namespaced/auth-system/ingresses/keycloak.yaml\ncd+++++++++ objects/namespaced/auth-system/pods/\n>f+++++++++ objects/namespaced/auth-system/pods/keycloak-0.txt\n>f+++++++++ objects/namespaced/auth-system/pods/keycloak-0.yaml\ncd+++++++++ objects/namespaced/auth-system/secrets/\n>f+++++++++ objects/namespaced/auth-system/secrets/keycloak-externaldb.txt\n>f+++++++++ objects/namespaced/auth-system/secrets/keycloak-externaldb.yaml\n>f+++++++++ objects/namespaced/auth-system/secrets/keycloak.199-204-45-153.nip.io-tls.txt\n>f+++++++++ objects/namespaced/auth-system/secrets/keycloak.199-204-45-153.nip.io-tls.yaml\n>f+++++++++ objects/namespaced/auth-system/secrets/keycloak.txt\n>f+++++++++ objects/namespaced/auth-system/secrets/keycloak.yaml\n>f+++++++++ objects/namespaced/auth-system/secrets/sh.helm.release.v1.keycloak.v1.txt\n>f+++++++++ objects/namespaced/auth-system/secrets/sh.helm.release.v1.keycloak.v1.yaml\ncd+++++++++ objects/namespaced/auth-system/serviceaccounts/\n>f+++++++++ objects/namespaced/auth-system/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/auth-system/serviceaccounts/default.yaml\n>f+++++++++ objects/namespaced/auth-system/serviceaccounts/keycloak.txt\n>f+++++++++ objects/namespaced/auth-system/serviceaccounts/keycloak.yaml\ncd+++++++++ objects/namespaced/auth-system/services/\n>f+++++++++ objects/namespaced/auth-system/services/keycloak-headless.txt\n>f+++++++++ objects/namespaced/auth-system/services/keycloak-headless.yaml\n>f+++++++++ objects/namespaced/auth-system/services/keycloak-metrics.txt\n>f+++++++++ objects/namespaced/auth-system/services/keycloak-metrics.yaml\n>f+++++++++ objects/namespaced/auth-system/services/keycloak.txt\n>f+++++++++ objects/namespaced/auth-system/services/keycloak.yaml\ncd+++++++++ objects/namespaced/auth-system/statefulsets/\n>f+++++++++ objects/namespaced/auth-system/statefulsets/keycloak.txt\n>f+++++++++ objects/namespaced/auth-system/statefulsets/keycloak.yaml\ncd+++++++++ objects/namespaced/cert-manager/\ncd+++++++++ objects/namespaced/cert-manager/configmaps/\n>f+++++++++ objects/namespaced/cert-manager/configmaps/cert-manager-webhook.txt\n>f+++++++++ objects/namespaced/cert-manager/configmaps/cert-manager-webhook.yaml\n>f+++++++++ objects/namespaced/cert-manager/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/cert-manager/configmaps/kube-root-ca.crt.yaml\ncd+++++++++ objects/namespaced/cert-manager/deployment/\n>f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager-cainjector.txt\n>f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager-cainjector.yaml\n>f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager-webhook.txt\n>f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager-webhook.yaml\n>f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager.txt\n>f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager.yaml\ncd+++++++++ objects/namespaced/cert-manager/endpoints/\n>f+++++++++ objects/namespaced/cert-manager/endpoints/cert-manager-webhook.txt\n>f+++++++++ objects/namespaced/cert-manager/endpoints/cert-manager-webhook.yaml\n>f+++++++++ objects/namespaced/cert-manager/endpoints/cert-manager.txt\n>f+++++++++ objects/namespaced/cert-manager/endpoints/cert-manager.yaml\ncd+++++++++ objects/namespaced/cert-manager/pods/\n>f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-75969b45cf-d5c4v.txt\n>f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-75969b45cf-d5c4v.yaml\n>f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-cainjector-84f45b698d-v64n5.txt\n>f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-cainjector-84f45b698d-v64n5.yaml\n>f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-webhook-9997c64fc-8w8sz.txt\n>f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-webhook-9997c64fc-8w8sz.yaml\ncd+++++++++ objects/namespaced/cert-manager/rolebindings/\n>f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager-cainjector:leaderelection.txt\n>f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager-cainjector:leaderelection.yaml\n>f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager-webhook:dynamic-serving.txt\n>f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager-webhook:dynamic-serving.yaml\n>f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager:leaderelection.txt\n>f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager:leaderelection.yaml\ncd+++++++++ objects/namespaced/cert-manager/roles/\n>f+++++++++ objects/namespaced/cert-manager/roles/cert-manager-cainjector:leaderelection.txt\n>f+++++++++ objects/namespaced/cert-manager/roles/cert-manager-cainjector:leaderelection.yaml\n>f+++++++++ objects/namespaced/cert-manager/roles/cert-manager-webhook:dynamic-serving.txt\n>f+++++++++ objects/namespaced/cert-manager/roles/cert-manager-webhook:dynamic-serving.yaml\n>f+++++++++ objects/namespaced/cert-manager/roles/cert-manager:leaderelection.txt\n>f+++++++++ objects/namespaced/cert-manager/roles/cert-manager:leaderelection.yaml\ncd+++++++++ objects/namespaced/cert-manager/secrets/\n>f+++++++++ objects/namespaced/cert-manager/secrets/cert-manager-selfsigned-ca.txt\n>f+++++++++ objects/namespaced/cert-manager/secrets/cert-manager-selfsigned-ca.yaml\n>f+++++++++ objects/namespaced/cert-manager/secrets/cert-manager-webhook-ca.txt\n>f+++++++++ objects/namespaced/cert-manager/secrets/cert-manager-webhook-ca.yaml\n>f+++++++++ objects/namespaced/cert-manager/secrets/kube-prometheus-stack-ca.txt\n>f+++++++++ objects/namespaced/cert-manager/secrets/kube-prometheus-stack-ca.yaml\n>f+++++++++ objects/namespaced/cert-manager/secrets/sh.helm.release.v1.cert-manager.v1.txt\n>f+++++++++ objects/namespaced/cert-manager/secrets/sh.helm.release.v1.cert-manager.v1.yaml\ncd+++++++++ objects/namespaced/cert-manager/serviceaccounts/\n>f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager-cainjector.txt\n>f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager-cainjector.yaml\n>f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager-webhook.txt\n>f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager-webhook.yaml\n>f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager.txt\n>f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager.yaml\n>f+++++++++ objects/namespaced/cert-manager/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/cert-manager/serviceaccounts/default.yaml\ncd+++++++++ objects/namespaced/cert-manager/services/\n>f+++++++++ objects/namespaced/cert-manager/services/cert-manager-webhook.txt\n>f+++++++++ objects/namespaced/cert-manager/services/cert-manager-webhook.yaml\n>f+++++++++ objects/namespaced/cert-manager/services/cert-manager.txt\n>f+++++++++ objects/namespaced/cert-manager/services/cert-manager.yaml\ncd+++++++++ objects/namespaced/default/\ncd+++++++++ objects/namespaced/default/configmaps/\n>f+++++++++ objects/namespaced/default/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/default/configmaps/kube-root-ca.crt.yaml\ncd+++++++++ objects/namespaced/default/endpoints/\n>f+++++++++ objects/namespaced/default/endpoints/kubernetes.txt\n>f+++++++++ objects/namespaced/default/endpoints/kubernetes.yaml\ncd+++++++++ objects/namespaced/default/serviceaccounts/\n>f+++++++++ objects/namespaced/default/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/default/serviceaccounts/default.yaml\ncd+++++++++ objects/namespaced/default/services/\n>f+++++++++ objects/namespaced/default/services/kubernetes.txt\n>f+++++++++ objects/namespaced/default/services/kubernetes.yaml\ncd+++++++++ objects/namespaced/envoy-gateway-system/\ncd+++++++++ objects/namespaced/envoy-gateway-system/configmaps/\n>f+++++++++ objects/namespaced/envoy-gateway-system/configmaps/envoy-gateway-config.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/configmaps/envoy-gateway-config.yaml\n>f+++++++++ objects/namespaced/envoy-gateway-system/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/configmaps/kube-root-ca.crt.yaml\ncd+++++++++ objects/namespaced/envoy-gateway-system/deployment/\n>f+++++++++ objects/namespaced/envoy-gateway-system/deployment/envoy-gateway.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/deployment/envoy-gateway.yaml\ncd+++++++++ objects/namespaced/envoy-gateway-system/endpoints/\n>f+++++++++ objects/namespaced/envoy-gateway-system/endpoints/envoy-gateway.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/endpoints/envoy-gateway.yaml\ncd+++++++++ objects/namespaced/envoy-gateway-system/pods/\n>f+++++++++ objects/namespaced/envoy-gateway-system/pods/envoy-gateway-78446f96c9-7zvgb.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/pods/envoy-gateway-78446f96c9-7zvgb.yaml\ncd+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/\n>f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-certgen.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-certgen.yaml\n>f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-infra-manager.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-infra-manager.yaml\n>f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-leader-election-rolebinding.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-leader-election-rolebinding.yaml\ncd+++++++++ objects/namespaced/envoy-gateway-system/roles/\n>f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-certgen.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-certgen.yaml\n>f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-infra-manager.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-infra-manager.yaml\n>f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-leader-election-role.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-leader-election-role.yaml\ncd+++++++++ objects/namespaced/envoy-gateway-system/secrets/\n>f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-gateway.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-gateway.yaml\n>f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-oidc-hmac.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-oidc-hmac.yaml\n>f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-rate-limit.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-rate-limit.yaml\n>f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy.yaml\n>f+++++++++ objects/namespaced/envoy-gateway-system/secrets/sh.helm.release.v1.envoy-gateway.v1.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/secrets/sh.helm.release.v1.envoy-gateway.v1.yaml\ncd+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/\n>f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/default.yaml\n>f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/envoy-gateway-gateway-helm-certgen.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/envoy-gateway-gateway-helm-certgen.yaml\n>f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/envoy-gateway.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/envoy-gateway.yaml\ncd+++++++++ objects/namespaced/envoy-gateway-system/services/\n>f+++++++++ objects/namespaced/envoy-gateway-system/services/envoy-gateway.txt\n>f+++++++++ objects/namespaced/envoy-gateway-system/services/envoy-gateway.yaml\ncd+++++++++ objects/namespaced/ingress-nginx/\ncd+++++++++ objects/namespaced/ingress-nginx/configmaps/\n>f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-controller.txt\n>f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-controller.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-tcp.txt\n>f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-tcp.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-udp.txt\n>f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-udp.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/ingress-nginx/configmaps/kube-root-ca.crt.yaml\ncd+++++++++ objects/namespaced/ingress-nginx/daemonsets/\n>f+++++++++ objects/namespaced/ingress-nginx/daemonsets/ingress-nginx-controller.txt\n>f+++++++++ objects/namespaced/ingress-nginx/daemonsets/ingress-nginx-controller.yaml\ncd+++++++++ objects/namespaced/ingress-nginx/deployment/\n>f+++++++++ objects/namespaced/ingress-nginx/deployment/ingress-nginx-defaultbackend.txt\n>f+++++++++ objects/namespaced/ingress-nginx/deployment/ingress-nginx-defaultbackend.yaml\ncd+++++++++ objects/namespaced/ingress-nginx/endpoints/\n>f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller-admission.txt\n>f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller-admission.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller-metrics.txt\n>f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller-metrics.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller.txt\n>f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-defaultbackend.txt\n>f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-defaultbackend.yaml\ncd+++++++++ objects/namespaced/ingress-nginx/pods/\n>f+++++++++ objects/namespaced/ingress-nginx/pods/ingress-nginx-controller-dmjt4.txt\n>f+++++++++ objects/namespaced/ingress-nginx/pods/ingress-nginx-controller-dmjt4.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/pods/ingress-nginx-defaultbackend-6987ff55cf-plsk9.txt\n>f+++++++++ objects/namespaced/ingress-nginx/pods/ingress-nginx-defaultbackend-6987ff55cf-plsk9.yaml\ncd+++++++++ objects/namespaced/ingress-nginx/rolebindings/\n>f+++++++++ objects/namespaced/ingress-nginx/rolebindings/ingress-nginx.txt\n>f+++++++++ objects/namespaced/ingress-nginx/rolebindings/ingress-nginx.yaml\ncd+++++++++ objects/namespaced/ingress-nginx/roles/\n>f+++++++++ objects/namespaced/ingress-nginx/roles/ingress-nginx.txt\n>f+++++++++ objects/namespaced/ingress-nginx/roles/ingress-nginx.yaml\ncd+++++++++ objects/namespaced/ingress-nginx/secrets/\n>f+++++++++ objects/namespaced/ingress-nginx/secrets/ingress-nginx-admission.txt\n>f+++++++++ objects/namespaced/ingress-nginx/secrets/ingress-nginx-admission.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/secrets/sh.helm.release.v1.ingress-nginx.v1.txt\n>f+++++++++ objects/namespaced/ingress-nginx/secrets/sh.helm.release.v1.ingress-nginx.v1.yaml\ncd+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/\n>f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/default.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/ingress-nginx-backend.txt\n>f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/ingress-nginx-backend.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/ingress-nginx.txt\n>f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/ingress-nginx.yaml\ncd+++++++++ objects/namespaced/ingress-nginx/services/\n>f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller-admission.txt\n>f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller-admission.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller-metrics.txt\n>f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller-metrics.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller.txt\n>f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller.yaml\n>f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-defaultbackend.txt\n>f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-defaultbackend.yaml\ncd+++++++++ objects/namespaced/kube-node-lease/\ncd+++++++++ objects/namespaced/kube-node-lease/configmaps/\n>f+++++++++ objects/namespaced/kube-node-lease/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/kube-node-lease/configmaps/kube-root-ca.crt.yaml\ncd+++++++++ objects/namespaced/kube-node-lease/serviceaccounts/\n>f+++++++++ objects/namespaced/kube-node-lease/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/kube-node-lease/serviceaccounts/default.yaml\ncd+++++++++ objects/namespaced/kube-public/\ncd+++++++++ objects/namespaced/kube-public/configmaps/\n>f+++++++++ objects/namespaced/kube-public/configmaps/cluster-info.txt\n>f+++++++++ objects/namespaced/kube-public/configmaps/cluster-info.yaml\n>f+++++++++ objects/namespaced/kube-public/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/kube-public/configmaps/kube-root-ca.crt.yaml\ncd+++++++++ objects/namespaced/kube-public/rolebindings/\n>f+++++++++ objects/namespaced/kube-public/rolebindings/kubeadm:bootstrap-signer-clusterinfo.txt\n>f+++++++++ objects/namespaced/kube-public/rolebindings/kubeadm:bootstrap-signer-clusterinfo.yaml\n>f+++++++++ objects/namespaced/kube-public/rolebindings/system:controller:bootstrap-signer.txt\n>f+++++++++ objects/namespaced/kube-public/rolebindings/system:controller:bootstrap-signer.yaml\ncd+++++++++ objects/namespaced/kube-public/roles/\n>f+++++++++ objects/namespaced/kube-public/roles/kubeadm:bootstrap-signer-clusterinfo.txt\n>f+++++++++ objects/namespaced/kube-public/roles/kubeadm:bootstrap-signer-clusterinfo.yaml\n>f+++++++++ objects/namespaced/kube-public/roles/system:controller:bootstrap-signer.txt\n>f+++++++++ objects/namespaced/kube-public/roles/system:controller:bootstrap-signer.yaml\ncd+++++++++ objects/namespaced/kube-public/serviceaccounts/\n>f+++++++++ objects/namespaced/kube-public/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/kube-public/serviceaccounts/default.yaml\ncd+++++++++ objects/namespaced/kube-system/\ncd+++++++++ objects/namespaced/kube-system/configmaps/\n>f+++++++++ objects/namespaced/kube-system/configmaps/cilium-config.txt\n>f+++++++++ objects/namespaced/kube-system/configmaps/cilium-config.yaml\n>f+++++++++ objects/namespaced/kube-system/configmaps/coredns.txt\n>f+++++++++ objects/namespaced/kube-system/configmaps/coredns.yaml\n>f+++++++++ objects/namespaced/kube-system/configmaps/extension-apiserver-authentication.txt\n>f+++++++++ objects/namespaced/kube-system/configmaps/extension-apiserver-authentication.yaml\n>f+++++++++ objects/namespaced/kube-system/configmaps/kube-apiserver-legacy-service-account-token-tracking.txt\n>f+++++++++ objects/namespaced/kube-system/configmaps/kube-apiserver-legacy-service-account-token-tracking.yaml\n>f+++++++++ objects/namespaced/kube-system/configmaps/kube-proxy.txt\n>f+++++++++ objects/namespaced/kube-system/configmaps/kube-proxy.yaml\n>f+++++++++ objects/namespaced/kube-system/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/kube-system/configmaps/kube-root-ca.crt.yaml\n>f+++++++++ objects/namespaced/kube-system/configmaps/kubeadm-config.txt\n>f+++++++++ objects/namespaced/kube-system/configmaps/kubeadm-config.yaml\n>f+++++++++ objects/namespaced/kube-system/configmaps/kubelet-config.txt\n>f+++++++++ objects/namespaced/kube-system/configmaps/kubelet-config.yaml\ncd+++++++++ objects/namespaced/kube-system/daemonsets/\n>f+++++++++ objects/namespaced/kube-system/daemonsets/cilium.txt\n>f+++++++++ objects/namespaced/kube-system/daemonsets/cilium.yaml\n>f+++++++++ objects/namespaced/kube-system/daemonsets/kube-proxy.txt\n>f+++++++++ objects/namespaced/kube-system/daemonsets/kube-proxy.yaml\ncd+++++++++ objects/namespaced/kube-system/deployment/\n>f+++++++++ objects/namespaced/kube-system/deployment/cilium-operator.txt\n>f+++++++++ objects/namespaced/kube-system/deployment/cilium-operator.yaml\n>f+++++++++ objects/namespaced/kube-system/deployment/coredns.txt\n>f+++++++++ objects/namespaced/kube-system/deployment/coredns.yaml\ncd+++++++++ objects/namespaced/kube-system/endpoints/\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-dns.txt\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-dns.yaml\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-coredns.txt\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-coredns.yaml\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-controller-manager.txt\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-controller-manager.yaml\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-etcd.txt\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-etcd.yaml\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-proxy.txt\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-proxy.yaml\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-scheduler.txt\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-scheduler.yaml\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kubelet.txt\n>f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kubelet.yaml\ncd+++++++++ objects/namespaced/kube-system/pods/\n>f+++++++++ objects/namespaced/kube-system/pods/cilium-operator-869df985b8-nqvtb.txt\n>f+++++++++ objects/namespaced/kube-system/pods/cilium-operator-869df985b8-nqvtb.yaml\n>f+++++++++ objects/namespaced/kube-system/pods/cilium-sk6n5.txt\n>f+++++++++ objects/namespaced/kube-system/pods/cilium-sk6n5.yaml\n>f+++++++++ objects/namespaced/kube-system/pods/coredns-67659f764b-mdrt6.txt\n>f+++++++++ objects/namespaced/kube-system/pods/coredns-67659f764b-mdrt6.yaml\n>f+++++++++ objects/namespaced/kube-system/pods/coredns-67659f764b-xb6lv.txt\n>f+++++++++ objects/namespaced/kube-system/pods/coredns-67659f764b-xb6lv.yaml\n>f+++++++++ objects/namespaced/kube-system/pods/etcd-instance.txt\n>f+++++++++ objects/namespaced/kube-system/pods/etcd-instance.yaml\n>f+++++++++ objects/namespaced/kube-system/pods/kube-apiserver-instance.txt\n>f+++++++++ objects/namespaced/kube-system/pods/kube-apiserver-instance.yaml\n>f+++++++++ objects/namespaced/kube-system/pods/kube-controller-manager-instance.txt\n>f+++++++++ objects/namespaced/kube-system/pods/kube-controller-manager-instance.yaml\n>f+++++++++ objects/namespaced/kube-system/pods/kube-proxy-cxlxv.txt\n>f+++++++++ objects/namespaced/kube-system/pods/kube-proxy-cxlxv.yaml\n>f+++++++++ objects/namespaced/kube-system/pods/kube-scheduler-instance.txt\n>f+++++++++ objects/namespaced/kube-system/pods/kube-scheduler-instance.yaml\n>f+++++++++ objects/namespaced/kube-system/pods/kube-vip-instance.txt\n>f+++++++++ objects/namespaced/kube-system/pods/kube-vip-instance.yaml\ncd+++++++++ objects/namespaced/kube-system/rolebindings/\n>f+++++++++ objects/namespaced/kube-system/rolebindings/cilium-config-agent.txt\n>f+++++++++ objects/namespaced/kube-system/rolebindings/cilium-config-agent.yaml\n>f+++++++++ objects/namespaced/kube-system/rolebindings/kube-proxy.txt\n>f+++++++++ objects/namespaced/kube-system/rolebindings/kube-proxy.yaml\n>f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:kubeadm-certs.txt\n>f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:kubeadm-certs.yaml\n>f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:kubelet-config.txt\n>f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:kubelet-config.yaml\n>f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:nodes-kubeadm-config.txt\n>f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:nodes-kubeadm-config.yaml\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system::extension-apiserver-authentication-reader.txt\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system::extension-apiserver-authentication-reader.yaml\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system::leader-locking-kube-controller-manager.txt\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system::leader-locking-kube-controller-manager.yaml\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system::leader-locking-kube-scheduler.txt\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system::leader-locking-kube-scheduler.yaml\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:bootstrap-signer.txt\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:bootstrap-signer.yaml\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:cloud-provider.txt\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:cloud-provider.yaml\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:token-cleaner.txt\n>f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:token-cleaner.yaml\ncd+++++++++ objects/namespaced/kube-system/roles/\n>f+++++++++ objects/namespaced/kube-system/roles/cilium-config-agent.txt\n>f+++++++++ objects/namespaced/kube-system/roles/cilium-config-agent.yaml\n>f+++++++++ objects/namespaced/kube-system/roles/extension-apiserver-authentication-reader.txt\n>f+++++++++ objects/namespaced/kube-system/roles/extension-apiserver-authentication-reader.yaml\n>f+++++++++ objects/namespaced/kube-system/roles/kube-proxy.txt\n>f+++++++++ objects/namespaced/kube-system/roles/kube-proxy.yaml\n>f+++++++++ objects/namespaced/kube-system/roles/kubeadm:kubeadm-certs.txt\n>f+++++++++ objects/namespaced/kube-system/roles/kubeadm:kubeadm-certs.yaml\n>f+++++++++ objects/namespaced/kube-system/roles/kubeadm:kubelet-config.txt\n>f+++++++++ objects/namespaced/kube-system/roles/kubeadm:kubelet-config.yaml\n>f+++++++++ objects/namespaced/kube-system/roles/kubeadm:nodes-kubeadm-config.txt\n>f+++++++++ objects/namespaced/kube-system/roles/kubeadm:nodes-kubeadm-config.yaml\n>f+++++++++ objects/namespaced/kube-system/roles/system::leader-locking-kube-controller-manager.txt\n>f+++++++++ objects/namespaced/kube-system/roles/system::leader-locking-kube-controller-manager.yaml\n>f+++++++++ objects/namespaced/kube-system/roles/system::leader-locking-kube-scheduler.txt\n>f+++++++++ objects/namespaced/kube-system/roles/system::leader-locking-kube-scheduler.yaml\n>f+++++++++ objects/namespaced/kube-system/roles/system:controller:bootstrap-signer.txt\n>f+++++++++ objects/namespaced/kube-system/roles/system:controller:bootstrap-signer.yaml\n>f+++++++++ objects/namespaced/kube-system/roles/system:controller:cloud-provider.txt\n>f+++++++++ objects/namespaced/kube-system/roles/system:controller:cloud-provider.yaml\n>f+++++++++ objects/namespaced/kube-system/roles/system:controller:token-cleaner.txt\n>f+++++++++ objects/namespaced/kube-system/roles/system:controller:token-cleaner.yaml\ncd+++++++++ objects/namespaced/kube-system/secrets/\n>f+++++++++ objects/namespaced/kube-system/secrets/bootstrap-token-3n8qak.txt\n>f+++++++++ objects/namespaced/kube-system/secrets/bootstrap-token-3n8qak.yaml\n>f+++++++++ objects/namespaced/kube-system/secrets/bootstrap-token-w4o225.txt\n>f+++++++++ objects/namespaced/kube-system/secrets/bootstrap-token-w4o225.yaml\n>f+++++++++ objects/namespaced/kube-system/secrets/kubeadm-certs.txt\n>f+++++++++ objects/namespaced/kube-system/secrets/kubeadm-certs.yaml\n>f+++++++++ objects/namespaced/kube-system/secrets/sh.helm.release.v1.cilium.v1.txt\n>f+++++++++ objects/namespaced/kube-system/secrets/sh.helm.release.v1.cilium.v1.yaml\ncd+++++++++ objects/namespaced/kube-system/serviceaccounts/\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/attachdetach-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/attachdetach-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/bootstrap-signer.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/bootstrap-signer.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/certificate-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/certificate-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/cilium-operator.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/cilium-operator.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/cilium.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/cilium.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/clusterrole-aggregation-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/clusterrole-aggregation-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/coredns.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/coredns.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/cronjob-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/cronjob-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/daemon-set-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/daemon-set-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/default.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/deployment-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/deployment-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/disruption-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/disruption-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpoint-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpoint-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpointslice-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpointslice-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpointslicemirroring-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpointslicemirroring-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/ephemeral-volume-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/ephemeral-volume-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/expand-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/expand-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/generic-garbage-collector.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/generic-garbage-collector.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/horizontal-pod-autoscaler.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/horizontal-pod-autoscaler.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/job-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/job-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/kube-proxy.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/kube-proxy.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/namespace-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/namespace-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/node-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/node-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/persistent-volume-binder.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/persistent-volume-binder.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/pod-garbage-collector.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/pod-garbage-collector.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/pv-protection-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/pv-protection-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/pvc-protection-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/pvc-protection-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/replicaset-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/replicaset-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/replication-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/replication-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/resourcequota-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/resourcequota-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/root-ca-cert-publisher.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/root-ca-cert-publisher.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/service-account-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/service-account-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/service-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/service-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/statefulset-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/statefulset-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/token-cleaner.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/token-cleaner.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/ttl-after-finished-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/ttl-after-finished-controller.yaml\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/ttl-controller.txt\n>f+++++++++ objects/namespaced/kube-system/serviceaccounts/ttl-controller.yaml\ncd+++++++++ objects/namespaced/kube-system/services/\n>f+++++++++ objects/namespaced/kube-system/services/kube-dns.txt\n>f+++++++++ objects/namespaced/kube-system/services/kube-dns.yaml\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-coredns.txt\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-coredns.yaml\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-controller-manager.txt\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-controller-manager.yaml\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-etcd.txt\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-etcd.yaml\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-proxy.txt\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-proxy.yaml\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-scheduler.txt\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-scheduler.yaml\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kubelet.txt\n>f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kubelet.yaml\ncd+++++++++ objects/namespaced/local-path-storage/\ncd+++++++++ objects/namespaced/local-path-storage/configmaps/\n>f+++++++++ objects/namespaced/local-path-storage/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/local-path-storage/configmaps/kube-root-ca.crt.yaml\n>f+++++++++ objects/namespaced/local-path-storage/configmaps/local-path-config.txt\n>f+++++++++ objects/namespaced/local-path-storage/configmaps/local-path-config.yaml\ncd+++++++++ objects/namespaced/local-path-storage/deployment/\n>f+++++++++ objects/namespaced/local-path-storage/deployment/local-path-provisioner.txt\n>f+++++++++ objects/namespaced/local-path-storage/deployment/local-path-provisioner.yaml\ncd+++++++++ objects/namespaced/local-path-storage/pods/\n>f+++++++++ objects/namespaced/local-path-storage/pods/local-path-provisioner-679c578f5-m52kp.txt\n>f+++++++++ objects/namespaced/local-path-storage/pods/local-path-provisioner-679c578f5-m52kp.yaml\ncd+++++++++ objects/namespaced/local-path-storage/secrets/\n>f+++++++++ objects/namespaced/local-path-storage/secrets/sh.helm.release.v1.local-path-provisioner.v1.txt\n>f+++++++++ objects/namespaced/local-path-storage/secrets/sh.helm.release.v1.local-path-provisioner.v1.yaml\ncd+++++++++ objects/namespaced/local-path-storage/serviceaccounts/\n>f+++++++++ objects/namespaced/local-path-storage/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/local-path-storage/serviceaccounts/default.yaml\n>f+++++++++ objects/namespaced/local-path-storage/serviceaccounts/local-path-provisioner.txt\n>f+++++++++ objects/namespaced/local-path-storage/serviceaccounts/local-path-provisioner.yaml\ncd+++++++++ objects/namespaced/monitoring/\ncd+++++++++ objects/namespaced/monitoring/configmaps/\n>f+++++++++ objects/namespaced/monitoring/configmaps/goldpinger-zap.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/goldpinger-zap.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/ipmi-exporter.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/ipmi-exporter.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-alertmanager-overview.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-alertmanager-overview.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-apiserver.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-apiserver.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-cluster-total.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-cluster-total.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-controller-manager.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-controller-manager.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster-advanced.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster-advanced.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-goldpinger.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-goldpinger.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-haproxy.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-haproxy.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-host-details.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-host-details.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-hosts-overview.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-hosts-overview.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-node-exporter-full.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-node-exporter-full.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-osd-device-details.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-osd-device-details.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-osds-overview.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-osds-overview.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-pool-detail.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-pool-detail.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-pool-overview.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-pool-overview.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-rbd-details.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-rbd-details.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-rbd-overview.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-rbd-overview.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-etcd.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-etcd.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-config-dashboards.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-config-dashboards.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-datasource.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-datasource.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-overview.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-overview.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-coredns.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-coredns.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-cluster.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-cluster.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-multicluster.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-multicluster.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-namespace.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-namespace.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-node.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-node.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-pod.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-pod.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-workload.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-workload.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-workloads-namespace.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-workloads-namespace.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-kubelet.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-kubelet.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-namespace-by-pod.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-namespace-by-pod.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-namespace-by-workload.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-namespace-by-workload.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-cluster-rsrc-use.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-cluster-rsrc-use.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-exporter.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-exporter.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-rsrc-use.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-rsrc-use.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-nodes-darwin.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-nodes-darwin.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-nodes.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-nodes.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-persistentvolumesusage.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-persistentvolumesusage.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-pod-total.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-pod-total.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-prometheus-tls.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-prometheus-tls.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-prometheus.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-prometheus.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-proxy.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-proxy.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-scheduler.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-scheduler.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-workload-total.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-workload-total.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/kube-root-ca.crt.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki-alerting-rules.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki-alerting-rules.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki-dashboards-1.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki-dashboards-1.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki-dashboards-2.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki-dashboards-2.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki-gateway.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki-gateway.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki-runtime.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki-runtime.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/loki.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/node-feature-discovery-worker-conf.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/node-feature-discovery-worker-conf.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/prometheus-kube-prometheus-stack-prometheus-rulefiles-0.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/prometheus-kube-prometheus-stack-prometheus-rulefiles-0.yaml\n>f+++++++++ objects/namespaced/monitoring/configmaps/vector.txt\n>f+++++++++ objects/namespaced/monitoring/configmaps/vector.yaml\ncd+++++++++ objects/namespaced/monitoring/daemonsets/\n>f+++++++++ objects/namespaced/monitoring/daemonsets/goldpinger.txt\n>f+++++++++ objects/namespaced/monitoring/daemonsets/goldpinger.yaml\n>f+++++++++ objects/namespaced/monitoring/daemonsets/ipmi-exporter.txt\n>f+++++++++ objects/namespaced/monitoring/daemonsets/ipmi-exporter.yaml\n>f+++++++++ objects/namespaced/monitoring/daemonsets/kube-prometheus-stack-prometheus-node-exporter.txt\n>f+++++++++ objects/namespaced/monitoring/daemonsets/kube-prometheus-stack-prometheus-node-exporter.yaml\n>f+++++++++ objects/namespaced/monitoring/daemonsets/node-feature-discovery-worker.txt\n>f+++++++++ objects/namespaced/monitoring/daemonsets/node-feature-discovery-worker.yaml\n>f+++++++++ objects/namespaced/monitoring/daemonsets/vector.txt\n>f+++++++++ objects/namespaced/monitoring/daemonsets/vector.yaml\ncd+++++++++ objects/namespaced/monitoring/deployment/\n>f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-grafana.txt\n>f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-grafana.yaml\n>f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-kube-state-metrics.txt\n>f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-kube-state-metrics.yaml\n>f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-operator.txt\n>f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-operator.yaml\n>f+++++++++ objects/namespaced/monitoring/deployment/loki-gateway.txt\n>f+++++++++ objects/namespaced/monitoring/deployment/loki-gateway.yaml\n>f+++++++++ objects/namespaced/monitoring/deployment/node-feature-discovery-master.txt\n>f+++++++++ objects/namespaced/monitoring/deployment/node-feature-discovery-master.yaml\n>f+++++++++ objects/namespaced/monitoring/deployment/prometheus-pushgateway.txt\n>f+++++++++ objects/namespaced/monitoring/deployment/prometheus-pushgateway.yaml\ncd+++++++++ objects/namespaced/monitoring/endpoints/\n>f+++++++++ objects/namespaced/monitoring/endpoints/alertmanager-operated.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/alertmanager-operated.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/goldpinger.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/goldpinger.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-alertmanager.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-alertmanager.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-grafana.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-grafana.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-kube-state-metrics.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-kube-state-metrics.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-operator.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-operator.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-prometheus-node-exporter.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-prometheus-node-exporter.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-prometheus.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-prometheus.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/loki-gateway.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/loki-gateway.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/loki-headless.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/loki-headless.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/loki-memberlist.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/loki-memberlist.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/loki.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/loki.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/node-feature-discovery-master.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/node-feature-discovery-master.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/prometheus-operated.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/prometheus-operated.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/prometheus-pushgateway.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/prometheus-pushgateway.yaml\n>f+++++++++ objects/namespaced/monitoring/endpoints/vector-headless.txt\n>f+++++++++ objects/namespaced/monitoring/endpoints/vector-headless.yaml\ncd+++++++++ objects/namespaced/monitoring/ingresses/\n>f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-alertmanager.txt\n>f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-alertmanager.yaml\n>f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-grafana.txt\n>f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-grafana.yaml\n>f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-prometheus.txt\n>f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-prometheus.yaml\ncd+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/\n>f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/alertmanager-kube-prometheus-stack-alertmanager-db-alertmanager-kube-prometheus-stack-alertmanager-0.txt\n>f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/alertmanager-kube-prometheus-stack-alertmanager-db-alertmanager-kube-prometheus-stack-alertmanager-0.yaml\n>f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/prometheus-kube-prometheus-stack-prometheus-db-prometheus-kube-prometheus-stack-prometheus-0.txt\n>f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/prometheus-kube-prometheus-stack-prometheus-db-prometheus-kube-prometheus-stack-prometheus-0.yaml\n>f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/storage-loki-0.txt\n>f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/storage-loki-0.yaml\ncd+++++++++ objects/namespaced/monitoring/pods/\n>f+++++++++ objects/namespaced/monitoring/pods/alertmanager-kube-prometheus-stack-alertmanager-0.txt\n>f+++++++++ objects/namespaced/monitoring/pods/alertmanager-kube-prometheus-stack-alertmanager-0.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/goldpinger-gz8xd.txt\n>f+++++++++ objects/namespaced/monitoring/pods/goldpinger-gz8xd.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-grafana-69fc56495c-5h4gd.txt\n>f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-grafana-69fc56495c-5h4gd.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj.txt\n>f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-operator-cc88b86d6-xspgk.txt\n>f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-operator-cc88b86d6-xspgk.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-prometheus-node-exporter-jmj6n.txt\n>f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-prometheus-node-exporter-jmj6n.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/loki-0.txt\n>f+++++++++ objects/namespaced/monitoring/pods/loki-0.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/loki-gateway-5cc45946cd-wp7ks.txt\n>f+++++++++ objects/namespaced/monitoring/pods/loki-gateway-5cc45946cd-wp7ks.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/node-feature-discovery-master-68cf466d5f-qknzk.txt\n>f+++++++++ objects/namespaced/monitoring/pods/node-feature-discovery-master-68cf466d5f-qknzk.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/node-feature-discovery-worker-7l5z7.txt\n>f+++++++++ objects/namespaced/monitoring/pods/node-feature-discovery-worker-7l5z7.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/prometheus-kube-prometheus-stack-prometheus-0.txt\n>f+++++++++ objects/namespaced/monitoring/pods/prometheus-kube-prometheus-stack-prometheus-0.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/prometheus-pushgateway-5c694d99cc-jnjt7.txt\n>f+++++++++ objects/namespaced/monitoring/pods/prometheus-pushgateway-5c694d99cc-jnjt7.yaml\n>f+++++++++ objects/namespaced/monitoring/pods/vector-75jgf.txt\n>f+++++++++ objects/namespaced/monitoring/pods/vector-75jgf.yaml\ncd+++++++++ objects/namespaced/monitoring/rolebindings/\n>f+++++++++ objects/namespaced/monitoring/rolebindings/kube-prometheus-stack-grafana.txt\n>f+++++++++ objects/namespaced/monitoring/rolebindings/kube-prometheus-stack-grafana.yaml\n>f+++++++++ objects/namespaced/monitoring/rolebindings/kube-prometheus-stack-pod-tls-sidecar.txt\n>f+++++++++ objects/namespaced/monitoring/rolebindings/kube-prometheus-stack-pod-tls-sidecar.yaml\ncd+++++++++ objects/namespaced/monitoring/roles/\n>f+++++++++ objects/namespaced/monitoring/roles/kube-prometheus-stack-grafana.txt\n>f+++++++++ objects/namespaced/monitoring/roles/kube-prometheus-stack-grafana.yaml\n>f+++++++++ objects/namespaced/monitoring/roles/kube-prometheus-stack-pod-tls-sidecar.txt\n>f+++++++++ objects/namespaced/monitoring/roles/kube-prometheus-stack-pod-tls-sidecar.yaml\ncd+++++++++ objects/namespaced/monitoring/secrets/\n>f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-generated.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-generated.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-tls-assets-0.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-tls-assets-0.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-web-config.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-web-config.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-tls.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-tls.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/grafana-tls.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/grafana-tls.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-admission.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-admission.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-client-secret.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-client-secret.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-cookie-secret.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-cookie-secret.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-oauth2-proxy.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-oauth2-proxy.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-etcd-client-cert.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-etcd-client-cert.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-grafana-client-secret.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-grafana-client-secret.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-grafana.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-grafana.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-client-secret.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-client-secret.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-cookie-secret.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-cookie-secret.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-node-exporter-jmj6n-tls.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-node-exporter-jmj6n-tls.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-oauth2-proxy.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-oauth2-proxy.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-0-tls.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-0-tls.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-tls-assets-0.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-tls-assets-0.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-web-config.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-web-config.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/prometheus-tls.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/prometheus-tls.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.goldpinger.v1.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.goldpinger.v1.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.kube-prometheus-stack.v1.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.kube-prometheus-stack.v1.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.loki.v1.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.loki.v1.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.node-feature-discovery.v1.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.node-feature-discovery.v1.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.prometheus-pushgateway.v1.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.prometheus-pushgateway.v1.yaml\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.vector.v1.txt\n>f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.vector.v1.yaml\ncd+++++++++ objects/namespaced/monitoring/serviceaccounts/\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/default.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/goldpinger.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/goldpinger.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-alertmanager.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-alertmanager.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-grafana.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-grafana.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-kube-state-metrics.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-kube-state-metrics.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-operator.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-operator.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-prometheus-node-exporter.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-prometheus-node-exporter.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-prometheus.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-prometheus.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/loki.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/loki.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/node-feature-discovery-worker.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/node-feature-discovery-worker.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/node-feature-discovery.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/node-feature-discovery.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/prometheus-pushgateway.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/prometheus-pushgateway.yaml\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/vector.txt\n>f+++++++++ objects/namespaced/monitoring/serviceaccounts/vector.yaml\ncd+++++++++ objects/namespaced/monitoring/services/\n>f+++++++++ objects/namespaced/monitoring/services/alertmanager-operated.txt\n>f+++++++++ objects/namespaced/monitoring/services/alertmanager-operated.yaml\n>f+++++++++ objects/namespaced/monitoring/services/goldpinger.txt\n>f+++++++++ objects/namespaced/monitoring/services/goldpinger.yaml\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-alertmanager.txt\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-alertmanager.yaml\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-grafana.txt\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-grafana.yaml\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-kube-state-metrics.txt\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-kube-state-metrics.yaml\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-operator.txt\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-operator.yaml\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-prometheus-node-exporter.txt\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-prometheus-node-exporter.yaml\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-prometheus.txt\n>f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-prometheus.yaml\n>f+++++++++ objects/namespaced/monitoring/services/loki-gateway.txt\n>f+++++++++ objects/namespaced/monitoring/services/loki-gateway.yaml\n>f+++++++++ objects/namespaced/monitoring/services/loki-headless.txt\n>f+++++++++ objects/namespaced/monitoring/services/loki-headless.yaml\n>f+++++++++ objects/namespaced/monitoring/services/loki-memberlist.txt\n>f+++++++++ objects/namespaced/monitoring/services/loki-memberlist.yaml\n>f+++++++++ objects/namespaced/monitoring/services/loki.txt\n>f+++++++++ objects/namespaced/monitoring/services/loki.yaml\n>f+++++++++ objects/namespaced/monitoring/services/node-feature-discovery-master.txt\n>f+++++++++ objects/namespaced/monitoring/services/node-feature-discovery-master.yaml\n>f+++++++++ objects/namespaced/monitoring/services/prometheus-operated.txt\n>f+++++++++ objects/namespaced/monitoring/services/prometheus-operated.yaml\n>f+++++++++ objects/namespaced/monitoring/services/prometheus-pushgateway.txt\n>f+++++++++ objects/namespaced/monitoring/services/prometheus-pushgateway.yaml\n>f+++++++++ objects/namespaced/monitoring/services/vector-headless.txt\n>f+++++++++ objects/namespaced/monitoring/services/vector-headless.yaml\ncd+++++++++ objects/namespaced/monitoring/statefulsets/\n>f+++++++++ objects/namespaced/monitoring/statefulsets/alertmanager-kube-prometheus-stack-alertmanager.txt\n>f+++++++++ objects/namespaced/monitoring/statefulsets/alertmanager-kube-prometheus-stack-alertmanager.yaml\n>f+++++++++ objects/namespaced/monitoring/statefulsets/loki.txt\n>f+++++++++ objects/namespaced/monitoring/statefulsets/loki.yaml\n>f+++++++++ objects/namespaced/monitoring/statefulsets/prometheus-kube-prometheus-stack-prometheus.txt\n>f+++++++++ objects/namespaced/monitoring/statefulsets/prometheus-kube-prometheus-stack-prometheus.yaml\ncd+++++++++ objects/namespaced/openstack/\ncd+++++++++ objects/namespaced/openstack/configmaps/\n>f+++++++++ objects/namespaced/openstack/configmaps/keepalived-bin.txt\n>f+++++++++ objects/namespaced/openstack/configmaps/keepalived-bin.yaml\n>f+++++++++ objects/namespaced/openstack/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/openstack/configmaps/kube-root-ca.crt.yaml\n>f+++++++++ objects/namespaced/openstack/configmaps/memcached-memcached-bin.txt\n>f+++++++++ objects/namespaced/openstack/configmaps/memcached-memcached-bin.yaml\n>f+++++++++ objects/namespaced/openstack/configmaps/messaging-topology-operator-leader-election.txt\n>f+++++++++ objects/namespaced/openstack/configmaps/messaging-topology-operator-leader-election.yaml\n>f+++++++++ objects/namespaced/openstack/configmaps/percona-xtradb-haproxy.txt\n>f+++++++++ objects/namespaced/openstack/configmaps/percona-xtradb-haproxy.yaml\n>f+++++++++ objects/namespaced/openstack/configmaps/percona-xtradb-pxc.txt\n>f+++++++++ objects/namespaced/openstack/configmaps/percona-xtradb-pxc.yaml\n>f+++++++++ objects/namespaced/openstack/configmaps/rabbitmq-cluster-operator-leader-election.txt\n>f+++++++++ objects/namespaced/openstack/configmaps/rabbitmq-cluster-operator-leader-election.yaml\ncd+++++++++ objects/namespaced/openstack/daemonsets/\n>f+++++++++ objects/namespaced/openstack/daemonsets/keepalived.txt\n>f+++++++++ objects/namespaced/openstack/daemonsets/keepalived.yaml\ncd+++++++++ objects/namespaced/openstack/deployment/\n>f+++++++++ objects/namespaced/openstack/deployment/memcached-memcached.txt\n>f+++++++++ objects/namespaced/openstack/deployment/memcached-memcached.yaml\n>f+++++++++ objects/namespaced/openstack/deployment/pxc-operator.txt\n>f+++++++++ objects/namespaced/openstack/deployment/pxc-operator.yaml\n>f+++++++++ objects/namespaced/openstack/deployment/rabbitmq-cluster-operator.txt\n>f+++++++++ objects/namespaced/openstack/deployment/rabbitmq-cluster-operator.yaml\n>f+++++++++ objects/namespaced/openstack/deployment/rabbitmq-messaging-topology-operator.txt\n>f+++++++++ objects/namespaced/openstack/deployment/rabbitmq-messaging-topology-operator.yaml\ncd+++++++++ objects/namespaced/openstack/endpoints/\n>f+++++++++ objects/namespaced/openstack/endpoints/memcached-metrics.txt\n>f+++++++++ objects/namespaced/openstack/endpoints/memcached-metrics.yaml\n>f+++++++++ objects/namespaced/openstack/endpoints/memcached.txt\n>f+++++++++ objects/namespaced/openstack/endpoints/memcached.yaml\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-cluster-operator.txt\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-cluster-operator.yaml\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy-metrics.txt\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy-metrics.yaml\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy-replicas.txt\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy-replicas.yaml\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy.txt\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy.yaml\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-pxc-unready.txt\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-pxc-unready.yaml\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-pxc.txt\n>f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-pxc.yaml\n>f+++++++++ objects/namespaced/openstack/endpoints/rabbitmq-messaging-topology-operator-webhook.txt\n>f+++++++++ objects/namespaced/openstack/endpoints/rabbitmq-messaging-topology-operator-webhook.yaml\ncd+++++++++ objects/namespaced/openstack/persistentvolumeclaims/\n>f+++++++++ objects/namespaced/openstack/persistentvolumeclaims/datadir-percona-xtradb-pxc-0.txt\n>f+++++++++ objects/namespaced/openstack/persistentvolumeclaims/datadir-percona-xtradb-pxc-0.yaml\ncd+++++++++ objects/namespaced/openstack/pods/\n>f+++++++++ objects/namespaced/openstack/pods/keepalived-wktjf.txt\n>f+++++++++ objects/namespaced/openstack/pods/keepalived-wktjf.yaml\n>f+++++++++ objects/namespaced/openstack/pods/memcached-memcached-c47d949f5-vc5cm.txt\n>f+++++++++ objects/namespaced/openstack/pods/memcached-memcached-c47d949f5-vc5cm.yaml\n>f+++++++++ objects/namespaced/openstack/pods/percona-xtradb-haproxy-0.txt\n>f+++++++++ objects/namespaced/openstack/pods/percona-xtradb-haproxy-0.yaml\n>f+++++++++ objects/namespaced/openstack/pods/percona-xtradb-pxc-0.txt\n>f+++++++++ objects/namespaced/openstack/pods/percona-xtradb-pxc-0.yaml\n>f+++++++++ objects/namespaced/openstack/pods/pxc-operator-59d66cbc78-clxvf.txt\n>f+++++++++ objects/namespaced/openstack/pods/pxc-operator-59d66cbc78-clxvf.yaml\n>f+++++++++ objects/namespaced/openstack/pods/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf.txt\n>f+++++++++ objects/namespaced/openstack/pods/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf.yaml\n>f+++++++++ objects/namespaced/openstack/pods/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc.txt\n>f+++++++++ objects/namespaced/openstack/pods/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc.yaml\ncd+++++++++ objects/namespaced/openstack/rolebindings/\n>f+++++++++ objects/namespaced/openstack/rolebindings/keepalived.txt\n>f+++++++++ objects/namespaced/openstack/rolebindings/keepalived.yaml\n>f+++++++++ objects/namespaced/openstack/rolebindings/pxc-operator.txt\n>f+++++++++ objects/namespaced/openstack/rolebindings/pxc-operator.yaml\n>f+++++++++ objects/namespaced/openstack/rolebindings/rabbitmq-cluster-operator.txt\n>f+++++++++ objects/namespaced/openstack/rolebindings/rabbitmq-cluster-operator.yaml\n>f+++++++++ objects/namespaced/openstack/rolebindings/rabbitmq-messaging-topology-operator.txt\n>f+++++++++ objects/namespaced/openstack/rolebindings/rabbitmq-messaging-topology-operator.yaml\ncd+++++++++ objects/namespaced/openstack/roles/\n>f+++++++++ objects/namespaced/openstack/roles/keepalived.txt\n>f+++++++++ objects/namespaced/openstack/roles/keepalived.yaml\n>f+++++++++ objects/namespaced/openstack/roles/pxc-operator.txt\n>f+++++++++ objects/namespaced/openstack/roles/pxc-operator.yaml\n>f+++++++++ objects/namespaced/openstack/roles/rabbitmq-cluster-operator.txt\n>f+++++++++ objects/namespaced/openstack/roles/rabbitmq-cluster-operator.yaml\n>f+++++++++ objects/namespaced/openstack/roles/rabbitmq-messaging-topology-operator.txt\n>f+++++++++ objects/namespaced/openstack/roles/rabbitmq-messaging-topology-operator.yaml\ncd+++++++++ objects/namespaced/openstack/secrets/\n>f+++++++++ objects/namespaced/openstack/secrets/internal-percona-xtradb.txt\n>f+++++++++ objects/namespaced/openstack/secrets/internal-percona-xtradb.yaml\n>f+++++++++ objects/namespaced/openstack/secrets/keepalived-etc.txt\n>f+++++++++ objects/namespaced/openstack/secrets/keepalived-etc.yaml\n>f+++++++++ objects/namespaced/openstack/secrets/percona-xtradb.txt\n>f+++++++++ objects/namespaced/openstack/secrets/percona-xtradb.yaml\n>f+++++++++ objects/namespaced/openstack/secrets/rabbitmq-messaging-topology-operator-webhook.txt\n>f+++++++++ objects/namespaced/openstack/secrets/rabbitmq-messaging-topology-operator-webhook.yaml\n>f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.memcached.v1.txt\n>f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.memcached.v1.yaml\n>f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.pxc-operator.v1.txt\n>f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.pxc-operator.v1.yaml\n>f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.rabbitmq-cluster-operator.v1.txt\n>f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.rabbitmq-cluster-operator.v1.yaml\ncd+++++++++ objects/namespaced/openstack/serviceaccounts/\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/default.yaml\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/keepalived.txt\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/keepalived.yaml\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/memcached-memcached.txt\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/memcached-memcached.yaml\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/pxc-operator.txt\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/pxc-operator.yaml\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/rabbitmq-cluster-operator.txt\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/rabbitmq-cluster-operator.yaml\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/rabbitmq-messaging-topology-operator.txt\n>f+++++++++ objects/namespaced/openstack/serviceaccounts/rabbitmq-messaging-topology-operator.yaml\ncd+++++++++ objects/namespaced/openstack/services/\n>f+++++++++ objects/namespaced/openstack/services/memcached-metrics.txt\n>f+++++++++ objects/namespaced/openstack/services/memcached-metrics.yaml\n>f+++++++++ objects/namespaced/openstack/services/memcached.txt\n>f+++++++++ objects/namespaced/openstack/services/memcached.yaml\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-cluster-operator.txt\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-cluster-operator.yaml\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy-metrics.txt\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy-metrics.yaml\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy-replicas.txt\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy-replicas.yaml\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy.txt\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy.yaml\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-pxc-unready.txt\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-pxc-unready.yaml\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-pxc.txt\n>f+++++++++ objects/namespaced/openstack/services/percona-xtradb-pxc.yaml\n>f+++++++++ objects/namespaced/openstack/services/rabbitmq-messaging-topology-operator-webhook.txt\n>f+++++++++ objects/namespaced/openstack/services/rabbitmq-messaging-topology-operator-webhook.yaml\ncd+++++++++ objects/namespaced/openstack/statefulsets/\n>f+++++++++ objects/namespaced/openstack/statefulsets/percona-xtradb-haproxy.txt\n>f+++++++++ objects/namespaced/openstack/statefulsets/percona-xtradb-haproxy.yaml\n>f+++++++++ objects/namespaced/openstack/statefulsets/percona-xtradb-pxc.txt\n>f+++++++++ objects/namespaced/openstack/statefulsets/percona-xtradb-pxc.yaml\ncd+++++++++ objects/namespaced/secretgen-controller/\ncd+++++++++ objects/namespaced/secretgen-controller/configmaps/\n>f+++++++++ objects/namespaced/secretgen-controller/configmaps/kube-root-ca.crt.txt\n>f+++++++++ objects/namespaced/secretgen-controller/configmaps/kube-root-ca.crt.yaml\ncd+++++++++ objects/namespaced/secretgen-controller/deployment/\n>f+++++++++ objects/namespaced/secretgen-controller/deployment/secretgen-controller.txt\n>f+++++++++ objects/namespaced/secretgen-controller/deployment/secretgen-controller.yaml\ncd+++++++++ objects/namespaced/secretgen-controller/pods/\n>f+++++++++ objects/namespaced/secretgen-controller/pods/secretgen-controller-5cf976ccc7-vwxz4.txt\n>f+++++++++ objects/namespaced/secretgen-controller/pods/secretgen-controller-5cf976ccc7-vwxz4.yaml\ncd+++++++++ objects/namespaced/secretgen-controller/serviceaccounts/\n>f+++++++++ objects/namespaced/secretgen-controller/serviceaccounts/default.txt\n>f+++++++++ objects/namespaced/secretgen-controller/serviceaccounts/default.yaml\n>f+++++++++ objects/namespaced/secretgen-controller/serviceaccounts/secretgen-controller-sa.txt\n>f+++++++++ objects/namespaced/secretgen-controller/serviceaccounts/secretgen-controller-sa.yaml\n",
                            "rc": 0,
                            "stdout_lines": [
                                "cd+++++++++ objects/",
                                "cd+++++++++ objects/cluster/",
                                "cd+++++++++ objects/cluster/clusterrole/",
                                ">f+++++++++ objects/cluster/clusterrole/admin.txt",
                                ">f+++++++++ objects/cluster/clusterrole/admin.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-cainjector.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-cainjector.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-approve:cert-manager-io.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-approve:cert-manager-io.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-certificates.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-certificates.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-certificatesigningrequests.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-certificatesigningrequests.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-challenges.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-challenges.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-clusterissuers.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-clusterissuers.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-ingress-shim.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-ingress-shim.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-issuers.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-issuers.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-orders.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-controller-orders.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-edit.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-edit.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-view.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-view.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-webhook:subjectaccessreviews.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cert-manager-webhook:subjectaccessreviews.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cilium-operator.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cilium-operator.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cilium.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cilium.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/cluster-admin.txt",
                                ">f+++++++++ objects/cluster/clusterrole/cluster-admin.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/edit.txt",
                                ">f+++++++++ objects/cluster/clusterrole/edit.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/envoy-gateway-gateway-helm-certgen:envoy-gateway-system.txt",
                                ">f+++++++++ objects/cluster/clusterrole/envoy-gateway-gateway-helm-certgen:envoy-gateway-system.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/envoy-gateway-gateway-helm-envoy-gateway-role.txt",
                                ">f+++++++++ objects/cluster/clusterrole/envoy-gateway-gateway-helm-envoy-gateway-role.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/goldpinger-clusterrole.txt",
                                ">f+++++++++ objects/cluster/clusterrole/goldpinger-clusterrole.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/ingress-nginx.txt",
                                ">f+++++++++ objects/cluster/clusterrole/ingress-nginx.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-grafana-clusterrole.txt",
                                ">f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-grafana-clusterrole.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-kube-state-metrics.txt",
                                ">f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-kube-state-metrics.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-operator.txt",
                                ">f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-operator.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-prometheus.txt",
                                ">f+++++++++ objects/cluster/clusterrole/kube-prometheus-stack-prometheus.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/kubeadm:get-nodes.txt",
                                ">f+++++++++ objects/cluster/clusterrole/kubeadm:get-nodes.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/local-path-provisioner.txt",
                                ">f+++++++++ objects/cluster/clusterrole/local-path-provisioner.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/loki-clusterrole.txt",
                                ">f+++++++++ objects/cluster/clusterrole/loki-clusterrole.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/node-feature-discovery.txt",
                                ">f+++++++++ objects/cluster/clusterrole/node-feature-discovery.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/rabbitmq-cluster-operator.txt",
                                ">f+++++++++ objects/cluster/clusterrole/rabbitmq-cluster-operator.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/rabbitmq-messaging-topology-operator.txt",
                                ">f+++++++++ objects/cluster/clusterrole/rabbitmq-messaging-topology-operator.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/secretgen-controller-cluster-role.txt",
                                ">f+++++++++ objects/cluster/clusterrole/secretgen-controller-cluster-role.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:aggregate-to-admin.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:aggregate-to-admin.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:aggregate-to-edit.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:aggregate-to-edit.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:aggregate-to-view.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:aggregate-to-view.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:auth-delegator.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:auth-delegator.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:basic-user.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:basic-user.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:certificatesigningrequests:nodeclient.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:certificatesigningrequests:nodeclient.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:certificatesigningrequests:selfnodeclient.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:certificatesigningrequests:selfnodeclient.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kube-apiserver-client-approver.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kube-apiserver-client-approver.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kube-apiserver-client-kubelet-approver.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kube-apiserver-client-kubelet-approver.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kubelet-serving-approver.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:kubelet-serving-approver.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:legacy-unknown-approver.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:certificates.k8s.io:legacy-unknown-approver.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:attachdetach-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:attachdetach-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:certificate-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:certificate-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:clusterrole-aggregation-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:clusterrole-aggregation-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:cronjob-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:cronjob-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:daemon-set-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:daemon-set-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:deployment-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:deployment-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:disruption-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:disruption-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:endpoint-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:endpoint-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:endpointslice-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:endpointslice-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:endpointslicemirroring-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:endpointslicemirroring-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:ephemeral-volume-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:ephemeral-volume-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:expand-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:expand-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:generic-garbage-collector.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:generic-garbage-collector.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:horizontal-pod-autoscaler.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:horizontal-pod-autoscaler.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:job-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:job-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:namespace-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:namespace-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:node-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:node-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:persistent-volume-binder.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:persistent-volume-binder.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:pod-garbage-collector.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:pod-garbage-collector.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:pv-protection-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:pv-protection-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:pvc-protection-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:pvc-protection-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:replicaset-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:replicaset-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:replication-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:replication-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:resourcequota-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:resourcequota-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:root-ca-cert-publisher.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:root-ca-cert-publisher.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:route-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:route-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:service-account-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:service-account-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:service-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:service-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:statefulset-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:statefulset-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:ttl-after-finished-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:ttl-after-finished-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:ttl-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:controller:ttl-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:coredns.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:coredns.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:discovery.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:discovery.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:heapster.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:heapster.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:kube-aggregator.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:kube-aggregator.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:kube-controller-manager.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:kube-controller-manager.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:kube-dns.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:kube-dns.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:kube-scheduler.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:kube-scheduler.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:kubelet-api-admin.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:kubelet-api-admin.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:monitoring.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:monitoring.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:node-bootstrapper.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:node-bootstrapper.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:node-problem-detector.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:node-problem-detector.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:node-proxier.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:node-proxier.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:node.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:node.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:persistent-volume-provisioner.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:persistent-volume-provisioner.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:public-info-viewer.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:public-info-viewer.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:service-account-issuer-discovery.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:service-account-issuer-discovery.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/system:volume-scheduler.txt",
                                ">f+++++++++ objects/cluster/clusterrole/system:volume-scheduler.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/vector.txt",
                                ">f+++++++++ objects/cluster/clusterrole/vector.yaml",
                                ">f+++++++++ objects/cluster/clusterrole/view.txt",
                                ">f+++++++++ objects/cluster/clusterrole/view.yaml",
                                "cd+++++++++ objects/cluster/clusterrolebinding/",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-cainjector.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-cainjector.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-approve:cert-manager-io.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-approve:cert-manager-io.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-certificates.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-certificates.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-certificatesigningrequests.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-certificatesigningrequests.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-challenges.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-challenges.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-clusterissuers.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-clusterissuers.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-ingress-shim.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-ingress-shim.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-issuers.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-issuers.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-orders.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-controller-orders.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-webhook:subjectaccessreviews.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cert-manager-webhook:subjectaccessreviews.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cilium-operator.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cilium-operator.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cilium.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cilium.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cluster-admin.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/cluster-admin.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/envoy-gateway-gateway-helm-certgen:envoy-gateway-system.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/envoy-gateway-gateway-helm-certgen:envoy-gateway-system.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/envoy-gateway-gateway-helm-envoy-gateway-rolebinding.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/envoy-gateway-gateway-helm-envoy-gateway-rolebinding.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/goldpinger-clusterrolebinding.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/goldpinger-clusterrolebinding.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/ingress-nginx.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/ingress-nginx.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-grafana-clusterrolebinding.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-grafana-clusterrolebinding.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-kube-state-metrics.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-kube-state-metrics.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-operator.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-operator.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-prometheus.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kube-prometheus-stack-prometheus.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kubeadm:get-nodes.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kubeadm:get-nodes.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kubeadm:kubelet-bootstrap.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kubeadm:kubelet-bootstrap.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-autoapprove-bootstrap.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-autoapprove-bootstrap.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-autoapprove-certificate-rotation.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-autoapprove-certificate-rotation.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-proxier.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/kubeadm:node-proxier.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/local-path-provisioner.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/local-path-provisioner.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/loki-clusterrolebinding.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/loki-clusterrolebinding.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/node-feature-discovery.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/node-feature-discovery.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/rabbitmq-cluster-operator.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/rabbitmq-cluster-operator.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/rabbitmq-messaging-topology-operator.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/rabbitmq-messaging-topology-operator.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/secretgen-controller-cluster-role-binding.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/secretgen-controller-cluster-role-binding.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:basic-user.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:basic-user.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:attachdetach-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:attachdetach-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:certificate-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:certificate-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:clusterrole-aggregation-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:clusterrole-aggregation-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:cronjob-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:cronjob-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:daemon-set-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:daemon-set-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:deployment-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:deployment-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:disruption-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:disruption-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpoint-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpoint-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpointslice-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpointslice-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpointslicemirroring-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:endpointslicemirroring-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:ephemeral-volume-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:ephemeral-volume-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:expand-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:expand-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:generic-garbage-collector.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:generic-garbage-collector.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:horizontal-pod-autoscaler.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:horizontal-pod-autoscaler.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:job-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:job-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:namespace-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:namespace-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:node-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:node-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:persistent-volume-binder.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:persistent-volume-binder.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:pod-garbage-collector.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:pod-garbage-collector.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:pv-protection-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:pv-protection-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:pvc-protection-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:pvc-protection-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:replicaset-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:replicaset-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:replication-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:replication-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:resourcequota-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:resourcequota-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:root-ca-cert-publisher.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:root-ca-cert-publisher.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:route-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:route-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:service-account-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:service-account-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:service-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:service-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:statefulset-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:statefulset-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:ttl-after-finished-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:ttl-after-finished-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:ttl-controller.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:controller:ttl-controller.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:coredns.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:coredns.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:discovery.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:discovery.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:kube-controller-manager.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:kube-controller-manager.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:kube-dns.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:kube-dns.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:kube-scheduler.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:kube-scheduler.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:monitoring.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:monitoring.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:node-proxier.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:node-proxier.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:node.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:node.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:public-info-viewer.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:public-info-viewer.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:service-account-issuer-discovery.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:service-account-issuer-discovery.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:volume-scheduler.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/system:volume-scheduler.yaml",
                                ">f+++++++++ objects/cluster/clusterrolebinding/vector.txt",
                                ">f+++++++++ objects/cluster/clusterrolebinding/vector.yaml",
                                "cd+++++++++ objects/cluster/namespace/",
                                ">f+++++++++ objects/cluster/namespace/auth-system.txt",
                                ">f+++++++++ objects/cluster/namespace/auth-system.yaml",
                                ">f+++++++++ objects/cluster/namespace/cert-manager.txt",
                                ">f+++++++++ objects/cluster/namespace/cert-manager.yaml",
                                ">f+++++++++ objects/cluster/namespace/default.txt",
                                ">f+++++++++ objects/cluster/namespace/default.yaml",
                                ">f+++++++++ objects/cluster/namespace/envoy-gateway-system.txt",
                                ">f+++++++++ objects/cluster/namespace/envoy-gateway-system.yaml",
                                ">f+++++++++ objects/cluster/namespace/ingress-nginx.txt",
                                ">f+++++++++ objects/cluster/namespace/ingress-nginx.yaml",
                                ">f+++++++++ objects/cluster/namespace/kube-node-lease.txt",
                                ">f+++++++++ objects/cluster/namespace/kube-node-lease.yaml",
                                ">f+++++++++ objects/cluster/namespace/kube-public.txt",
                                ">f+++++++++ objects/cluster/namespace/kube-public.yaml",
                                ">f+++++++++ objects/cluster/namespace/kube-system.txt",
                                ">f+++++++++ objects/cluster/namespace/kube-system.yaml",
                                ">f+++++++++ objects/cluster/namespace/local-path-storage.txt",
                                ">f+++++++++ objects/cluster/namespace/local-path-storage.yaml",
                                ">f+++++++++ objects/cluster/namespace/monitoring.txt",
                                ">f+++++++++ objects/cluster/namespace/monitoring.yaml",
                                ">f+++++++++ objects/cluster/namespace/openstack.txt",
                                ">f+++++++++ objects/cluster/namespace/openstack.yaml",
                                ">f+++++++++ objects/cluster/namespace/secretgen-controller.txt",
                                ">f+++++++++ objects/cluster/namespace/secretgen-controller.yaml",
                                "cd+++++++++ objects/cluster/node/",
                                ">f+++++++++ objects/cluster/node/instance.txt",
                                ">f+++++++++ objects/cluster/node/instance.yaml",
                                "cd+++++++++ objects/cluster/storageclass/",
                                ">f+++++++++ objects/cluster/storageclass/general.txt",
                                ">f+++++++++ objects/cluster/storageclass/general.yaml",
                                "cd+++++++++ objects/namespaced/",
                                "cd+++++++++ objects/namespaced/auth-system/",
                                "cd+++++++++ objects/namespaced/auth-system/configmaps/",
                                ">f+++++++++ objects/namespaced/auth-system/configmaps/keycloak-env-vars.txt",
                                ">f+++++++++ objects/namespaced/auth-system/configmaps/keycloak-env-vars.yaml",
                                ">f+++++++++ objects/namespaced/auth-system/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/auth-system/configmaps/kube-root-ca.crt.yaml",
                                "cd+++++++++ objects/namespaced/auth-system/endpoints/",
                                ">f+++++++++ objects/namespaced/auth-system/endpoints/keycloak-headless.txt",
                                ">f+++++++++ objects/namespaced/auth-system/endpoints/keycloak-headless.yaml",
                                ">f+++++++++ objects/namespaced/auth-system/endpoints/keycloak-metrics.txt",
                                ">f+++++++++ objects/namespaced/auth-system/endpoints/keycloak-metrics.yaml",
                                ">f+++++++++ objects/namespaced/auth-system/endpoints/keycloak.txt",
                                ">f+++++++++ objects/namespaced/auth-system/endpoints/keycloak.yaml",
                                "cd+++++++++ objects/namespaced/auth-system/ingresses/",
                                ">f+++++++++ objects/namespaced/auth-system/ingresses/keycloak.txt",
                                ">f+++++++++ objects/namespaced/auth-system/ingresses/keycloak.yaml",
                                "cd+++++++++ objects/namespaced/auth-system/pods/",
                                ">f+++++++++ objects/namespaced/auth-system/pods/keycloak-0.txt",
                                ">f+++++++++ objects/namespaced/auth-system/pods/keycloak-0.yaml",
                                "cd+++++++++ objects/namespaced/auth-system/secrets/",
                                ">f+++++++++ objects/namespaced/auth-system/secrets/keycloak-externaldb.txt",
                                ">f+++++++++ objects/namespaced/auth-system/secrets/keycloak-externaldb.yaml",
                                ">f+++++++++ objects/namespaced/auth-system/secrets/keycloak.199-204-45-153.nip.io-tls.txt",
                                ">f+++++++++ objects/namespaced/auth-system/secrets/keycloak.199-204-45-153.nip.io-tls.yaml",
                                ">f+++++++++ objects/namespaced/auth-system/secrets/keycloak.txt",
                                ">f+++++++++ objects/namespaced/auth-system/secrets/keycloak.yaml",
                                ">f+++++++++ objects/namespaced/auth-system/secrets/sh.helm.release.v1.keycloak.v1.txt",
                                ">f+++++++++ objects/namespaced/auth-system/secrets/sh.helm.release.v1.keycloak.v1.yaml",
                                "cd+++++++++ objects/namespaced/auth-system/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/auth-system/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/auth-system/serviceaccounts/default.yaml",
                                ">f+++++++++ objects/namespaced/auth-system/serviceaccounts/keycloak.txt",
                                ">f+++++++++ objects/namespaced/auth-system/serviceaccounts/keycloak.yaml",
                                "cd+++++++++ objects/namespaced/auth-system/services/",
                                ">f+++++++++ objects/namespaced/auth-system/services/keycloak-headless.txt",
                                ">f+++++++++ objects/namespaced/auth-system/services/keycloak-headless.yaml",
                                ">f+++++++++ objects/namespaced/auth-system/services/keycloak-metrics.txt",
                                ">f+++++++++ objects/namespaced/auth-system/services/keycloak-metrics.yaml",
                                ">f+++++++++ objects/namespaced/auth-system/services/keycloak.txt",
                                ">f+++++++++ objects/namespaced/auth-system/services/keycloak.yaml",
                                "cd+++++++++ objects/namespaced/auth-system/statefulsets/",
                                ">f+++++++++ objects/namespaced/auth-system/statefulsets/keycloak.txt",
                                ">f+++++++++ objects/namespaced/auth-system/statefulsets/keycloak.yaml",
                                "cd+++++++++ objects/namespaced/cert-manager/",
                                "cd+++++++++ objects/namespaced/cert-manager/configmaps/",
                                ">f+++++++++ objects/namespaced/cert-manager/configmaps/cert-manager-webhook.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/configmaps/cert-manager-webhook.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/configmaps/kube-root-ca.crt.yaml",
                                "cd+++++++++ objects/namespaced/cert-manager/deployment/",
                                ">f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager-cainjector.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager-cainjector.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager-webhook.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager-webhook.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/deployment/cert-manager.yaml",
                                "cd+++++++++ objects/namespaced/cert-manager/endpoints/",
                                ">f+++++++++ objects/namespaced/cert-manager/endpoints/cert-manager-webhook.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/endpoints/cert-manager-webhook.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/endpoints/cert-manager.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/endpoints/cert-manager.yaml",
                                "cd+++++++++ objects/namespaced/cert-manager/pods/",
                                ">f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-75969b45cf-d5c4v.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-75969b45cf-d5c4v.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-cainjector-84f45b698d-v64n5.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-cainjector-84f45b698d-v64n5.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-webhook-9997c64fc-8w8sz.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/pods/cert-manager-webhook-9997c64fc-8w8sz.yaml",
                                "cd+++++++++ objects/namespaced/cert-manager/rolebindings/",
                                ">f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager-cainjector:leaderelection.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager-cainjector:leaderelection.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager-webhook:dynamic-serving.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager-webhook:dynamic-serving.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager:leaderelection.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/rolebindings/cert-manager:leaderelection.yaml",
                                "cd+++++++++ objects/namespaced/cert-manager/roles/",
                                ">f+++++++++ objects/namespaced/cert-manager/roles/cert-manager-cainjector:leaderelection.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/roles/cert-manager-cainjector:leaderelection.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/roles/cert-manager-webhook:dynamic-serving.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/roles/cert-manager-webhook:dynamic-serving.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/roles/cert-manager:leaderelection.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/roles/cert-manager:leaderelection.yaml",
                                "cd+++++++++ objects/namespaced/cert-manager/secrets/",
                                ">f+++++++++ objects/namespaced/cert-manager/secrets/cert-manager-selfsigned-ca.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/secrets/cert-manager-selfsigned-ca.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/secrets/cert-manager-webhook-ca.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/secrets/cert-manager-webhook-ca.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/secrets/kube-prometheus-stack-ca.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/secrets/kube-prometheus-stack-ca.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/secrets/sh.helm.release.v1.cert-manager.v1.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/secrets/sh.helm.release.v1.cert-manager.v1.yaml",
                                "cd+++++++++ objects/namespaced/cert-manager/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager-cainjector.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager-cainjector.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager-webhook.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager-webhook.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/serviceaccounts/cert-manager.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/serviceaccounts/default.yaml",
                                "cd+++++++++ objects/namespaced/cert-manager/services/",
                                ">f+++++++++ objects/namespaced/cert-manager/services/cert-manager-webhook.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/services/cert-manager-webhook.yaml",
                                ">f+++++++++ objects/namespaced/cert-manager/services/cert-manager.txt",
                                ">f+++++++++ objects/namespaced/cert-manager/services/cert-manager.yaml",
                                "cd+++++++++ objects/namespaced/default/",
                                "cd+++++++++ objects/namespaced/default/configmaps/",
                                ">f+++++++++ objects/namespaced/default/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/default/configmaps/kube-root-ca.crt.yaml",
                                "cd+++++++++ objects/namespaced/default/endpoints/",
                                ">f+++++++++ objects/namespaced/default/endpoints/kubernetes.txt",
                                ">f+++++++++ objects/namespaced/default/endpoints/kubernetes.yaml",
                                "cd+++++++++ objects/namespaced/default/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/default/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/default/serviceaccounts/default.yaml",
                                "cd+++++++++ objects/namespaced/default/services/",
                                ">f+++++++++ objects/namespaced/default/services/kubernetes.txt",
                                ">f+++++++++ objects/namespaced/default/services/kubernetes.yaml",
                                "cd+++++++++ objects/namespaced/envoy-gateway-system/",
                                "cd+++++++++ objects/namespaced/envoy-gateway-system/configmaps/",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/configmaps/envoy-gateway-config.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/configmaps/envoy-gateway-config.yaml",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/configmaps/kube-root-ca.crt.yaml",
                                "cd+++++++++ objects/namespaced/envoy-gateway-system/deployment/",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/deployment/envoy-gateway.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/deployment/envoy-gateway.yaml",
                                "cd+++++++++ objects/namespaced/envoy-gateway-system/endpoints/",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/endpoints/envoy-gateway.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/endpoints/envoy-gateway.yaml",
                                "cd+++++++++ objects/namespaced/envoy-gateway-system/pods/",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/pods/envoy-gateway-78446f96c9-7zvgb.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/pods/envoy-gateway-78446f96c9-7zvgb.yaml",
                                "cd+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-certgen.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-certgen.yaml",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-infra-manager.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-infra-manager.yaml",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-leader-election-rolebinding.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/rolebindings/envoy-gateway-gateway-helm-leader-election-rolebinding.yaml",
                                "cd+++++++++ objects/namespaced/envoy-gateway-system/roles/",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-certgen.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-certgen.yaml",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-infra-manager.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-infra-manager.yaml",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-leader-election-role.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/roles/envoy-gateway-gateway-helm-leader-election-role.yaml",
                                "cd+++++++++ objects/namespaced/envoy-gateway-system/secrets/",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-gateway.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-gateway.yaml",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-oidc-hmac.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-oidc-hmac.yaml",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-rate-limit.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy-rate-limit.yaml",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/secrets/envoy.yaml",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/secrets/sh.helm.release.v1.envoy-gateway.v1.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/secrets/sh.helm.release.v1.envoy-gateway.v1.yaml",
                                "cd+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/default.yaml",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/envoy-gateway-gateway-helm-certgen.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/envoy-gateway-gateway-helm-certgen.yaml",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/envoy-gateway.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/serviceaccounts/envoy-gateway.yaml",
                                "cd+++++++++ objects/namespaced/envoy-gateway-system/services/",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/services/envoy-gateway.txt",
                                ">f+++++++++ objects/namespaced/envoy-gateway-system/services/envoy-gateway.yaml",
                                "cd+++++++++ objects/namespaced/ingress-nginx/",
                                "cd+++++++++ objects/namespaced/ingress-nginx/configmaps/",
                                ">f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-controller.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-controller.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-tcp.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-tcp.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-udp.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/configmaps/ingress-nginx-udp.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/configmaps/kube-root-ca.crt.yaml",
                                "cd+++++++++ objects/namespaced/ingress-nginx/daemonsets/",
                                ">f+++++++++ objects/namespaced/ingress-nginx/daemonsets/ingress-nginx-controller.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/daemonsets/ingress-nginx-controller.yaml",
                                "cd+++++++++ objects/namespaced/ingress-nginx/deployment/",
                                ">f+++++++++ objects/namespaced/ingress-nginx/deployment/ingress-nginx-defaultbackend.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/deployment/ingress-nginx-defaultbackend.yaml",
                                "cd+++++++++ objects/namespaced/ingress-nginx/endpoints/",
                                ">f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller-admission.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller-admission.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller-metrics.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller-metrics.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-controller.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-defaultbackend.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/endpoints/ingress-nginx-defaultbackend.yaml",
                                "cd+++++++++ objects/namespaced/ingress-nginx/pods/",
                                ">f+++++++++ objects/namespaced/ingress-nginx/pods/ingress-nginx-controller-dmjt4.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/pods/ingress-nginx-controller-dmjt4.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/pods/ingress-nginx-defaultbackend-6987ff55cf-plsk9.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/pods/ingress-nginx-defaultbackend-6987ff55cf-plsk9.yaml",
                                "cd+++++++++ objects/namespaced/ingress-nginx/rolebindings/",
                                ">f+++++++++ objects/namespaced/ingress-nginx/rolebindings/ingress-nginx.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/rolebindings/ingress-nginx.yaml",
                                "cd+++++++++ objects/namespaced/ingress-nginx/roles/",
                                ">f+++++++++ objects/namespaced/ingress-nginx/roles/ingress-nginx.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/roles/ingress-nginx.yaml",
                                "cd+++++++++ objects/namespaced/ingress-nginx/secrets/",
                                ">f+++++++++ objects/namespaced/ingress-nginx/secrets/ingress-nginx-admission.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/secrets/ingress-nginx-admission.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/secrets/sh.helm.release.v1.ingress-nginx.v1.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/secrets/sh.helm.release.v1.ingress-nginx.v1.yaml",
                                "cd+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/default.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/ingress-nginx-backend.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/ingress-nginx-backend.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/ingress-nginx.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/serviceaccounts/ingress-nginx.yaml",
                                "cd+++++++++ objects/namespaced/ingress-nginx/services/",
                                ">f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller-admission.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller-admission.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller-metrics.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller-metrics.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-controller.yaml",
                                ">f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-defaultbackend.txt",
                                ">f+++++++++ objects/namespaced/ingress-nginx/services/ingress-nginx-defaultbackend.yaml",
                                "cd+++++++++ objects/namespaced/kube-node-lease/",
                                "cd+++++++++ objects/namespaced/kube-node-lease/configmaps/",
                                ">f+++++++++ objects/namespaced/kube-node-lease/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/kube-node-lease/configmaps/kube-root-ca.crt.yaml",
                                "cd+++++++++ objects/namespaced/kube-node-lease/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/kube-node-lease/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/kube-node-lease/serviceaccounts/default.yaml",
                                "cd+++++++++ objects/namespaced/kube-public/",
                                "cd+++++++++ objects/namespaced/kube-public/configmaps/",
                                ">f+++++++++ objects/namespaced/kube-public/configmaps/cluster-info.txt",
                                ">f+++++++++ objects/namespaced/kube-public/configmaps/cluster-info.yaml",
                                ">f+++++++++ objects/namespaced/kube-public/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/kube-public/configmaps/kube-root-ca.crt.yaml",
                                "cd+++++++++ objects/namespaced/kube-public/rolebindings/",
                                ">f+++++++++ objects/namespaced/kube-public/rolebindings/kubeadm:bootstrap-signer-clusterinfo.txt",
                                ">f+++++++++ objects/namespaced/kube-public/rolebindings/kubeadm:bootstrap-signer-clusterinfo.yaml",
                                ">f+++++++++ objects/namespaced/kube-public/rolebindings/system:controller:bootstrap-signer.txt",
                                ">f+++++++++ objects/namespaced/kube-public/rolebindings/system:controller:bootstrap-signer.yaml",
                                "cd+++++++++ objects/namespaced/kube-public/roles/",
                                ">f+++++++++ objects/namespaced/kube-public/roles/kubeadm:bootstrap-signer-clusterinfo.txt",
                                ">f+++++++++ objects/namespaced/kube-public/roles/kubeadm:bootstrap-signer-clusterinfo.yaml",
                                ">f+++++++++ objects/namespaced/kube-public/roles/system:controller:bootstrap-signer.txt",
                                ">f+++++++++ objects/namespaced/kube-public/roles/system:controller:bootstrap-signer.yaml",
                                "cd+++++++++ objects/namespaced/kube-public/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/kube-public/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/kube-public/serviceaccounts/default.yaml",
                                "cd+++++++++ objects/namespaced/kube-system/",
                                "cd+++++++++ objects/namespaced/kube-system/configmaps/",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/cilium-config.txt",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/cilium-config.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/coredns.txt",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/coredns.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/extension-apiserver-authentication.txt",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/extension-apiserver-authentication.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/kube-apiserver-legacy-service-account-token-tracking.txt",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/kube-apiserver-legacy-service-account-token-tracking.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/kube-proxy.txt",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/kube-proxy.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/kube-root-ca.crt.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/kubeadm-config.txt",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/kubeadm-config.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/kubelet-config.txt",
                                ">f+++++++++ objects/namespaced/kube-system/configmaps/kubelet-config.yaml",
                                "cd+++++++++ objects/namespaced/kube-system/daemonsets/",
                                ">f+++++++++ objects/namespaced/kube-system/daemonsets/cilium.txt",
                                ">f+++++++++ objects/namespaced/kube-system/daemonsets/cilium.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/daemonsets/kube-proxy.txt",
                                ">f+++++++++ objects/namespaced/kube-system/daemonsets/kube-proxy.yaml",
                                "cd+++++++++ objects/namespaced/kube-system/deployment/",
                                ">f+++++++++ objects/namespaced/kube-system/deployment/cilium-operator.txt",
                                ">f+++++++++ objects/namespaced/kube-system/deployment/cilium-operator.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/deployment/coredns.txt",
                                ">f+++++++++ objects/namespaced/kube-system/deployment/coredns.yaml",
                                "cd+++++++++ objects/namespaced/kube-system/endpoints/",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-dns.txt",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-dns.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-coredns.txt",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-coredns.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-controller-manager.txt",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-controller-manager.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-etcd.txt",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-etcd.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-proxy.txt",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-proxy.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-scheduler.txt",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kube-scheduler.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kubelet.txt",
                                ">f+++++++++ objects/namespaced/kube-system/endpoints/kube-prometheus-stack-kubelet.yaml",
                                "cd+++++++++ objects/namespaced/kube-system/pods/",
                                ">f+++++++++ objects/namespaced/kube-system/pods/cilium-operator-869df985b8-nqvtb.txt",
                                ">f+++++++++ objects/namespaced/kube-system/pods/cilium-operator-869df985b8-nqvtb.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/pods/cilium-sk6n5.txt",
                                ">f+++++++++ objects/namespaced/kube-system/pods/cilium-sk6n5.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/pods/coredns-67659f764b-mdrt6.txt",
                                ">f+++++++++ objects/namespaced/kube-system/pods/coredns-67659f764b-mdrt6.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/pods/coredns-67659f764b-xb6lv.txt",
                                ">f+++++++++ objects/namespaced/kube-system/pods/coredns-67659f764b-xb6lv.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/pods/etcd-instance.txt",
                                ">f+++++++++ objects/namespaced/kube-system/pods/etcd-instance.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/pods/kube-apiserver-instance.txt",
                                ">f+++++++++ objects/namespaced/kube-system/pods/kube-apiserver-instance.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/pods/kube-controller-manager-instance.txt",
                                ">f+++++++++ objects/namespaced/kube-system/pods/kube-controller-manager-instance.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/pods/kube-proxy-cxlxv.txt",
                                ">f+++++++++ objects/namespaced/kube-system/pods/kube-proxy-cxlxv.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/pods/kube-scheduler-instance.txt",
                                ">f+++++++++ objects/namespaced/kube-system/pods/kube-scheduler-instance.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/pods/kube-vip-instance.txt",
                                ">f+++++++++ objects/namespaced/kube-system/pods/kube-vip-instance.yaml",
                                "cd+++++++++ objects/namespaced/kube-system/rolebindings/",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/cilium-config-agent.txt",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/cilium-config-agent.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/kube-proxy.txt",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/kube-proxy.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:kubeadm-certs.txt",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:kubeadm-certs.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:kubelet-config.txt",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:kubelet-config.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:nodes-kubeadm-config.txt",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/kubeadm:nodes-kubeadm-config.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system::extension-apiserver-authentication-reader.txt",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system::extension-apiserver-authentication-reader.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system::leader-locking-kube-controller-manager.txt",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system::leader-locking-kube-controller-manager.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system::leader-locking-kube-scheduler.txt",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system::leader-locking-kube-scheduler.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:bootstrap-signer.txt",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:bootstrap-signer.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:cloud-provider.txt",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:cloud-provider.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:token-cleaner.txt",
                                ">f+++++++++ objects/namespaced/kube-system/rolebindings/system:controller:token-cleaner.yaml",
                                "cd+++++++++ objects/namespaced/kube-system/roles/",
                                ">f+++++++++ objects/namespaced/kube-system/roles/cilium-config-agent.txt",
                                ">f+++++++++ objects/namespaced/kube-system/roles/cilium-config-agent.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/roles/extension-apiserver-authentication-reader.txt",
                                ">f+++++++++ objects/namespaced/kube-system/roles/extension-apiserver-authentication-reader.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/roles/kube-proxy.txt",
                                ">f+++++++++ objects/namespaced/kube-system/roles/kube-proxy.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/roles/kubeadm:kubeadm-certs.txt",
                                ">f+++++++++ objects/namespaced/kube-system/roles/kubeadm:kubeadm-certs.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/roles/kubeadm:kubelet-config.txt",
                                ">f+++++++++ objects/namespaced/kube-system/roles/kubeadm:kubelet-config.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/roles/kubeadm:nodes-kubeadm-config.txt",
                                ">f+++++++++ objects/namespaced/kube-system/roles/kubeadm:nodes-kubeadm-config.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/roles/system::leader-locking-kube-controller-manager.txt",
                                ">f+++++++++ objects/namespaced/kube-system/roles/system::leader-locking-kube-controller-manager.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/roles/system::leader-locking-kube-scheduler.txt",
                                ">f+++++++++ objects/namespaced/kube-system/roles/system::leader-locking-kube-scheduler.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/roles/system:controller:bootstrap-signer.txt",
                                ">f+++++++++ objects/namespaced/kube-system/roles/system:controller:bootstrap-signer.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/roles/system:controller:cloud-provider.txt",
                                ">f+++++++++ objects/namespaced/kube-system/roles/system:controller:cloud-provider.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/roles/system:controller:token-cleaner.txt",
                                ">f+++++++++ objects/namespaced/kube-system/roles/system:controller:token-cleaner.yaml",
                                "cd+++++++++ objects/namespaced/kube-system/secrets/",
                                ">f+++++++++ objects/namespaced/kube-system/secrets/bootstrap-token-3n8qak.txt",
                                ">f+++++++++ objects/namespaced/kube-system/secrets/bootstrap-token-3n8qak.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/secrets/bootstrap-token-w4o225.txt",
                                ">f+++++++++ objects/namespaced/kube-system/secrets/bootstrap-token-w4o225.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/secrets/kubeadm-certs.txt",
                                ">f+++++++++ objects/namespaced/kube-system/secrets/kubeadm-certs.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/secrets/sh.helm.release.v1.cilium.v1.txt",
                                ">f+++++++++ objects/namespaced/kube-system/secrets/sh.helm.release.v1.cilium.v1.yaml",
                                "cd+++++++++ objects/namespaced/kube-system/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/attachdetach-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/attachdetach-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/bootstrap-signer.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/bootstrap-signer.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/certificate-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/certificate-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/cilium-operator.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/cilium-operator.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/cilium.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/cilium.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/clusterrole-aggregation-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/clusterrole-aggregation-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/coredns.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/coredns.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/cronjob-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/cronjob-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/daemon-set-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/daemon-set-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/default.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/deployment-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/deployment-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/disruption-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/disruption-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpoint-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpoint-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpointslice-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpointslice-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpointslicemirroring-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/endpointslicemirroring-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/ephemeral-volume-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/ephemeral-volume-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/expand-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/expand-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/generic-garbage-collector.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/generic-garbage-collector.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/horizontal-pod-autoscaler.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/horizontal-pod-autoscaler.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/job-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/job-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/kube-proxy.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/kube-proxy.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/namespace-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/namespace-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/node-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/node-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/persistent-volume-binder.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/persistent-volume-binder.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/pod-garbage-collector.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/pod-garbage-collector.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/pv-protection-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/pv-protection-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/pvc-protection-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/pvc-protection-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/replicaset-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/replicaset-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/replication-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/replication-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/resourcequota-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/resourcequota-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/root-ca-cert-publisher.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/root-ca-cert-publisher.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/service-account-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/service-account-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/service-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/service-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/statefulset-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/statefulset-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/token-cleaner.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/token-cleaner.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/ttl-after-finished-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/ttl-after-finished-controller.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/ttl-controller.txt",
                                ">f+++++++++ objects/namespaced/kube-system/serviceaccounts/ttl-controller.yaml",
                                "cd+++++++++ objects/namespaced/kube-system/services/",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-dns.txt",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-dns.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-coredns.txt",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-coredns.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-controller-manager.txt",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-controller-manager.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-etcd.txt",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-etcd.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-proxy.txt",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-proxy.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-scheduler.txt",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kube-scheduler.yaml",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kubelet.txt",
                                ">f+++++++++ objects/namespaced/kube-system/services/kube-prometheus-stack-kubelet.yaml",
                                "cd+++++++++ objects/namespaced/local-path-storage/",
                                "cd+++++++++ objects/namespaced/local-path-storage/configmaps/",
                                ">f+++++++++ objects/namespaced/local-path-storage/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/local-path-storage/configmaps/kube-root-ca.crt.yaml",
                                ">f+++++++++ objects/namespaced/local-path-storage/configmaps/local-path-config.txt",
                                ">f+++++++++ objects/namespaced/local-path-storage/configmaps/local-path-config.yaml",
                                "cd+++++++++ objects/namespaced/local-path-storage/deployment/",
                                ">f+++++++++ objects/namespaced/local-path-storage/deployment/local-path-provisioner.txt",
                                ">f+++++++++ objects/namespaced/local-path-storage/deployment/local-path-provisioner.yaml",
                                "cd+++++++++ objects/namespaced/local-path-storage/pods/",
                                ">f+++++++++ objects/namespaced/local-path-storage/pods/local-path-provisioner-679c578f5-m52kp.txt",
                                ">f+++++++++ objects/namespaced/local-path-storage/pods/local-path-provisioner-679c578f5-m52kp.yaml",
                                "cd+++++++++ objects/namespaced/local-path-storage/secrets/",
                                ">f+++++++++ objects/namespaced/local-path-storage/secrets/sh.helm.release.v1.local-path-provisioner.v1.txt",
                                ">f+++++++++ objects/namespaced/local-path-storage/secrets/sh.helm.release.v1.local-path-provisioner.v1.yaml",
                                "cd+++++++++ objects/namespaced/local-path-storage/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/local-path-storage/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/local-path-storage/serviceaccounts/default.yaml",
                                ">f+++++++++ objects/namespaced/local-path-storage/serviceaccounts/local-path-provisioner.txt",
                                ">f+++++++++ objects/namespaced/local-path-storage/serviceaccounts/local-path-provisioner.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/",
                                "cd+++++++++ objects/namespaced/monitoring/configmaps/",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/goldpinger-zap.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/goldpinger-zap.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/ipmi-exporter.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/ipmi-exporter.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-alertmanager-overview.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-alertmanager-overview.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-apiserver.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-apiserver.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-cluster-total.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-cluster-total.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-controller-manager.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-controller-manager.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster-advanced.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster-advanced.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-ceph-cluster.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-goldpinger.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-goldpinger.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-haproxy.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-haproxy.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-host-details.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-host-details.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-hosts-overview.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-hosts-overview.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-node-exporter-full.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-node-exporter-full.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-osd-device-details.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-osd-device-details.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-osds-overview.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-osds-overview.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-pool-detail.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-pool-detail.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-pool-overview.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-pool-overview.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-rbd-details.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-rbd-details.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-rbd-overview.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-dashboard-rbd-overview.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-etcd.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-etcd.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-config-dashboards.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-config-dashboards.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-datasource.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-datasource.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-overview.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana-overview.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-grafana.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-coredns.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-coredns.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-cluster.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-cluster.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-multicluster.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-multicluster.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-namespace.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-namespace.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-node.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-node.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-pod.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-pod.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-workload.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-workload.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-workloads-namespace.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-k8s-resources-workloads-namespace.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-kubelet.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-kubelet.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-namespace-by-pod.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-namespace-by-pod.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-namespace-by-workload.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-namespace-by-workload.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-cluster-rsrc-use.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-cluster-rsrc-use.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-exporter.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-exporter.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-rsrc-use.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-node-rsrc-use.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-nodes-darwin.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-nodes-darwin.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-nodes.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-nodes.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-persistentvolumesusage.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-persistentvolumesusage.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-pod-total.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-pod-total.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-prometheus-tls.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-prometheus-tls.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-prometheus.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-prometheus.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-proxy.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-proxy.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-scheduler.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-scheduler.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-workload-total.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-prometheus-stack-workload-total.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/kube-root-ca.crt.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki-alerting-rules.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki-alerting-rules.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki-dashboards-1.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki-dashboards-1.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki-dashboards-2.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki-dashboards-2.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki-gateway.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki-gateway.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki-runtime.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki-runtime.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/loki.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/node-feature-discovery-worker-conf.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/node-feature-discovery-worker-conf.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/prometheus-kube-prometheus-stack-prometheus-rulefiles-0.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/prometheus-kube-prometheus-stack-prometheus-rulefiles-0.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/vector.txt",
                                ">f+++++++++ objects/namespaced/monitoring/configmaps/vector.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/daemonsets/",
                                ">f+++++++++ objects/namespaced/monitoring/daemonsets/goldpinger.txt",
                                ">f+++++++++ objects/namespaced/monitoring/daemonsets/goldpinger.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/daemonsets/ipmi-exporter.txt",
                                ">f+++++++++ objects/namespaced/monitoring/daemonsets/ipmi-exporter.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/daemonsets/kube-prometheus-stack-prometheus-node-exporter.txt",
                                ">f+++++++++ objects/namespaced/monitoring/daemonsets/kube-prometheus-stack-prometheus-node-exporter.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/daemonsets/node-feature-discovery-worker.txt",
                                ">f+++++++++ objects/namespaced/monitoring/daemonsets/node-feature-discovery-worker.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/daemonsets/vector.txt",
                                ">f+++++++++ objects/namespaced/monitoring/daemonsets/vector.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/deployment/",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-grafana.txt",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-grafana.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-kube-state-metrics.txt",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-kube-state-metrics.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-operator.txt",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/kube-prometheus-stack-operator.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/loki-gateway.txt",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/loki-gateway.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/node-feature-discovery-master.txt",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/node-feature-discovery-master.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/prometheus-pushgateway.txt",
                                ">f+++++++++ objects/namespaced/monitoring/deployment/prometheus-pushgateway.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/endpoints/",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/alertmanager-operated.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/alertmanager-operated.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/goldpinger.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/goldpinger.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-alertmanager.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-alertmanager.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-grafana.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-grafana.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-kube-state-metrics.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-kube-state-metrics.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-operator.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-operator.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-prometheus-node-exporter.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-prometheus-node-exporter.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-prometheus.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/kube-prometheus-stack-prometheus.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/loki-gateway.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/loki-gateway.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/loki-headless.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/loki-headless.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/loki-memberlist.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/loki-memberlist.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/loki.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/loki.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/node-feature-discovery-master.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/node-feature-discovery-master.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/prometheus-operated.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/prometheus-operated.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/prometheus-pushgateway.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/prometheus-pushgateway.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/vector-headless.txt",
                                ">f+++++++++ objects/namespaced/monitoring/endpoints/vector-headless.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/ingresses/",
                                ">f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-alertmanager.txt",
                                ">f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-alertmanager.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-grafana.txt",
                                ">f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-grafana.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-prometheus.txt",
                                ">f+++++++++ objects/namespaced/monitoring/ingresses/kube-prometheus-stack-prometheus.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/",
                                ">f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/alertmanager-kube-prometheus-stack-alertmanager-db-alertmanager-kube-prometheus-stack-alertmanager-0.txt",
                                ">f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/alertmanager-kube-prometheus-stack-alertmanager-db-alertmanager-kube-prometheus-stack-alertmanager-0.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/prometheus-kube-prometheus-stack-prometheus-db-prometheus-kube-prometheus-stack-prometheus-0.txt",
                                ">f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/prometheus-kube-prometheus-stack-prometheus-db-prometheus-kube-prometheus-stack-prometheus-0.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/storage-loki-0.txt",
                                ">f+++++++++ objects/namespaced/monitoring/persistentvolumeclaims/storage-loki-0.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/pods/",
                                ">f+++++++++ objects/namespaced/monitoring/pods/alertmanager-kube-prometheus-stack-alertmanager-0.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/alertmanager-kube-prometheus-stack-alertmanager-0.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/goldpinger-gz8xd.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/goldpinger-gz8xd.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-grafana-69fc56495c-5h4gd.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-grafana-69fc56495c-5h4gd.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-operator-cc88b86d6-xspgk.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-operator-cc88b86d6-xspgk.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-prometheus-node-exporter-jmj6n.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/kube-prometheus-stack-prometheus-node-exporter-jmj6n.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/loki-0.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/loki-0.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/loki-gateway-5cc45946cd-wp7ks.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/loki-gateway-5cc45946cd-wp7ks.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/node-feature-discovery-master-68cf466d5f-qknzk.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/node-feature-discovery-master-68cf466d5f-qknzk.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/node-feature-discovery-worker-7l5z7.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/node-feature-discovery-worker-7l5z7.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/prometheus-kube-prometheus-stack-prometheus-0.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/prometheus-kube-prometheus-stack-prometheus-0.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/prometheus-pushgateway-5c694d99cc-jnjt7.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/prometheus-pushgateway-5c694d99cc-jnjt7.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/pods/vector-75jgf.txt",
                                ">f+++++++++ objects/namespaced/monitoring/pods/vector-75jgf.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/rolebindings/",
                                ">f+++++++++ objects/namespaced/monitoring/rolebindings/kube-prometheus-stack-grafana.txt",
                                ">f+++++++++ objects/namespaced/monitoring/rolebindings/kube-prometheus-stack-grafana.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/rolebindings/kube-prometheus-stack-pod-tls-sidecar.txt",
                                ">f+++++++++ objects/namespaced/monitoring/rolebindings/kube-prometheus-stack-pod-tls-sidecar.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/roles/",
                                ">f+++++++++ objects/namespaced/monitoring/roles/kube-prometheus-stack-grafana.txt",
                                ">f+++++++++ objects/namespaced/monitoring/roles/kube-prometheus-stack-grafana.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/roles/kube-prometheus-stack-pod-tls-sidecar.txt",
                                ">f+++++++++ objects/namespaced/monitoring/roles/kube-prometheus-stack-pod-tls-sidecar.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/secrets/",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-generated.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-generated.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-tls-assets-0.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-tls-assets-0.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-web-config.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager-web-config.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-kube-prometheus-stack-alertmanager.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-tls.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/alertmanager-tls.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/grafana-tls.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/grafana-tls.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-admission.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-admission.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-client-secret.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-client-secret.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-cookie-secret.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-cookie-secret.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-oauth2-proxy.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-alertmanager-oauth2-proxy.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-etcd-client-cert.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-etcd-client-cert.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-grafana-client-secret.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-grafana-client-secret.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-grafana.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-grafana.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-client-secret.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-client-secret.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-cookie-secret.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-cookie-secret.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-node-exporter-jmj6n-tls.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-node-exporter-jmj6n-tls.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-oauth2-proxy.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/kube-prometheus-stack-prometheus-oauth2-proxy.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-0-tls.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-0-tls.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-tls-assets-0.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-tls-assets-0.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-web-config.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus-web-config.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/prometheus-kube-prometheus-stack-prometheus.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/prometheus-tls.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/prometheus-tls.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.goldpinger.v1.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.goldpinger.v1.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.kube-prometheus-stack.v1.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.kube-prometheus-stack.v1.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.loki.v1.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.loki.v1.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.node-feature-discovery.v1.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.node-feature-discovery.v1.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.prometheus-pushgateway.v1.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.prometheus-pushgateway.v1.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.vector.v1.txt",
                                ">f+++++++++ objects/namespaced/monitoring/secrets/sh.helm.release.v1.vector.v1.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/default.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/goldpinger.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/goldpinger.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-alertmanager.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-alertmanager.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-grafana.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-grafana.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-kube-state-metrics.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-kube-state-metrics.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-operator.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-operator.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-prometheus-node-exporter.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-prometheus-node-exporter.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-prometheus.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/kube-prometheus-stack-prometheus.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/loki.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/loki.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/node-feature-discovery-worker.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/node-feature-discovery-worker.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/node-feature-discovery.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/node-feature-discovery.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/prometheus-pushgateway.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/prometheus-pushgateway.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/vector.txt",
                                ">f+++++++++ objects/namespaced/monitoring/serviceaccounts/vector.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/services/",
                                ">f+++++++++ objects/namespaced/monitoring/services/alertmanager-operated.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/alertmanager-operated.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/goldpinger.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/goldpinger.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-alertmanager.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-alertmanager.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-grafana.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-grafana.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-kube-state-metrics.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-kube-state-metrics.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-operator.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-operator.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-prometheus-node-exporter.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-prometheus-node-exporter.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-prometheus.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/kube-prometheus-stack-prometheus.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/loki-gateway.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/loki-gateway.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/loki-headless.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/loki-headless.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/loki-memberlist.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/loki-memberlist.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/loki.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/loki.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/node-feature-discovery-master.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/node-feature-discovery-master.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/prometheus-operated.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/prometheus-operated.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/prometheus-pushgateway.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/prometheus-pushgateway.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/services/vector-headless.txt",
                                ">f+++++++++ objects/namespaced/monitoring/services/vector-headless.yaml",
                                "cd+++++++++ objects/namespaced/monitoring/statefulsets/",
                                ">f+++++++++ objects/namespaced/monitoring/statefulsets/alertmanager-kube-prometheus-stack-alertmanager.txt",
                                ">f+++++++++ objects/namespaced/monitoring/statefulsets/alertmanager-kube-prometheus-stack-alertmanager.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/statefulsets/loki.txt",
                                ">f+++++++++ objects/namespaced/monitoring/statefulsets/loki.yaml",
                                ">f+++++++++ objects/namespaced/monitoring/statefulsets/prometheus-kube-prometheus-stack-prometheus.txt",
                                ">f+++++++++ objects/namespaced/monitoring/statefulsets/prometheus-kube-prometheus-stack-prometheus.yaml",
                                "cd+++++++++ objects/namespaced/openstack/",
                                "cd+++++++++ objects/namespaced/openstack/configmaps/",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/keepalived-bin.txt",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/keepalived-bin.yaml",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/kube-root-ca.crt.yaml",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/memcached-memcached-bin.txt",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/memcached-memcached-bin.yaml",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/messaging-topology-operator-leader-election.txt",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/messaging-topology-operator-leader-election.yaml",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/percona-xtradb-haproxy.txt",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/percona-xtradb-haproxy.yaml",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/percona-xtradb-pxc.txt",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/percona-xtradb-pxc.yaml",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/rabbitmq-cluster-operator-leader-election.txt",
                                ">f+++++++++ objects/namespaced/openstack/configmaps/rabbitmq-cluster-operator-leader-election.yaml",
                                "cd+++++++++ objects/namespaced/openstack/daemonsets/",
                                ">f+++++++++ objects/namespaced/openstack/daemonsets/keepalived.txt",
                                ">f+++++++++ objects/namespaced/openstack/daemonsets/keepalived.yaml",
                                "cd+++++++++ objects/namespaced/openstack/deployment/",
                                ">f+++++++++ objects/namespaced/openstack/deployment/memcached-memcached.txt",
                                ">f+++++++++ objects/namespaced/openstack/deployment/memcached-memcached.yaml",
                                ">f+++++++++ objects/namespaced/openstack/deployment/pxc-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/deployment/pxc-operator.yaml",
                                ">f+++++++++ objects/namespaced/openstack/deployment/rabbitmq-cluster-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/deployment/rabbitmq-cluster-operator.yaml",
                                ">f+++++++++ objects/namespaced/openstack/deployment/rabbitmq-messaging-topology-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/deployment/rabbitmq-messaging-topology-operator.yaml",
                                "cd+++++++++ objects/namespaced/openstack/endpoints/",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/memcached-metrics.txt",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/memcached-metrics.yaml",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/memcached.txt",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/memcached.yaml",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-cluster-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-cluster-operator.yaml",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy-metrics.txt",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy-metrics.yaml",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy-replicas.txt",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy-replicas.yaml",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy.txt",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-haproxy.yaml",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-pxc-unready.txt",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-pxc-unready.yaml",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-pxc.txt",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/percona-xtradb-pxc.yaml",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/rabbitmq-messaging-topology-operator-webhook.txt",
                                ">f+++++++++ objects/namespaced/openstack/endpoints/rabbitmq-messaging-topology-operator-webhook.yaml",
                                "cd+++++++++ objects/namespaced/openstack/persistentvolumeclaims/",
                                ">f+++++++++ objects/namespaced/openstack/persistentvolumeclaims/datadir-percona-xtradb-pxc-0.txt",
                                ">f+++++++++ objects/namespaced/openstack/persistentvolumeclaims/datadir-percona-xtradb-pxc-0.yaml",
                                "cd+++++++++ objects/namespaced/openstack/pods/",
                                ">f+++++++++ objects/namespaced/openstack/pods/keepalived-wktjf.txt",
                                ">f+++++++++ objects/namespaced/openstack/pods/keepalived-wktjf.yaml",
                                ">f+++++++++ objects/namespaced/openstack/pods/memcached-memcached-c47d949f5-vc5cm.txt",
                                ">f+++++++++ objects/namespaced/openstack/pods/memcached-memcached-c47d949f5-vc5cm.yaml",
                                ">f+++++++++ objects/namespaced/openstack/pods/percona-xtradb-haproxy-0.txt",
                                ">f+++++++++ objects/namespaced/openstack/pods/percona-xtradb-haproxy-0.yaml",
                                ">f+++++++++ objects/namespaced/openstack/pods/percona-xtradb-pxc-0.txt",
                                ">f+++++++++ objects/namespaced/openstack/pods/percona-xtradb-pxc-0.yaml",
                                ">f+++++++++ objects/namespaced/openstack/pods/pxc-operator-59d66cbc78-clxvf.txt",
                                ">f+++++++++ objects/namespaced/openstack/pods/pxc-operator-59d66cbc78-clxvf.yaml",
                                ">f+++++++++ objects/namespaced/openstack/pods/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf.txt",
                                ">f+++++++++ objects/namespaced/openstack/pods/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf.yaml",
                                ">f+++++++++ objects/namespaced/openstack/pods/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc.txt",
                                ">f+++++++++ objects/namespaced/openstack/pods/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc.yaml",
                                "cd+++++++++ objects/namespaced/openstack/rolebindings/",
                                ">f+++++++++ objects/namespaced/openstack/rolebindings/keepalived.txt",
                                ">f+++++++++ objects/namespaced/openstack/rolebindings/keepalived.yaml",
                                ">f+++++++++ objects/namespaced/openstack/rolebindings/pxc-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/rolebindings/pxc-operator.yaml",
                                ">f+++++++++ objects/namespaced/openstack/rolebindings/rabbitmq-cluster-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/rolebindings/rabbitmq-cluster-operator.yaml",
                                ">f+++++++++ objects/namespaced/openstack/rolebindings/rabbitmq-messaging-topology-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/rolebindings/rabbitmq-messaging-topology-operator.yaml",
                                "cd+++++++++ objects/namespaced/openstack/roles/",
                                ">f+++++++++ objects/namespaced/openstack/roles/keepalived.txt",
                                ">f+++++++++ objects/namespaced/openstack/roles/keepalived.yaml",
                                ">f+++++++++ objects/namespaced/openstack/roles/pxc-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/roles/pxc-operator.yaml",
                                ">f+++++++++ objects/namespaced/openstack/roles/rabbitmq-cluster-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/roles/rabbitmq-cluster-operator.yaml",
                                ">f+++++++++ objects/namespaced/openstack/roles/rabbitmq-messaging-topology-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/roles/rabbitmq-messaging-topology-operator.yaml",
                                "cd+++++++++ objects/namespaced/openstack/secrets/",
                                ">f+++++++++ objects/namespaced/openstack/secrets/internal-percona-xtradb.txt",
                                ">f+++++++++ objects/namespaced/openstack/secrets/internal-percona-xtradb.yaml",
                                ">f+++++++++ objects/namespaced/openstack/secrets/keepalived-etc.txt",
                                ">f+++++++++ objects/namespaced/openstack/secrets/keepalived-etc.yaml",
                                ">f+++++++++ objects/namespaced/openstack/secrets/percona-xtradb.txt",
                                ">f+++++++++ objects/namespaced/openstack/secrets/percona-xtradb.yaml",
                                ">f+++++++++ objects/namespaced/openstack/secrets/rabbitmq-messaging-topology-operator-webhook.txt",
                                ">f+++++++++ objects/namespaced/openstack/secrets/rabbitmq-messaging-topology-operator-webhook.yaml",
                                ">f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.memcached.v1.txt",
                                ">f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.memcached.v1.yaml",
                                ">f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.pxc-operator.v1.txt",
                                ">f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.pxc-operator.v1.yaml",
                                ">f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.rabbitmq-cluster-operator.v1.txt",
                                ">f+++++++++ objects/namespaced/openstack/secrets/sh.helm.release.v1.rabbitmq-cluster-operator.v1.yaml",
                                "cd+++++++++ objects/namespaced/openstack/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/default.yaml",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/keepalived.txt",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/keepalived.yaml",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/memcached-memcached.txt",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/memcached-memcached.yaml",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/pxc-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/pxc-operator.yaml",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/rabbitmq-cluster-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/rabbitmq-cluster-operator.yaml",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/rabbitmq-messaging-topology-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/serviceaccounts/rabbitmq-messaging-topology-operator.yaml",
                                "cd+++++++++ objects/namespaced/openstack/services/",
                                ">f+++++++++ objects/namespaced/openstack/services/memcached-metrics.txt",
                                ">f+++++++++ objects/namespaced/openstack/services/memcached-metrics.yaml",
                                ">f+++++++++ objects/namespaced/openstack/services/memcached.txt",
                                ">f+++++++++ objects/namespaced/openstack/services/memcached.yaml",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-cluster-operator.txt",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-cluster-operator.yaml",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy-metrics.txt",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy-metrics.yaml",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy-replicas.txt",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy-replicas.yaml",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy.txt",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-haproxy.yaml",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-pxc-unready.txt",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-pxc-unready.yaml",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-pxc.txt",
                                ">f+++++++++ objects/namespaced/openstack/services/percona-xtradb-pxc.yaml",
                                ">f+++++++++ objects/namespaced/openstack/services/rabbitmq-messaging-topology-operator-webhook.txt",
                                ">f+++++++++ objects/namespaced/openstack/services/rabbitmq-messaging-topology-operator-webhook.yaml",
                                "cd+++++++++ objects/namespaced/openstack/statefulsets/",
                                ">f+++++++++ objects/namespaced/openstack/statefulsets/percona-xtradb-haproxy.txt",
                                ">f+++++++++ objects/namespaced/openstack/statefulsets/percona-xtradb-haproxy.yaml",
                                ">f+++++++++ objects/namespaced/openstack/statefulsets/percona-xtradb-pxc.txt",
                                ">f+++++++++ objects/namespaced/openstack/statefulsets/percona-xtradb-pxc.yaml",
                                "cd+++++++++ objects/namespaced/secretgen-controller/",
                                "cd+++++++++ objects/namespaced/secretgen-controller/configmaps/",
                                ">f+++++++++ objects/namespaced/secretgen-controller/configmaps/kube-root-ca.crt.txt",
                                ">f+++++++++ objects/namespaced/secretgen-controller/configmaps/kube-root-ca.crt.yaml",
                                "cd+++++++++ objects/namespaced/secretgen-controller/deployment/",
                                ">f+++++++++ objects/namespaced/secretgen-controller/deployment/secretgen-controller.txt",
                                ">f+++++++++ objects/namespaced/secretgen-controller/deployment/secretgen-controller.yaml",
                                "cd+++++++++ objects/namespaced/secretgen-controller/pods/",
                                ">f+++++++++ objects/namespaced/secretgen-controller/pods/secretgen-controller-5cf976ccc7-vwxz4.txt",
                                ">f+++++++++ objects/namespaced/secretgen-controller/pods/secretgen-controller-5cf976ccc7-vwxz4.yaml",
                                "cd+++++++++ objects/namespaced/secretgen-controller/serviceaccounts/",
                                ">f+++++++++ objects/namespaced/secretgen-controller/serviceaccounts/default.txt",
                                ">f+++++++++ objects/namespaced/secretgen-controller/serviceaccounts/default.yaml",
                                ">f+++++++++ objects/namespaced/secretgen-controller/serviceaccounts/secretgen-controller-sa.txt",
                                ">f+++++++++ objects/namespaced/secretgen-controller/serviceaccounts/secretgen-controller-sa.yaml"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000017",
                        "name": "describe-kubernetes-objects",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/describe-kubernetes-objects"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:22.332574Z",
                            "start": "2026-04-04T10:26:21.604059Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-00000000001d",
                        "name": "Downloads logs to executor"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/tmp/logs/pod-logs",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/tmp/logs/pod-logs",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": null,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/tmp/logs/pod-logs",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/tmp/logs/pod-logs",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-00000000001f",
                        "name": "gather-pod-logs",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-pod-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:22.563738Z",
                            "start": "2026-04-04T10:26:22.345936Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000021",
                        "name": "creating directory for pod logs"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/tmp/logs/pod-logs/failed-pods",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/tmp/logs/pod-logs/failed-pods",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": null,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/tmp/logs/pod-logs/failed-pods",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/tmp/logs/pod-logs/failed-pods",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-00000000001f",
                        "name": "gather-pod-logs",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-pod-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:22.790535Z",
                            "start": "2026-04-04T10:26:22.571496Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000022",
                        "name": "creating directory for failed pod logs"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "shell",
                            "changed": true,
                            "cmd": "set -e\nfunction get_namespaces () {\n  kubectl get namespaces -o name | awk -F '/' '{ print $NF }'\n}\nfunction get_pods () {\n  NAMESPACE=$1\n  kubectl get pods -n ${NAMESPACE} -o name | awk -F '/' '{ print $NF }' | xargs -I {} echo ${NAMESPACE} {}\n}\nexport -f get_pods\nfunction get_pod_logs () {\n  NAMESPACE=${1% *}\n  POD=${1#* }\n  INIT_CONTAINERS=$(kubectl get pod $POD -n ${NAMESPACE} -o json | jq -r '.spec.initContainers[]?.name')\n  CONTAINERS=$(kubectl get pod $POD -n ${NAMESPACE} -o json | jq -r '.spec.containers[].name')\n  for CONTAINER in ${INIT_CONTAINERS} ${CONTAINERS}; do\n    echo \"${NAMESPACE}/${POD}/${CONTAINER}\"\n    mkdir -p \"/tmp/logs/pod-logs/${NAMESPACE}/${POD}\"\n    mkdir -p \"/tmp/logs/pod-logs/failed-pods/${NAMESPACE}/${POD}\"\n    kubectl logs ${POD} -n ${NAMESPACE} -c ${CONTAINER} > \"/tmp/logs/pod-logs/${NAMESPACE}/${POD}/${CONTAINER}.txt\"\n    kubectl logs --previous ${POD} -n ${NAMESPACE} -c ${CONTAINER} > \"/tmp/logs/pod-logs/failed-pods/${NAMESPACE}/${POD}/${CONTAINER}.txt\"\n  done\n}\nexport -f get_pod_logs\nget_namespaces |  xargs -r -I {} bash -c 'get_pods \"$@\"' _ {} |  xargs -r -I {} bash -c 'get_pod_logs \"$@\"' _ {}",
                            "delta": "0:00:20.491698",
                            "end": "2026-04-04 10:26:43.500684",
                            "failed": true,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -e\nfunction get_namespaces () {\n  kubectl get namespaces -o name | awk -F '/' '{ print $NF }'\n}\nfunction get_pods () {\n  NAMESPACE=$1\n  kubectl get pods -n ${NAMESPACE} -o name | awk -F '/' '{ print $NF }' | xargs -I {} echo ${NAMESPACE} {}\n}\nexport -f get_pods\nfunction get_pod_logs () {\n  NAMESPACE=${1% *}\n  POD=${1#* }\n  INIT_CONTAINERS=$(kubectl get pod $POD -n ${NAMESPACE} -o json | jq -r '.spec.initContainers[]?.name')\n  CONTAINERS=$(kubectl get pod $POD -n ${NAMESPACE} -o json | jq -r '.spec.containers[].name')\n  for CONTAINER in ${INIT_CONTAINERS} ${CONTAINERS}; do\n    echo \"${NAMESPACE}/${POD}/${CONTAINER}\"\n    mkdir -p \"/tmp/logs/pod-logs/${NAMESPACE}/${POD}\"\n    mkdir -p \"/tmp/logs/pod-logs/failed-pods/${NAMESPACE}/${POD}\"\n    kubectl logs ${POD} -n ${NAMESPACE} -c ${CONTAINER} > \"/tmp/logs/pod-logs/${NAMESPACE}/${POD}/${CONTAINER}.txt\"\n    kubectl logs --previous ${POD} -n ${NAMESPACE} -c ${CONTAINER} > \"/tmp/logs/pod-logs/failed-pods/${NAMESPACE}/${POD}/${CONTAINER}.txt\"\n  done\n}\nexport -f get_pod_logs\nget_namespaces |  xargs -r -I {} bash -c 'get_pods \"$@\"' _ {} |  xargs -r -I {} bash -c 'get_pod_logs \"$@\"' _ {}",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-cc15-c3f6-000000000023-1-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "non-zero return code",
                            "rc": 123,
                            "start": "2026-04-04 10:26:23.008986",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "auth-system/keycloak-0/keycloak\nError from server (BadRequest): previous terminated container \"keycloak\" in pod \"keycloak-0\" not found\ncert-manager/cert-manager-75969b45cf-d5c4v/cert-manager-controller\nError from server (BadRequest): previous terminated container \"cert-manager-controller\" in pod \"cert-manager-75969b45cf-d5c4v\" not found\ncert-manager/cert-manager-cainjector-84f45b698d-v64n5/cert-manager-cainjector\nError from server (BadRequest): previous terminated container \"cert-manager-cainjector\" in pod \"cert-manager-cainjector-84f45b698d-v64n5\" not found\ncert-manager/cert-manager-webhook-9997c64fc-8w8sz/cert-manager-webhook\nError from server (BadRequest): previous terminated container \"cert-manager-webhook\" in pod \"cert-manager-webhook-9997c64fc-8w8sz\" not found\nenvoy-gateway-system/envoy-gateway-78446f96c9-7zvgb/envoy-gateway\nError from server (BadRequest): previous terminated container \"envoy-gateway\" in pod \"envoy-gateway-78446f96c9-7zvgb\" not found\ningress-nginx/ingress-nginx-controller-dmjt4/controller\nError from server (BadRequest): previous terminated container \"controller\" in pod \"ingress-nginx-controller-dmjt4\" not found\ningress-nginx/ingress-nginx-defaultbackend-6987ff55cf-plsk9/ingress-nginx-default-backend\nError from server (BadRequest): previous terminated container \"ingress-nginx-default-backend\" in pod \"ingress-nginx-defaultbackend-6987ff55cf-plsk9\" not found\nkube-system/cilium-operator-869df985b8-nqvtb/cilium-operator\nError from server (BadRequest): previous terminated container \"cilium-operator\" in pod \"cilium-operator-869df985b8-nqvtb\" not found\nkube-system/cilium-sk6n5/config\nError from server (BadRequest): previous terminated container \"config\" in pod \"cilium-sk6n5\" not found\nkube-system/cilium-sk6n5/mount-cgroup\nError from server (BadRequest): previous terminated container \"mount-cgroup\" in pod \"cilium-sk6n5\" not found\nkube-system/cilium-sk6n5/apply-sysctl-overwrites\nError from server (BadRequest): previous terminated container \"apply-sysctl-overwrites\" in pod \"cilium-sk6n5\" not found\nkube-system/cilium-sk6n5/mount-bpf-fs\nError from server (BadRequest): previous terminated container \"mount-bpf-fs\" in pod \"cilium-sk6n5\" not found\nkube-system/cilium-sk6n5/clean-cilium-state\nError from server (BadRequest): previous terminated container \"clean-cilium-state\" in pod \"cilium-sk6n5\" not found\nkube-system/cilium-sk6n5/install-cni-binaries\nError from server (BadRequest): previous terminated container \"install-cni-binaries\" in pod \"cilium-sk6n5\" not found\nkube-system/cilium-sk6n5/cilium-agent\nError from server (BadRequest): previous terminated container \"cilium-agent\" in pod \"cilium-sk6n5\" not found\nkube-system/coredns-67659f764b-mdrt6/coredns\nError from server (BadRequest): previous terminated container \"coredns\" in pod \"coredns-67659f764b-mdrt6\" not found\nkube-system/coredns-67659f764b-xb6lv/coredns\nError from server (BadRequest): previous terminated container \"coredns\" in pod \"coredns-67659f764b-xb6lv\" not found\nkube-system/etcd-instance/etcd\nError from server (BadRequest): previous terminated container \"etcd\" in pod \"etcd-instance\" not found\nkube-system/kube-apiserver-instance/kube-apiserver\nError from server (BadRequest): previous terminated container \"kube-apiserver\" in pod \"kube-apiserver-instance\" not found\nkube-system/kube-controller-manager-instance/kube-controller-manager\nError from server (BadRequest): previous terminated container \"kube-controller-manager\" in pod \"kube-controller-manager-instance\" not found\nkube-system/kube-proxy-cxlxv/kube-proxy\nError from server (BadRequest): previous terminated container \"kube-proxy\" in pod \"kube-proxy-cxlxv\" not found\nkube-system/kube-scheduler-instance/kube-scheduler\nError from server (BadRequest): previous terminated container \"kube-scheduler\" in pod \"kube-scheduler-instance\" not found\nkube-system/kube-vip-instance/kube-vip\nError from server (BadRequest): previous terminated container \"kube-vip\" in pod \"kube-vip-instance\" not found\nlocal-path-storage/local-path-provisioner-679c578f5-m52kp/local-path-provisioner\nError from server (BadRequest): previous terminated container \"local-path-provisioner\" in pod \"local-path-provisioner-679c578f5-m52kp\" not found\nmonitoring/alertmanager-kube-prometheus-stack-alertmanager-0/init-config-reloader\nError from server (BadRequest): previous terminated container \"init-config-reloader\" in pod \"alertmanager-kube-prometheus-stack-alertmanager-0\" not found\nmonitoring/alertmanager-kube-prometheus-stack-alertmanager-0/alertmanager\nError from server (BadRequest): previous terminated container \"alertmanager\" in pod \"alertmanager-kube-prometheus-stack-alertmanager-0\" not found\nmonitoring/alertmanager-kube-prometheus-stack-alertmanager-0/config-reloader\nError from server (BadRequest): previous terminated container \"config-reloader\" in pod \"alertmanager-kube-prometheus-stack-alertmanager-0\" not found\nmonitoring/alertmanager-kube-prometheus-stack-alertmanager-0/oauth2-proxy\nError from server (BadRequest): previous terminated container \"oauth2-proxy\" in pod \"alertmanager-kube-prometheus-stack-alertmanager-0\" not found\nmonitoring/goldpinger-gz8xd/goldpinger-daemon\nError from server (BadRequest): previous terminated container \"goldpinger-daemon\" in pod \"goldpinger-gz8xd\" not found\nmonitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-dashboard\nError from server (BadRequest): previous terminated container \"grafana-sc-dashboard\" in pod \"kube-prometheus-stack-grafana-69fc56495c-5h4gd\" not found\nmonitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-datasources\nError from server (BadRequest): previous terminated container \"grafana-sc-datasources\" in pod \"kube-prometheus-stack-grafana-69fc56495c-5h4gd\" not found\nmonitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana\nError from server (BadRequest): previous terminated container \"grafana\" in pod \"kube-prometheus-stack-grafana-69fc56495c-5h4gd\" not found\nmonitoring/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj/kube-state-metrics\nError from server (BadRequest): previous terminated container \"kube-state-metrics\" in pod \"kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj\" not found\nmonitoring/kube-prometheus-stack-operator-cc88b86d6-xspgk/kube-prometheus-stack\nError from server (BadRequest): previous terminated container \"kube-prometheus-stack\" in pod \"kube-prometheus-stack-operator-cc88b86d6-xspgk\" not found\nmonitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/node-exporter\nmonitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/pod-tls-sidecar\nError from server (BadRequest): previous terminated container \"pod-tls-sidecar\" in pod \"kube-prometheus-stack-prometheus-node-exporter-jmj6n\" not found\nmonitoring/loki-0/loki\nError from server (BadRequest): previous terminated container \"loki\" in pod \"loki-0\" not found\nmonitoring/loki-gateway-5cc45946cd-wp7ks/nginx\nError from server (BadRequest): previous terminated container \"nginx\" in pod \"loki-gateway-5cc45946cd-wp7ks\" not found\nmonitoring/node-feature-discovery-master-68cf466d5f-qknzk/master\nError from server (BadRequest): previous terminated container \"master\" in pod \"node-feature-discovery-master-68cf466d5f-qknzk\" not found\nmonitoring/node-feature-discovery-worker-7l5z7/worker\nError from server (BadRequest): previous terminated container \"worker\" in pod \"node-feature-discovery-worker-7l5z7\" not found\nmonitoring/prometheus-kube-prometheus-stack-prometheus-0/init-config-reloader\nError from server (BadRequest): previous terminated container \"init-config-reloader\" in pod \"prometheus-kube-prometheus-stack-prometheus-0\" not found\nmonitoring/prometheus-kube-prometheus-stack-prometheus-0/prometheus\nError from server (BadRequest): previous terminated container \"prometheus\" in pod \"prometheus-kube-prometheus-stack-prometheus-0\" not found\nmonitoring/prometheus-kube-prometheus-stack-prometheus-0/config-reloader\nError from server (BadRequest): previous terminated container \"config-reloader\" in pod \"prometheus-kube-prometheus-stack-prometheus-0\" not found\nmonitoring/prometheus-kube-prometheus-stack-prometheus-0/pod-tls-sidecar\nError from server (BadRequest): previous terminated container \"pod-tls-sidecar\" in pod \"prometheus-kube-prometheus-stack-prometheus-0\" not found\nmonitoring/prometheus-kube-prometheus-stack-prometheus-0/oauth2-proxy\nError from server (BadRequest): previous terminated container \"oauth2-proxy\" in pod \"prometheus-kube-prometheus-stack-prometheus-0\" not found\nmonitoring/prometheus-pushgateway-5c694d99cc-jnjt7/pushgateway\nError from server (BadRequest): previous terminated container \"pushgateway\" in pod \"prometheus-pushgateway-5c694d99cc-jnjt7\" not found\nmonitoring/vector-75jgf/vector\nError from server (BadRequest): previous terminated container \"vector\" in pod \"vector-75jgf\" not found\nopenstack/keepalived-wktjf/init\nError from server (BadRequest): previous terminated container \"init\" in pod \"keepalived-wktjf\" not found\nopenstack/keepalived-wktjf/wait-for-ip\nError from server (BadRequest): container \"wait-for-ip\" in pod \"keepalived-wktjf\" is waiting to start: PodInitializing\nError from server (BadRequest): previous terminated container \"wait-for-ip\" in pod \"keepalived-wktjf\" not found\nopenstack/keepalived-wktjf/keepalived\nError from server (BadRequest): container \"keepalived\" in pod \"keepalived-wktjf\" is waiting to start: PodInitializing\nError from server (BadRequest): previous terminated container \"keepalived\" in pod \"keepalived-wktjf\" not found\nopenstack/memcached-memcached-c47d949f5-vc5cm/init\nError from server (BadRequest): previous terminated container \"init\" in pod \"memcached-memcached-c47d949f5-vc5cm\" not found\nopenstack/memcached-memcached-c47d949f5-vc5cm/memcached\nError from server (BadRequest): previous terminated container \"memcached\" in pod \"memcached-memcached-c47d949f5-vc5cm\" not found\nopenstack/memcached-memcached-c47d949f5-vc5cm/memcached-exporter\nError from server (BadRequest): previous terminated container \"memcached-exporter\" in pod \"memcached-memcached-c47d949f5-vc5cm\" not found\nopenstack/percona-xtradb-haproxy-0/pxc-init\nError from server (BadRequest): previous terminated container \"pxc-init\" in pod \"percona-xtradb-haproxy-0\" not found\nopenstack/percona-xtradb-haproxy-0/haproxy\nError from server (BadRequest): previous terminated container \"haproxy\" in pod \"percona-xtradb-haproxy-0\" not found\nopenstack/percona-xtradb-haproxy-0/pxc-monit\nError from server (BadRequest): previous terminated container \"pxc-monit\" in pod \"percona-xtradb-haproxy-0\" not found\nopenstack/percona-xtradb-pxc-0/pxc-init\nError from server (BadRequest): previous terminated container \"pxc-init\" in pod \"percona-xtradb-pxc-0\" not found\nopenstack/percona-xtradb-pxc-0/pxc\nError from server (BadRequest): previous terminated container \"pxc\" in pod \"percona-xtradb-pxc-0\" not found\nopenstack/percona-xtradb-pxc-0/exporter\nError from server (BadRequest): previous terminated container \"exporter\" in pod \"percona-xtradb-pxc-0\" not found\nopenstack/pxc-operator-59d66cbc78-clxvf/percona-xtradb-cluster-operator\nError from server (BadRequest): previous terminated container \"percona-xtradb-cluster-operator\" in pod \"pxc-operator-59d66cbc78-clxvf\" not found\nopenstack/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf/rabbitmq-cluster-operator\nError from server (BadRequest): previous terminated container \"rabbitmq-cluster-operator\" in pod \"rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf\" not found\nopenstack/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc/rabbitmq-cluster-operator\nError from server (BadRequest): previous terminated container \"rabbitmq-cluster-operator\" in pod \"rabbitmq-messaging-topology-operator-69996b7f7f-52mnc\" not found\nsecretgen-controller/secretgen-controller-5cf976ccc7-vwxz4/secretgen-controller\nError from server (BadRequest): previous terminated container \"secretgen-controller\" in pod \"secretgen-controller-5cf976ccc7-vwxz4\" not found",
                            "stdout_lines": [
                                "auth-system/keycloak-0/keycloak",
                                "Error from server (BadRequest): previous terminated container \"keycloak\" in pod \"keycloak-0\" not found",
                                "cert-manager/cert-manager-75969b45cf-d5c4v/cert-manager-controller",
                                "Error from server (BadRequest): previous terminated container \"cert-manager-controller\" in pod \"cert-manager-75969b45cf-d5c4v\" not found",
                                "cert-manager/cert-manager-cainjector-84f45b698d-v64n5/cert-manager-cainjector",
                                "Error from server (BadRequest): previous terminated container \"cert-manager-cainjector\" in pod \"cert-manager-cainjector-84f45b698d-v64n5\" not found",
                                "cert-manager/cert-manager-webhook-9997c64fc-8w8sz/cert-manager-webhook",
                                "Error from server (BadRequest): previous terminated container \"cert-manager-webhook\" in pod \"cert-manager-webhook-9997c64fc-8w8sz\" not found",
                                "envoy-gateway-system/envoy-gateway-78446f96c9-7zvgb/envoy-gateway",
                                "Error from server (BadRequest): previous terminated container \"envoy-gateway\" in pod \"envoy-gateway-78446f96c9-7zvgb\" not found",
                                "ingress-nginx/ingress-nginx-controller-dmjt4/controller",
                                "Error from server (BadRequest): previous terminated container \"controller\" in pod \"ingress-nginx-controller-dmjt4\" not found",
                                "ingress-nginx/ingress-nginx-defaultbackend-6987ff55cf-plsk9/ingress-nginx-default-backend",
                                "Error from server (BadRequest): previous terminated container \"ingress-nginx-default-backend\" in pod \"ingress-nginx-defaultbackend-6987ff55cf-plsk9\" not found",
                                "kube-system/cilium-operator-869df985b8-nqvtb/cilium-operator",
                                "Error from server (BadRequest): previous terminated container \"cilium-operator\" in pod \"cilium-operator-869df985b8-nqvtb\" not found",
                                "kube-system/cilium-sk6n5/config",
                                "Error from server (BadRequest): previous terminated container \"config\" in pod \"cilium-sk6n5\" not found",
                                "kube-system/cilium-sk6n5/mount-cgroup",
                                "Error from server (BadRequest): previous terminated container \"mount-cgroup\" in pod \"cilium-sk6n5\" not found",
                                "kube-system/cilium-sk6n5/apply-sysctl-overwrites",
                                "Error from server (BadRequest): previous terminated container \"apply-sysctl-overwrites\" in pod \"cilium-sk6n5\" not found",
                                "kube-system/cilium-sk6n5/mount-bpf-fs",
                                "Error from server (BadRequest): previous terminated container \"mount-bpf-fs\" in pod \"cilium-sk6n5\" not found",
                                "kube-system/cilium-sk6n5/clean-cilium-state",
                                "Error from server (BadRequest): previous terminated container \"clean-cilium-state\" in pod \"cilium-sk6n5\" not found",
                                "kube-system/cilium-sk6n5/install-cni-binaries",
                                "Error from server (BadRequest): previous terminated container \"install-cni-binaries\" in pod \"cilium-sk6n5\" not found",
                                "kube-system/cilium-sk6n5/cilium-agent",
                                "Error from server (BadRequest): previous terminated container \"cilium-agent\" in pod \"cilium-sk6n5\" not found",
                                "kube-system/coredns-67659f764b-mdrt6/coredns",
                                "Error from server (BadRequest): previous terminated container \"coredns\" in pod \"coredns-67659f764b-mdrt6\" not found",
                                "kube-system/coredns-67659f764b-xb6lv/coredns",
                                "Error from server (BadRequest): previous terminated container \"coredns\" in pod \"coredns-67659f764b-xb6lv\" not found",
                                "kube-system/etcd-instance/etcd",
                                "Error from server (BadRequest): previous terminated container \"etcd\" in pod \"etcd-instance\" not found",
                                "kube-system/kube-apiserver-instance/kube-apiserver",
                                "Error from server (BadRequest): previous terminated container \"kube-apiserver\" in pod \"kube-apiserver-instance\" not found",
                                "kube-system/kube-controller-manager-instance/kube-controller-manager",
                                "Error from server (BadRequest): previous terminated container \"kube-controller-manager\" in pod \"kube-controller-manager-instance\" not found",
                                "kube-system/kube-proxy-cxlxv/kube-proxy",
                                "Error from server (BadRequest): previous terminated container \"kube-proxy\" in pod \"kube-proxy-cxlxv\" not found",
                                "kube-system/kube-scheduler-instance/kube-scheduler",
                                "Error from server (BadRequest): previous terminated container \"kube-scheduler\" in pod \"kube-scheduler-instance\" not found",
                                "kube-system/kube-vip-instance/kube-vip",
                                "Error from server (BadRequest): previous terminated container \"kube-vip\" in pod \"kube-vip-instance\" not found",
                                "local-path-storage/local-path-provisioner-679c578f5-m52kp/local-path-provisioner",
                                "Error from server (BadRequest): previous terminated container \"local-path-provisioner\" in pod \"local-path-provisioner-679c578f5-m52kp\" not found",
                                "monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/init-config-reloader",
                                "Error from server (BadRequest): previous terminated container \"init-config-reloader\" in pod \"alertmanager-kube-prometheus-stack-alertmanager-0\" not found",
                                "monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/alertmanager",
                                "Error from server (BadRequest): previous terminated container \"alertmanager\" in pod \"alertmanager-kube-prometheus-stack-alertmanager-0\" not found",
                                "monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/config-reloader",
                                "Error from server (BadRequest): previous terminated container \"config-reloader\" in pod \"alertmanager-kube-prometheus-stack-alertmanager-0\" not found",
                                "monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/oauth2-proxy",
                                "Error from server (BadRequest): previous terminated container \"oauth2-proxy\" in pod \"alertmanager-kube-prometheus-stack-alertmanager-0\" not found",
                                "monitoring/goldpinger-gz8xd/goldpinger-daemon",
                                "Error from server (BadRequest): previous terminated container \"goldpinger-daemon\" in pod \"goldpinger-gz8xd\" not found",
                                "monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-dashboard",
                                "Error from server (BadRequest): previous terminated container \"grafana-sc-dashboard\" in pod \"kube-prometheus-stack-grafana-69fc56495c-5h4gd\" not found",
                                "monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-datasources",
                                "Error from server (BadRequest): previous terminated container \"grafana-sc-datasources\" in pod \"kube-prometheus-stack-grafana-69fc56495c-5h4gd\" not found",
                                "monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana",
                                "Error from server (BadRequest): previous terminated container \"grafana\" in pod \"kube-prometheus-stack-grafana-69fc56495c-5h4gd\" not found",
                                "monitoring/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj/kube-state-metrics",
                                "Error from server (BadRequest): previous terminated container \"kube-state-metrics\" in pod \"kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj\" not found",
                                "monitoring/kube-prometheus-stack-operator-cc88b86d6-xspgk/kube-prometheus-stack",
                                "Error from server (BadRequest): previous terminated container \"kube-prometheus-stack\" in pod \"kube-prometheus-stack-operator-cc88b86d6-xspgk\" not found",
                                "monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/node-exporter",
                                "monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/pod-tls-sidecar",
                                "Error from server (BadRequest): previous terminated container \"pod-tls-sidecar\" in pod \"kube-prometheus-stack-prometheus-node-exporter-jmj6n\" not found",
                                "monitoring/loki-0/loki",
                                "Error from server (BadRequest): previous terminated container \"loki\" in pod \"loki-0\" not found",
                                "monitoring/loki-gateway-5cc45946cd-wp7ks/nginx",
                                "Error from server (BadRequest): previous terminated container \"nginx\" in pod \"loki-gateway-5cc45946cd-wp7ks\" not found",
                                "monitoring/node-feature-discovery-master-68cf466d5f-qknzk/master",
                                "Error from server (BadRequest): previous terminated container \"master\" in pod \"node-feature-discovery-master-68cf466d5f-qknzk\" not found",
                                "monitoring/node-feature-discovery-worker-7l5z7/worker",
                                "Error from server (BadRequest): previous terminated container \"worker\" in pod \"node-feature-discovery-worker-7l5z7\" not found",
                                "monitoring/prometheus-kube-prometheus-stack-prometheus-0/init-config-reloader",
                                "Error from server (BadRequest): previous terminated container \"init-config-reloader\" in pod \"prometheus-kube-prometheus-stack-prometheus-0\" not found",
                                "monitoring/prometheus-kube-prometheus-stack-prometheus-0/prometheus",
                                "Error from server (BadRequest): previous terminated container \"prometheus\" in pod \"prometheus-kube-prometheus-stack-prometheus-0\" not found",
                                "monitoring/prometheus-kube-prometheus-stack-prometheus-0/config-reloader",
                                "Error from server (BadRequest): previous terminated container \"config-reloader\" in pod \"prometheus-kube-prometheus-stack-prometheus-0\" not found",
                                "monitoring/prometheus-kube-prometheus-stack-prometheus-0/pod-tls-sidecar",
                                "Error from server (BadRequest): previous terminated container \"pod-tls-sidecar\" in pod \"prometheus-kube-prometheus-stack-prometheus-0\" not found",
                                "monitoring/prometheus-kube-prometheus-stack-prometheus-0/oauth2-proxy",
                                "Error from server (BadRequest): previous terminated container \"oauth2-proxy\" in pod \"prometheus-kube-prometheus-stack-prometheus-0\" not found",
                                "monitoring/prometheus-pushgateway-5c694d99cc-jnjt7/pushgateway",
                                "Error from server (BadRequest): previous terminated container \"pushgateway\" in pod \"prometheus-pushgateway-5c694d99cc-jnjt7\" not found",
                                "monitoring/vector-75jgf/vector",
                                "Error from server (BadRequest): previous terminated container \"vector\" in pod \"vector-75jgf\" not found",
                                "openstack/keepalived-wktjf/init",
                                "Error from server (BadRequest): previous terminated container \"init\" in pod \"keepalived-wktjf\" not found",
                                "openstack/keepalived-wktjf/wait-for-ip",
                                "Error from server (BadRequest): container \"wait-for-ip\" in pod \"keepalived-wktjf\" is waiting to start: PodInitializing",
                                "Error from server (BadRequest): previous terminated container \"wait-for-ip\" in pod \"keepalived-wktjf\" not found",
                                "openstack/keepalived-wktjf/keepalived",
                                "Error from server (BadRequest): container \"keepalived\" in pod \"keepalived-wktjf\" is waiting to start: PodInitializing",
                                "Error from server (BadRequest): previous terminated container \"keepalived\" in pod \"keepalived-wktjf\" not found",
                                "openstack/memcached-memcached-c47d949f5-vc5cm/init",
                                "Error from server (BadRequest): previous terminated container \"init\" in pod \"memcached-memcached-c47d949f5-vc5cm\" not found",
                                "openstack/memcached-memcached-c47d949f5-vc5cm/memcached",
                                "Error from server (BadRequest): previous terminated container \"memcached\" in pod \"memcached-memcached-c47d949f5-vc5cm\" not found",
                                "openstack/memcached-memcached-c47d949f5-vc5cm/memcached-exporter",
                                "Error from server (BadRequest): previous terminated container \"memcached-exporter\" in pod \"memcached-memcached-c47d949f5-vc5cm\" not found",
                                "openstack/percona-xtradb-haproxy-0/pxc-init",
                                "Error from server (BadRequest): previous terminated container \"pxc-init\" in pod \"percona-xtradb-haproxy-0\" not found",
                                "openstack/percona-xtradb-haproxy-0/haproxy",
                                "Error from server (BadRequest): previous terminated container \"haproxy\" in pod \"percona-xtradb-haproxy-0\" not found",
                                "openstack/percona-xtradb-haproxy-0/pxc-monit",
                                "Error from server (BadRequest): previous terminated container \"pxc-monit\" in pod \"percona-xtradb-haproxy-0\" not found",
                                "openstack/percona-xtradb-pxc-0/pxc-init",
                                "Error from server (BadRequest): previous terminated container \"pxc-init\" in pod \"percona-xtradb-pxc-0\" not found",
                                "openstack/percona-xtradb-pxc-0/pxc",
                                "Error from server (BadRequest): previous terminated container \"pxc\" in pod \"percona-xtradb-pxc-0\" not found",
                                "openstack/percona-xtradb-pxc-0/exporter",
                                "Error from server (BadRequest): previous terminated container \"exporter\" in pod \"percona-xtradb-pxc-0\" not found",
                                "openstack/pxc-operator-59d66cbc78-clxvf/percona-xtradb-cluster-operator",
                                "Error from server (BadRequest): previous terminated container \"percona-xtradb-cluster-operator\" in pod \"pxc-operator-59d66cbc78-clxvf\" not found",
                                "openstack/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf/rabbitmq-cluster-operator",
                                "Error from server (BadRequest): previous terminated container \"rabbitmq-cluster-operator\" in pod \"rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf\" not found",
                                "openstack/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc/rabbitmq-cluster-operator",
                                "Error from server (BadRequest): previous terminated container \"rabbitmq-cluster-operator\" in pod \"rabbitmq-messaging-topology-operator-69996b7f7f-52mnc\" not found",
                                "secretgen-controller/secretgen-controller-5cf976ccc7-vwxz4/secretgen-controller",
                                "Error from server (BadRequest): previous terminated container \"secretgen-controller\" in pod \"secretgen-controller-5cf976ccc7-vwxz4\" not found"
                            ],
                            "zuul_log_id": "0242ac17-0011-cc15-c3f6-000000000023-1-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-00000000001f",
                        "name": "gather-pod-logs",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-pod-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:43.866791Z",
                            "start": "2026-04-04T10:26:22.816993Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000023",
                        "name": "retrieve all kubernetes logs, current and previous (if they exist)"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "synchronize",
                            "changed": true,
                            "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --rsh='/usr/bin/ssh -S none -o Port=22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null' --rsync-path='sudo -u root rsync' --out-format='<<CHANGED>>%i %n%L' zuul@199.204.45.153:/tmp/logs/pod-logs /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                            "invocation": {
                                "module_args": {
                                    "_local_rsync_password": null,
                                    "_local_rsync_path": "rsync",
                                    "_ssh_args": null,
                                    "_substitute_controller": false,
                                    "archive": true,
                                    "checksum": false,
                                    "compress": true,
                                    "copy_links": false,
                                    "delay_updates": true,
                                    "delete": false,
                                    "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                                    "dest_port": 22,
                                    "dirs": false,
                                    "existing_only": false,
                                    "group": null,
                                    "link_dest": null,
                                    "links": null,
                                    "mode": "pull",
                                    "owner": null,
                                    "partial": false,
                                    "perms": null,
                                    "private_key": null,
                                    "recursive": null,
                                    "rsync_opts": [],
                                    "rsync_path": "sudo -u root rsync",
                                    "rsync_timeout": 0,
                                    "set_remote_user": true,
                                    "src": "zuul@199.204.45.153:/tmp/logs/pod-logs",
                                    "ssh_connection_multiplexing": false,
                                    "times": null,
                                    "use_ssh_args": false,
                                    "verify_host": false
                                }
                            },
                            "msg": "cd+++++++++ pod-logs/\ncd+++++++++ pod-logs/auth-system/\ncd+++++++++ pod-logs/auth-system/keycloak-0/\n>f+++++++++ pod-logs/auth-system/keycloak-0/keycloak.txt\ncd+++++++++ pod-logs/cert-manager/\ncd+++++++++ pod-logs/cert-manager/cert-manager-75969b45cf-d5c4v/\n>f+++++++++ pod-logs/cert-manager/cert-manager-75969b45cf-d5c4v/cert-manager-controller.txt\ncd+++++++++ pod-logs/cert-manager/cert-manager-cainjector-84f45b698d-v64n5/\n>f+++++++++ pod-logs/cert-manager/cert-manager-cainjector-84f45b698d-v64n5/cert-manager-cainjector.txt\ncd+++++++++ pod-logs/cert-manager/cert-manager-webhook-9997c64fc-8w8sz/\n>f+++++++++ pod-logs/cert-manager/cert-manager-webhook-9997c64fc-8w8sz/cert-manager-webhook.txt\ncd+++++++++ pod-logs/envoy-gateway-system/\ncd+++++++++ pod-logs/envoy-gateway-system/envoy-gateway-78446f96c9-7zvgb/\n>f+++++++++ pod-logs/envoy-gateway-system/envoy-gateway-78446f96c9-7zvgb/envoy-gateway.txt\ncd+++++++++ pod-logs/failed-pods/\ncd+++++++++ pod-logs/failed-pods/auth-system/\ncd+++++++++ pod-logs/failed-pods/auth-system/keycloak-0/\n>f+++++++++ pod-logs/failed-pods/auth-system/keycloak-0/keycloak.txt\ncd+++++++++ pod-logs/failed-pods/cert-manager/\ncd+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-75969b45cf-d5c4v/\n>f+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-75969b45cf-d5c4v/cert-manager-controller.txt\ncd+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-cainjector-84f45b698d-v64n5/\n>f+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-cainjector-84f45b698d-v64n5/cert-manager-cainjector.txt\ncd+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-webhook-9997c64fc-8w8sz/\n>f+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-webhook-9997c64fc-8w8sz/cert-manager-webhook.txt\ncd+++++++++ pod-logs/failed-pods/envoy-gateway-system/\ncd+++++++++ pod-logs/failed-pods/envoy-gateway-system/envoy-gateway-78446f96c9-7zvgb/\n>f+++++++++ pod-logs/failed-pods/envoy-gateway-system/envoy-gateway-78446f96c9-7zvgb/envoy-gateway.txt\ncd+++++++++ pod-logs/failed-pods/ingress-nginx/\ncd+++++++++ pod-logs/failed-pods/ingress-nginx/ingress-nginx-controller-dmjt4/\n>f+++++++++ pod-logs/failed-pods/ingress-nginx/ingress-nginx-controller-dmjt4/controller.txt\ncd+++++++++ pod-logs/failed-pods/ingress-nginx/ingress-nginx-defaultbackend-6987ff55cf-plsk9/\n>f+++++++++ pod-logs/failed-pods/ingress-nginx/ingress-nginx-defaultbackend-6987ff55cf-plsk9/ingress-nginx-default-backend.txt\ncd+++++++++ pod-logs/failed-pods/kube-system/\ncd+++++++++ pod-logs/failed-pods/kube-system/cilium-operator-869df985b8-nqvtb/\n>f+++++++++ pod-logs/failed-pods/kube-system/cilium-operator-869df985b8-nqvtb/cilium-operator.txt\ncd+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/\n>f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/apply-sysctl-overwrites.txt\n>f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/cilium-agent.txt\n>f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/clean-cilium-state.txt\n>f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/config.txt\n>f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/install-cni-binaries.txt\n>f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/mount-bpf-fs.txt\n>f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/mount-cgroup.txt\ncd+++++++++ pod-logs/failed-pods/kube-system/coredns-67659f764b-mdrt6/\n>f+++++++++ pod-logs/failed-pods/kube-system/coredns-67659f764b-mdrt6/coredns.txt\ncd+++++++++ pod-logs/failed-pods/kube-system/coredns-67659f764b-xb6lv/\n>f+++++++++ pod-logs/failed-pods/kube-system/coredns-67659f764b-xb6lv/coredns.txt\ncd+++++++++ pod-logs/failed-pods/kube-system/etcd-instance/\n>f+++++++++ pod-logs/failed-pods/kube-system/etcd-instance/etcd.txt\ncd+++++++++ pod-logs/failed-pods/kube-system/kube-apiserver-instance/\n>f+++++++++ pod-logs/failed-pods/kube-system/kube-apiserver-instance/kube-apiserver.txt\ncd+++++++++ pod-logs/failed-pods/kube-system/kube-controller-manager-instance/\n>f+++++++++ pod-logs/failed-pods/kube-system/kube-controller-manager-instance/kube-controller-manager.txt\ncd+++++++++ pod-logs/failed-pods/kube-system/kube-proxy-cxlxv/\n>f+++++++++ pod-logs/failed-pods/kube-system/kube-proxy-cxlxv/kube-proxy.txt\ncd+++++++++ pod-logs/failed-pods/kube-system/kube-scheduler-instance/\n>f+++++++++ pod-logs/failed-pods/kube-system/kube-scheduler-instance/kube-scheduler.txt\ncd+++++++++ pod-logs/failed-pods/kube-system/kube-vip-instance/\n>f+++++++++ pod-logs/failed-pods/kube-system/kube-vip-instance/kube-vip.txt\ncd+++++++++ pod-logs/failed-pods/local-path-storage/\ncd+++++++++ pod-logs/failed-pods/local-path-storage/local-path-provisioner-679c578f5-m52kp/\n>f+++++++++ pod-logs/failed-pods/local-path-storage/local-path-provisioner-679c578f5-m52kp/local-path-provisioner.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/\ncd+++++++++ pod-logs/failed-pods/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/\n>f+++++++++ pod-logs/failed-pods/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/alertmanager.txt\n>f+++++++++ pod-logs/failed-pods/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/config-reloader.txt\n>f+++++++++ pod-logs/failed-pods/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/init-config-reloader.txt\n>f+++++++++ pod-logs/failed-pods/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/oauth2-proxy.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/goldpinger-gz8xd/\n>f+++++++++ pod-logs/failed-pods/monitoring/goldpinger-gz8xd/goldpinger-daemon.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/\n>f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-dashboard.txt\n>f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-datasources.txt\n>f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj/\n>f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj/kube-state-metrics.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-operator-cc88b86d6-xspgk/\n>f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-operator-cc88b86d6-xspgk/kube-prometheus-stack.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/\n>f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/node-exporter.txt\n>f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/pod-tls-sidecar.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/loki-0/\n>f+++++++++ pod-logs/failed-pods/monitoring/loki-0/loki.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/loki-gateway-5cc45946cd-wp7ks/\n>f+++++++++ pod-logs/failed-pods/monitoring/loki-gateway-5cc45946cd-wp7ks/nginx.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/node-feature-discovery-master-68cf466d5f-qknzk/\n>f+++++++++ pod-logs/failed-pods/monitoring/node-feature-discovery-master-68cf466d5f-qknzk/master.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/node-feature-discovery-worker-7l5z7/\n>f+++++++++ pod-logs/failed-pods/monitoring/node-feature-discovery-worker-7l5z7/worker.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/\n>f+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/config-reloader.txt\n>f+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/init-config-reloader.txt\n>f+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/oauth2-proxy.txt\n>f+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/pod-tls-sidecar.txt\n>f+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/prometheus.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/prometheus-pushgateway-5c694d99cc-jnjt7/\n>f+++++++++ pod-logs/failed-pods/monitoring/prometheus-pushgateway-5c694d99cc-jnjt7/pushgateway.txt\ncd+++++++++ pod-logs/failed-pods/monitoring/vector-75jgf/\n>f+++++++++ pod-logs/failed-pods/monitoring/vector-75jgf/vector.txt\ncd+++++++++ pod-logs/failed-pods/openstack/\ncd+++++++++ pod-logs/failed-pods/openstack/keepalived-wktjf/\n>f+++++++++ pod-logs/failed-pods/openstack/keepalived-wktjf/init.txt\n>f+++++++++ pod-logs/failed-pods/openstack/keepalived-wktjf/keepalived.txt\n>f+++++++++ pod-logs/failed-pods/openstack/keepalived-wktjf/wait-for-ip.txt\ncd+++++++++ pod-logs/failed-pods/openstack/memcached-memcached-c47d949f5-vc5cm/\n>f+++++++++ pod-logs/failed-pods/openstack/memcached-memcached-c47d949f5-vc5cm/init.txt\n>f+++++++++ pod-logs/failed-pods/openstack/memcached-memcached-c47d949f5-vc5cm/memcached-exporter.txt\n>f+++++++++ pod-logs/failed-pods/openstack/memcached-memcached-c47d949f5-vc5cm/memcached.txt\ncd+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-haproxy-0/\n>f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-haproxy-0/haproxy.txt\n>f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-haproxy-0/pxc-init.txt\n>f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-haproxy-0/pxc-monit.txt\ncd+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-pxc-0/\n>f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-pxc-0/exporter.txt\n>f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-pxc-0/pxc-init.txt\n>f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-pxc-0/pxc.txt\ncd+++++++++ pod-logs/failed-pods/openstack/pxc-operator-59d66cbc78-clxvf/\n>f+++++++++ pod-logs/failed-pods/openstack/pxc-operator-59d66cbc78-clxvf/percona-xtradb-cluster-operator.txt\ncd+++++++++ pod-logs/failed-pods/openstack/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf/\n>f+++++++++ pod-logs/failed-pods/openstack/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf/rabbitmq-cluster-operator.txt\ncd+++++++++ pod-logs/failed-pods/openstack/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc/\n>f+++++++++ pod-logs/failed-pods/openstack/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc/rabbitmq-cluster-operator.txt\ncd+++++++++ pod-logs/failed-pods/secretgen-controller/\ncd+++++++++ pod-logs/failed-pods/secretgen-controller/secretgen-controller-5cf976ccc7-vwxz4/\n>f+++++++++ pod-logs/failed-pods/secretgen-controller/secretgen-controller-5cf976ccc7-vwxz4/secretgen-controller.txt\ncd+++++++++ pod-logs/ingress-nginx/\ncd+++++++++ pod-logs/ingress-nginx/ingress-nginx-controller-dmjt4/\n>f+++++++++ pod-logs/ingress-nginx/ingress-nginx-controller-dmjt4/controller.txt\ncd+++++++++ pod-logs/ingress-nginx/ingress-nginx-defaultbackend-6987ff55cf-plsk9/\n>f+++++++++ pod-logs/ingress-nginx/ingress-nginx-defaultbackend-6987ff55cf-plsk9/ingress-nginx-default-backend.txt\ncd+++++++++ pod-logs/kube-system/\ncd+++++++++ pod-logs/kube-system/cilium-operator-869df985b8-nqvtb/\n>f+++++++++ pod-logs/kube-system/cilium-operator-869df985b8-nqvtb/cilium-operator.txt\ncd+++++++++ pod-logs/kube-system/cilium-sk6n5/\n>f+++++++++ pod-logs/kube-system/cilium-sk6n5/apply-sysctl-overwrites.txt\n>f+++++++++ pod-logs/kube-system/cilium-sk6n5/cilium-agent.txt\n>f+++++++++ pod-logs/kube-system/cilium-sk6n5/clean-cilium-state.txt\n>f+++++++++ pod-logs/kube-system/cilium-sk6n5/config.txt\n>f+++++++++ pod-logs/kube-system/cilium-sk6n5/install-cni-binaries.txt\n>f+++++++++ pod-logs/kube-system/cilium-sk6n5/mount-bpf-fs.txt\n>f+++++++++ pod-logs/kube-system/cilium-sk6n5/mount-cgroup.txt\ncd+++++++++ pod-logs/kube-system/coredns-67659f764b-mdrt6/\n>f+++++++++ pod-logs/kube-system/coredns-67659f764b-mdrt6/coredns.txt\ncd+++++++++ pod-logs/kube-system/coredns-67659f764b-xb6lv/\n>f+++++++++ pod-logs/kube-system/coredns-67659f764b-xb6lv/coredns.txt\ncd+++++++++ pod-logs/kube-system/etcd-instance/\n>f+++++++++ pod-logs/kube-system/etcd-instance/etcd.txt\ncd+++++++++ pod-logs/kube-system/kube-apiserver-instance/\n>f+++++++++ pod-logs/kube-system/kube-apiserver-instance/kube-apiserver.txt\ncd+++++++++ pod-logs/kube-system/kube-controller-manager-instance/\n>f+++++++++ pod-logs/kube-system/kube-controller-manager-instance/kube-controller-manager.txt\ncd+++++++++ pod-logs/kube-system/kube-proxy-cxlxv/\n>f+++++++++ pod-logs/kube-system/kube-proxy-cxlxv/kube-proxy.txt\ncd+++++++++ pod-logs/kube-system/kube-scheduler-instance/\n>f+++++++++ pod-logs/kube-system/kube-scheduler-instance/kube-scheduler.txt\ncd+++++++++ pod-logs/kube-system/kube-vip-instance/\n>f+++++++++ pod-logs/kube-system/kube-vip-instance/kube-vip.txt\ncd+++++++++ pod-logs/local-path-storage/\ncd+++++++++ pod-logs/local-path-storage/local-path-provisioner-679c578f5-m52kp/\n>f+++++++++ pod-logs/local-path-storage/local-path-provisioner-679c578f5-m52kp/local-path-provisioner.txt\ncd+++++++++ pod-logs/monitoring/\ncd+++++++++ pod-logs/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/\n>f+++++++++ pod-logs/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/alertmanager.txt\n>f+++++++++ pod-logs/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/config-reloader.txt\n>f+++++++++ pod-logs/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/init-config-reloader.txt\n>f+++++++++ pod-logs/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/oauth2-proxy.txt\ncd+++++++++ pod-logs/monitoring/goldpinger-gz8xd/\n>f+++++++++ pod-logs/monitoring/goldpinger-gz8xd/goldpinger-daemon.txt\ncd+++++++++ pod-logs/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/\n>f+++++++++ pod-logs/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-dashboard.txt\n>f+++++++++ pod-logs/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-datasources.txt\n>f+++++++++ pod-logs/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana.txt\ncd+++++++++ pod-logs/monitoring/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj/\n>f+++++++++ pod-logs/monitoring/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj/kube-state-metrics.txt\ncd+++++++++ pod-logs/monitoring/kube-prometheus-stack-operator-cc88b86d6-xspgk/\n>f+++++++++ pod-logs/monitoring/kube-prometheus-stack-operator-cc88b86d6-xspgk/kube-prometheus-stack.txt\ncd+++++++++ pod-logs/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/\n>f+++++++++ pod-logs/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/node-exporter.txt\n>f+++++++++ pod-logs/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/pod-tls-sidecar.txt\ncd+++++++++ pod-logs/monitoring/loki-0/\n>f+++++++++ pod-logs/monitoring/loki-0/loki.txt\ncd+++++++++ pod-logs/monitoring/loki-gateway-5cc45946cd-wp7ks/\n>f+++++++++ pod-logs/monitoring/loki-gateway-5cc45946cd-wp7ks/nginx.txt\ncd+++++++++ pod-logs/monitoring/node-feature-discovery-master-68cf466d5f-qknzk/\n>f+++++++++ pod-logs/monitoring/node-feature-discovery-master-68cf466d5f-qknzk/master.txt\ncd+++++++++ pod-logs/monitoring/node-feature-discovery-worker-7l5z7/\n>f+++++++++ pod-logs/monitoring/node-feature-discovery-worker-7l5z7/worker.txt\ncd+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/\n>f+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/config-reloader.txt\n>f+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/init-config-reloader.txt\n>f+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/oauth2-proxy.txt\n>f+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/pod-tls-sidecar.txt\n>f+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/prometheus.txt\ncd+++++++++ pod-logs/monitoring/prometheus-pushgateway-5c694d99cc-jnjt7/\n>f+++++++++ pod-logs/monitoring/prometheus-pushgateway-5c694d99cc-jnjt7/pushgateway.txt\ncd+++++++++ pod-logs/monitoring/vector-75jgf/\n>f+++++++++ pod-logs/monitoring/vector-75jgf/vector.txt\ncd+++++++++ pod-logs/openstack/\ncd+++++++++ pod-logs/openstack/keepalived-wktjf/\n>f+++++++++ pod-logs/openstack/keepalived-wktjf/init.txt\n>f+++++++++ pod-logs/openstack/keepalived-wktjf/keepalived.txt\n>f+++++++++ pod-logs/openstack/keepalived-wktjf/wait-for-ip.txt\ncd+++++++++ pod-logs/openstack/memcached-memcached-c47d949f5-vc5cm/\n>f+++++++++ pod-logs/openstack/memcached-memcached-c47d949f5-vc5cm/init.txt\n>f+++++++++ pod-logs/openstack/memcached-memcached-c47d949f5-vc5cm/memcached-exporter.txt\n>f+++++++++ pod-logs/openstack/memcached-memcached-c47d949f5-vc5cm/memcached.txt\ncd+++++++++ pod-logs/openstack/percona-xtradb-haproxy-0/\n>f+++++++++ pod-logs/openstack/percona-xtradb-haproxy-0/haproxy.txt\n>f+++++++++ pod-logs/openstack/percona-xtradb-haproxy-0/pxc-init.txt\n>f+++++++++ pod-logs/openstack/percona-xtradb-haproxy-0/pxc-monit.txt\ncd+++++++++ pod-logs/openstack/percona-xtradb-pxc-0/\n>f+++++++++ pod-logs/openstack/percona-xtradb-pxc-0/exporter.txt\n>f+++++++++ pod-logs/openstack/percona-xtradb-pxc-0/pxc-init.txt\n>f+++++++++ pod-logs/openstack/percona-xtradb-pxc-0/pxc.txt\ncd+++++++++ pod-logs/openstack/pxc-operator-59d66cbc78-clxvf/\n>f+++++++++ pod-logs/openstack/pxc-operator-59d66cbc78-clxvf/percona-xtradb-cluster-operator.txt\ncd+++++++++ pod-logs/openstack/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf/\n>f+++++++++ pod-logs/openstack/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf/rabbitmq-cluster-operator.txt\ncd+++++++++ pod-logs/openstack/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc/\n>f+++++++++ pod-logs/openstack/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc/rabbitmq-cluster-operator.txt\ncd+++++++++ pod-logs/secretgen-controller/\ncd+++++++++ pod-logs/secretgen-controller/secretgen-controller-5cf976ccc7-vwxz4/\n>f+++++++++ pod-logs/secretgen-controller/secretgen-controller-5cf976ccc7-vwxz4/secretgen-controller.txt\n",
                            "rc": 0,
                            "stdout_lines": [
                                "cd+++++++++ pod-logs/",
                                "cd+++++++++ pod-logs/auth-system/",
                                "cd+++++++++ pod-logs/auth-system/keycloak-0/",
                                ">f+++++++++ pod-logs/auth-system/keycloak-0/keycloak.txt",
                                "cd+++++++++ pod-logs/cert-manager/",
                                "cd+++++++++ pod-logs/cert-manager/cert-manager-75969b45cf-d5c4v/",
                                ">f+++++++++ pod-logs/cert-manager/cert-manager-75969b45cf-d5c4v/cert-manager-controller.txt",
                                "cd+++++++++ pod-logs/cert-manager/cert-manager-cainjector-84f45b698d-v64n5/",
                                ">f+++++++++ pod-logs/cert-manager/cert-manager-cainjector-84f45b698d-v64n5/cert-manager-cainjector.txt",
                                "cd+++++++++ pod-logs/cert-manager/cert-manager-webhook-9997c64fc-8w8sz/",
                                ">f+++++++++ pod-logs/cert-manager/cert-manager-webhook-9997c64fc-8w8sz/cert-manager-webhook.txt",
                                "cd+++++++++ pod-logs/envoy-gateway-system/",
                                "cd+++++++++ pod-logs/envoy-gateway-system/envoy-gateway-78446f96c9-7zvgb/",
                                ">f+++++++++ pod-logs/envoy-gateway-system/envoy-gateway-78446f96c9-7zvgb/envoy-gateway.txt",
                                "cd+++++++++ pod-logs/failed-pods/",
                                "cd+++++++++ pod-logs/failed-pods/auth-system/",
                                "cd+++++++++ pod-logs/failed-pods/auth-system/keycloak-0/",
                                ">f+++++++++ pod-logs/failed-pods/auth-system/keycloak-0/keycloak.txt",
                                "cd+++++++++ pod-logs/failed-pods/cert-manager/",
                                "cd+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-75969b45cf-d5c4v/",
                                ">f+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-75969b45cf-d5c4v/cert-manager-controller.txt",
                                "cd+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-cainjector-84f45b698d-v64n5/",
                                ">f+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-cainjector-84f45b698d-v64n5/cert-manager-cainjector.txt",
                                "cd+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-webhook-9997c64fc-8w8sz/",
                                ">f+++++++++ pod-logs/failed-pods/cert-manager/cert-manager-webhook-9997c64fc-8w8sz/cert-manager-webhook.txt",
                                "cd+++++++++ pod-logs/failed-pods/envoy-gateway-system/",
                                "cd+++++++++ pod-logs/failed-pods/envoy-gateway-system/envoy-gateway-78446f96c9-7zvgb/",
                                ">f+++++++++ pod-logs/failed-pods/envoy-gateway-system/envoy-gateway-78446f96c9-7zvgb/envoy-gateway.txt",
                                "cd+++++++++ pod-logs/failed-pods/ingress-nginx/",
                                "cd+++++++++ pod-logs/failed-pods/ingress-nginx/ingress-nginx-controller-dmjt4/",
                                ">f+++++++++ pod-logs/failed-pods/ingress-nginx/ingress-nginx-controller-dmjt4/controller.txt",
                                "cd+++++++++ pod-logs/failed-pods/ingress-nginx/ingress-nginx-defaultbackend-6987ff55cf-plsk9/",
                                ">f+++++++++ pod-logs/failed-pods/ingress-nginx/ingress-nginx-defaultbackend-6987ff55cf-plsk9/ingress-nginx-default-backend.txt",
                                "cd+++++++++ pod-logs/failed-pods/kube-system/",
                                "cd+++++++++ pod-logs/failed-pods/kube-system/cilium-operator-869df985b8-nqvtb/",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/cilium-operator-869df985b8-nqvtb/cilium-operator.txt",
                                "cd+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/apply-sysctl-overwrites.txt",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/cilium-agent.txt",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/clean-cilium-state.txt",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/config.txt",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/install-cni-binaries.txt",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/mount-bpf-fs.txt",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/cilium-sk6n5/mount-cgroup.txt",
                                "cd+++++++++ pod-logs/failed-pods/kube-system/coredns-67659f764b-mdrt6/",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/coredns-67659f764b-mdrt6/coredns.txt",
                                "cd+++++++++ pod-logs/failed-pods/kube-system/coredns-67659f764b-xb6lv/",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/coredns-67659f764b-xb6lv/coredns.txt",
                                "cd+++++++++ pod-logs/failed-pods/kube-system/etcd-instance/",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/etcd-instance/etcd.txt",
                                "cd+++++++++ pod-logs/failed-pods/kube-system/kube-apiserver-instance/",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/kube-apiserver-instance/kube-apiserver.txt",
                                "cd+++++++++ pod-logs/failed-pods/kube-system/kube-controller-manager-instance/",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/kube-controller-manager-instance/kube-controller-manager.txt",
                                "cd+++++++++ pod-logs/failed-pods/kube-system/kube-proxy-cxlxv/",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/kube-proxy-cxlxv/kube-proxy.txt",
                                "cd+++++++++ pod-logs/failed-pods/kube-system/kube-scheduler-instance/",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/kube-scheduler-instance/kube-scheduler.txt",
                                "cd+++++++++ pod-logs/failed-pods/kube-system/kube-vip-instance/",
                                ">f+++++++++ pod-logs/failed-pods/kube-system/kube-vip-instance/kube-vip.txt",
                                "cd+++++++++ pod-logs/failed-pods/local-path-storage/",
                                "cd+++++++++ pod-logs/failed-pods/local-path-storage/local-path-provisioner-679c578f5-m52kp/",
                                ">f+++++++++ pod-logs/failed-pods/local-path-storage/local-path-provisioner-679c578f5-m52kp/local-path-provisioner.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/alertmanager.txt",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/config-reloader.txt",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/init-config-reloader.txt",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/oauth2-proxy.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/goldpinger-gz8xd/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/goldpinger-gz8xd/goldpinger-daemon.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-dashboard.txt",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-datasources.txt",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj/kube-state-metrics.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-operator-cc88b86d6-xspgk/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-operator-cc88b86d6-xspgk/kube-prometheus-stack.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/node-exporter.txt",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/pod-tls-sidecar.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/loki-0/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/loki-0/loki.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/loki-gateway-5cc45946cd-wp7ks/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/loki-gateway-5cc45946cd-wp7ks/nginx.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/node-feature-discovery-master-68cf466d5f-qknzk/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/node-feature-discovery-master-68cf466d5f-qknzk/master.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/node-feature-discovery-worker-7l5z7/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/node-feature-discovery-worker-7l5z7/worker.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/config-reloader.txt",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/init-config-reloader.txt",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/oauth2-proxy.txt",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/pod-tls-sidecar.txt",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/prometheus-kube-prometheus-stack-prometheus-0/prometheus.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/prometheus-pushgateway-5c694d99cc-jnjt7/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/prometheus-pushgateway-5c694d99cc-jnjt7/pushgateway.txt",
                                "cd+++++++++ pod-logs/failed-pods/monitoring/vector-75jgf/",
                                ">f+++++++++ pod-logs/failed-pods/monitoring/vector-75jgf/vector.txt",
                                "cd+++++++++ pod-logs/failed-pods/openstack/",
                                "cd+++++++++ pod-logs/failed-pods/openstack/keepalived-wktjf/",
                                ">f+++++++++ pod-logs/failed-pods/openstack/keepalived-wktjf/init.txt",
                                ">f+++++++++ pod-logs/failed-pods/openstack/keepalived-wktjf/keepalived.txt",
                                ">f+++++++++ pod-logs/failed-pods/openstack/keepalived-wktjf/wait-for-ip.txt",
                                "cd+++++++++ pod-logs/failed-pods/openstack/memcached-memcached-c47d949f5-vc5cm/",
                                ">f+++++++++ pod-logs/failed-pods/openstack/memcached-memcached-c47d949f5-vc5cm/init.txt",
                                ">f+++++++++ pod-logs/failed-pods/openstack/memcached-memcached-c47d949f5-vc5cm/memcached-exporter.txt",
                                ">f+++++++++ pod-logs/failed-pods/openstack/memcached-memcached-c47d949f5-vc5cm/memcached.txt",
                                "cd+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-haproxy-0/",
                                ">f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-haproxy-0/haproxy.txt",
                                ">f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-haproxy-0/pxc-init.txt",
                                ">f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-haproxy-0/pxc-monit.txt",
                                "cd+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-pxc-0/",
                                ">f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-pxc-0/exporter.txt",
                                ">f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-pxc-0/pxc-init.txt",
                                ">f+++++++++ pod-logs/failed-pods/openstack/percona-xtradb-pxc-0/pxc.txt",
                                "cd+++++++++ pod-logs/failed-pods/openstack/pxc-operator-59d66cbc78-clxvf/",
                                ">f+++++++++ pod-logs/failed-pods/openstack/pxc-operator-59d66cbc78-clxvf/percona-xtradb-cluster-operator.txt",
                                "cd+++++++++ pod-logs/failed-pods/openstack/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf/",
                                ">f+++++++++ pod-logs/failed-pods/openstack/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf/rabbitmq-cluster-operator.txt",
                                "cd+++++++++ pod-logs/failed-pods/openstack/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc/",
                                ">f+++++++++ pod-logs/failed-pods/openstack/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc/rabbitmq-cluster-operator.txt",
                                "cd+++++++++ pod-logs/failed-pods/secretgen-controller/",
                                "cd+++++++++ pod-logs/failed-pods/secretgen-controller/secretgen-controller-5cf976ccc7-vwxz4/",
                                ">f+++++++++ pod-logs/failed-pods/secretgen-controller/secretgen-controller-5cf976ccc7-vwxz4/secretgen-controller.txt",
                                "cd+++++++++ pod-logs/ingress-nginx/",
                                "cd+++++++++ pod-logs/ingress-nginx/ingress-nginx-controller-dmjt4/",
                                ">f+++++++++ pod-logs/ingress-nginx/ingress-nginx-controller-dmjt4/controller.txt",
                                "cd+++++++++ pod-logs/ingress-nginx/ingress-nginx-defaultbackend-6987ff55cf-plsk9/",
                                ">f+++++++++ pod-logs/ingress-nginx/ingress-nginx-defaultbackend-6987ff55cf-plsk9/ingress-nginx-default-backend.txt",
                                "cd+++++++++ pod-logs/kube-system/",
                                "cd+++++++++ pod-logs/kube-system/cilium-operator-869df985b8-nqvtb/",
                                ">f+++++++++ pod-logs/kube-system/cilium-operator-869df985b8-nqvtb/cilium-operator.txt",
                                "cd+++++++++ pod-logs/kube-system/cilium-sk6n5/",
                                ">f+++++++++ pod-logs/kube-system/cilium-sk6n5/apply-sysctl-overwrites.txt",
                                ">f+++++++++ pod-logs/kube-system/cilium-sk6n5/cilium-agent.txt",
                                ">f+++++++++ pod-logs/kube-system/cilium-sk6n5/clean-cilium-state.txt",
                                ">f+++++++++ pod-logs/kube-system/cilium-sk6n5/config.txt",
                                ">f+++++++++ pod-logs/kube-system/cilium-sk6n5/install-cni-binaries.txt",
                                ">f+++++++++ pod-logs/kube-system/cilium-sk6n5/mount-bpf-fs.txt",
                                ">f+++++++++ pod-logs/kube-system/cilium-sk6n5/mount-cgroup.txt",
                                "cd+++++++++ pod-logs/kube-system/coredns-67659f764b-mdrt6/",
                                ">f+++++++++ pod-logs/kube-system/coredns-67659f764b-mdrt6/coredns.txt",
                                "cd+++++++++ pod-logs/kube-system/coredns-67659f764b-xb6lv/",
                                ">f+++++++++ pod-logs/kube-system/coredns-67659f764b-xb6lv/coredns.txt",
                                "cd+++++++++ pod-logs/kube-system/etcd-instance/",
                                ">f+++++++++ pod-logs/kube-system/etcd-instance/etcd.txt",
                                "cd+++++++++ pod-logs/kube-system/kube-apiserver-instance/",
                                ">f+++++++++ pod-logs/kube-system/kube-apiserver-instance/kube-apiserver.txt",
                                "cd+++++++++ pod-logs/kube-system/kube-controller-manager-instance/",
                                ">f+++++++++ pod-logs/kube-system/kube-controller-manager-instance/kube-controller-manager.txt",
                                "cd+++++++++ pod-logs/kube-system/kube-proxy-cxlxv/",
                                ">f+++++++++ pod-logs/kube-system/kube-proxy-cxlxv/kube-proxy.txt",
                                "cd+++++++++ pod-logs/kube-system/kube-scheduler-instance/",
                                ">f+++++++++ pod-logs/kube-system/kube-scheduler-instance/kube-scheduler.txt",
                                "cd+++++++++ pod-logs/kube-system/kube-vip-instance/",
                                ">f+++++++++ pod-logs/kube-system/kube-vip-instance/kube-vip.txt",
                                "cd+++++++++ pod-logs/local-path-storage/",
                                "cd+++++++++ pod-logs/local-path-storage/local-path-provisioner-679c578f5-m52kp/",
                                ">f+++++++++ pod-logs/local-path-storage/local-path-provisioner-679c578f5-m52kp/local-path-provisioner.txt",
                                "cd+++++++++ pod-logs/monitoring/",
                                "cd+++++++++ pod-logs/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/",
                                ">f+++++++++ pod-logs/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/alertmanager.txt",
                                ">f+++++++++ pod-logs/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/config-reloader.txt",
                                ">f+++++++++ pod-logs/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/init-config-reloader.txt",
                                ">f+++++++++ pod-logs/monitoring/alertmanager-kube-prometheus-stack-alertmanager-0/oauth2-proxy.txt",
                                "cd+++++++++ pod-logs/monitoring/goldpinger-gz8xd/",
                                ">f+++++++++ pod-logs/monitoring/goldpinger-gz8xd/goldpinger-daemon.txt",
                                "cd+++++++++ pod-logs/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/",
                                ">f+++++++++ pod-logs/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-dashboard.txt",
                                ">f+++++++++ pod-logs/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana-sc-datasources.txt",
                                ">f+++++++++ pod-logs/monitoring/kube-prometheus-stack-grafana-69fc56495c-5h4gd/grafana.txt",
                                "cd+++++++++ pod-logs/monitoring/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj/",
                                ">f+++++++++ pod-logs/monitoring/kube-prometheus-stack-kube-state-metrics-c7bd54d7-twjsj/kube-state-metrics.txt",
                                "cd+++++++++ pod-logs/monitoring/kube-prometheus-stack-operator-cc88b86d6-xspgk/",
                                ">f+++++++++ pod-logs/monitoring/kube-prometheus-stack-operator-cc88b86d6-xspgk/kube-prometheus-stack.txt",
                                "cd+++++++++ pod-logs/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/",
                                ">f+++++++++ pod-logs/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/node-exporter.txt",
                                ">f+++++++++ pod-logs/monitoring/kube-prometheus-stack-prometheus-node-exporter-jmj6n/pod-tls-sidecar.txt",
                                "cd+++++++++ pod-logs/monitoring/loki-0/",
                                ">f+++++++++ pod-logs/monitoring/loki-0/loki.txt",
                                "cd+++++++++ pod-logs/monitoring/loki-gateway-5cc45946cd-wp7ks/",
                                ">f+++++++++ pod-logs/monitoring/loki-gateway-5cc45946cd-wp7ks/nginx.txt",
                                "cd+++++++++ pod-logs/monitoring/node-feature-discovery-master-68cf466d5f-qknzk/",
                                ">f+++++++++ pod-logs/monitoring/node-feature-discovery-master-68cf466d5f-qknzk/master.txt",
                                "cd+++++++++ pod-logs/monitoring/node-feature-discovery-worker-7l5z7/",
                                ">f+++++++++ pod-logs/monitoring/node-feature-discovery-worker-7l5z7/worker.txt",
                                "cd+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/",
                                ">f+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/config-reloader.txt",
                                ">f+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/init-config-reloader.txt",
                                ">f+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/oauth2-proxy.txt",
                                ">f+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/pod-tls-sidecar.txt",
                                ">f+++++++++ pod-logs/monitoring/prometheus-kube-prometheus-stack-prometheus-0/prometheus.txt",
                                "cd+++++++++ pod-logs/monitoring/prometheus-pushgateway-5c694d99cc-jnjt7/",
                                ">f+++++++++ pod-logs/monitoring/prometheus-pushgateway-5c694d99cc-jnjt7/pushgateway.txt",
                                "cd+++++++++ pod-logs/monitoring/vector-75jgf/",
                                ">f+++++++++ pod-logs/monitoring/vector-75jgf/vector.txt",
                                "cd+++++++++ pod-logs/openstack/",
                                "cd+++++++++ pod-logs/openstack/keepalived-wktjf/",
                                ">f+++++++++ pod-logs/openstack/keepalived-wktjf/init.txt",
                                ">f+++++++++ pod-logs/openstack/keepalived-wktjf/keepalived.txt",
                                ">f+++++++++ pod-logs/openstack/keepalived-wktjf/wait-for-ip.txt",
                                "cd+++++++++ pod-logs/openstack/memcached-memcached-c47d949f5-vc5cm/",
                                ">f+++++++++ pod-logs/openstack/memcached-memcached-c47d949f5-vc5cm/init.txt",
                                ">f+++++++++ pod-logs/openstack/memcached-memcached-c47d949f5-vc5cm/memcached-exporter.txt",
                                ">f+++++++++ pod-logs/openstack/memcached-memcached-c47d949f5-vc5cm/memcached.txt",
                                "cd+++++++++ pod-logs/openstack/percona-xtradb-haproxy-0/",
                                ">f+++++++++ pod-logs/openstack/percona-xtradb-haproxy-0/haproxy.txt",
                                ">f+++++++++ pod-logs/openstack/percona-xtradb-haproxy-0/pxc-init.txt",
                                ">f+++++++++ pod-logs/openstack/percona-xtradb-haproxy-0/pxc-monit.txt",
                                "cd+++++++++ pod-logs/openstack/percona-xtradb-pxc-0/",
                                ">f+++++++++ pod-logs/openstack/percona-xtradb-pxc-0/exporter.txt",
                                ">f+++++++++ pod-logs/openstack/percona-xtradb-pxc-0/pxc-init.txt",
                                ">f+++++++++ pod-logs/openstack/percona-xtradb-pxc-0/pxc.txt",
                                "cd+++++++++ pod-logs/openstack/pxc-operator-59d66cbc78-clxvf/",
                                ">f+++++++++ pod-logs/openstack/pxc-operator-59d66cbc78-clxvf/percona-xtradb-cluster-operator.txt",
                                "cd+++++++++ pod-logs/openstack/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf/",
                                ">f+++++++++ pod-logs/openstack/rabbitmq-cluster-operator-6b5c8b7cdc-6w8jf/rabbitmq-cluster-operator.txt",
                                "cd+++++++++ pod-logs/openstack/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc/",
                                ">f+++++++++ pod-logs/openstack/rabbitmq-messaging-topology-operator-69996b7f7f-52mnc/rabbitmq-cluster-operator.txt",
                                "cd+++++++++ pod-logs/secretgen-controller/",
                                "cd+++++++++ pod-logs/secretgen-controller/secretgen-controller-5cf976ccc7-vwxz4/",
                                ">f+++++++++ pod-logs/secretgen-controller/secretgen-controller-5cf976ccc7-vwxz4/secretgen-controller.txt"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-00000000001f",
                        "name": "gather-pod-logs",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-pod-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:44.422950Z",
                            "start": "2026-04-04T10:26:43.872146Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000024",
                        "name": "Downloads pod logs to executor"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/tmp/logs/prometheus",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/tmp/logs/prometheus",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": null,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/tmp/logs/prometheus",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/tmp/logs/prometheus",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000026",
                        "name": "gather-prom-metrics",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-prom-metrics"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:44.644862Z",
                            "start": "2026-04-04T10:26:44.436673Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000028",
                        "name": "creating directory for helm release descriptions"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "shell",
                            "changed": true,
                            "cmd": "set -e\nNAMESPACES=$(kubectl get namespaces -o json | jq -r '.items[].metadata.name')\nfor NS in $NAMESPACES; do\n  SERVICES=$(kubectl get svc -n $NS -o json | jq -r '.items[] | select(.spec.ports[].name==\"metrics\") | .metadata.name')\n  for SVC in $SERVICES; do\n    PORT=$(kubectl get svc $SVC -n $NS -o json | jq -r '.spec.ports[] | select(.name==\"metrics\") | .port')\n    echo \"Scraping $SVC.$NS:$PORT/metrics:\"\n    curl \"$SVC.$NS:$PORT/metrics\" >> \"/tmp/logs\"/prometheus/$NS-$SVC.txt || true\n  done\ndone",
                            "delta": "0:00:01.347876",
                            "end": "2026-04-04 10:26:46.225628",
                            "failed": true,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -e\nNAMESPACES=$(kubectl get namespaces -o json | jq -r '.items[].metadata.name')\nfor NS in $NAMESPACES; do\n  SERVICES=$(kubectl get svc -n $NS -o json | jq -r '.items[] | select(.spec.ports[].name==\"metrics\") | .metadata.name')\n  for SVC in $SERVICES; do\n    PORT=$(kubectl get svc $SVC -n $NS -o json | jq -r '.spec.ports[] | select(.name==\"metrics\") | .port')\n    echo \"Scraping $SVC.$NS:$PORT/metrics:\"\n    curl \"$SVC.$NS:$PORT/metrics\" >> \"/tmp/logs\"/prometheus/$NS-$SVC.txt || true\n  done\ndone",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-cc15-c3f6-000000000029-1-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "non-zero return code",
                            "rc": 5,
                            "start": "2026-04-04 10:26:44.877752",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Scraping envoy-gateway.envoy-gateway-system:19001/metrics:\n  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current\n                                 Dload  Upload   Total   Spent    Left  Speed\n\r  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0curl: (6) Could not resolve host: envoy-gateway.envoy-gateway-system\nScraping ingress-nginx-controller-metrics.ingress-nginx:10254/metrics:\n  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current\n                                 Dload  Upload   Total   Spent    Left  Speed\n\r  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0curl: (6) Could not resolve host: ingress-nginx-controller-metrics.ingress-nginx\nScraping kube-dns.kube-system:9153/metrics:\n  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current\n                                 Dload  Upload   Total   Spent    Left  Speed\n\r  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0curl: (6) Could not resolve host: kube-dns.kube-system\njq: error (at <stdin>:888): Cannot iterate over null (null)",
                            "stdout_lines": [
                                "Scraping envoy-gateway.envoy-gateway-system:19001/metrics:",
                                "  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current",
                                "                                 Dload  Upload   Total   Spent    Left  Speed",
                                "",
                                "  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0curl: (6) Could not resolve host: envoy-gateway.envoy-gateway-system",
                                "Scraping ingress-nginx-controller-metrics.ingress-nginx:10254/metrics:",
                                "  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current",
                                "                                 Dload  Upload   Total   Spent    Left  Speed",
                                "",
                                "  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0curl: (6) Could not resolve host: ingress-nginx-controller-metrics.ingress-nginx",
                                "Scraping kube-dns.kube-system:9153/metrics:",
                                "  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current",
                                "                                 Dload  Upload   Total   Spent    Left  Speed",
                                "",
                                "  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0curl: (6) Could not resolve host: kube-dns.kube-system",
                                "jq: error (at <stdin>:888): Cannot iterate over null (null)"
                            ],
                            "zuul_log_id": "0242ac17-0011-cc15-c3f6-000000000029-1-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000026",
                        "name": "gather-prom-metrics",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-prom-metrics"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:46.695092Z",
                            "start": "2026-04-04T10:26:44.669932Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000029",
                        "name": "Get metrics from exporter services in all namespaces"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "shell",
                            "changed": true,
                            "cmd": "set -e\nmgr_endpoints=$(kubectl get endpoints -n ceph -l component=manager -o json | jq -r '.items[].subsets[].addresses[].ip')\necho \"ceph-mgr endpoints: $mgr_endpoints\"\nfor endpoint in $mgr_endpoints; do\n  echo \"checking ceph-mgr at $endpoint\"\n  metrics_curl=\"curl $endpoint:9283/metrics\"\n  op=$(eval \"$metrics_curl\")\n  if [[ -n $op ]]; then\n    curl $endpoint:9283/metrics >> \"/tmp/logs\"/prometheus/ceph-ceph-mgr.txt\n    break\n  else\n    echo \"$endpoint is a standby ceph-mgr. Trying next endpoint\"\n  fi\ndone",
                            "delta": "0:00:00.090315",
                            "end": "2026-04-04 10:26:47.010367",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -e\nmgr_endpoints=$(kubectl get endpoints -n ceph -l component=manager -o json | jq -r '.items[].subsets[].addresses[].ip')\necho \"ceph-mgr endpoints: $mgr_endpoints\"\nfor endpoint in $mgr_endpoints; do\n  echo \"checking ceph-mgr at $endpoint\"\n  metrics_curl=\"curl $endpoint:9283/metrics\"\n  op=$(eval \"$metrics_curl\")\n  if [[ -n $op ]]; then\n    curl $endpoint:9283/metrics >> \"/tmp/logs\"/prometheus/ceph-ceph-mgr.txt\n    break\n  else\n    echo \"$endpoint is a standby ceph-mgr. Trying next endpoint\"\n  fi\ndone",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-cc15-c3f6-00000000002a-1-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-04-04 10:26:46.920052",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "ceph-mgr endpoints: ",
                            "stdout_lines": [
                                "ceph-mgr endpoints: "
                            ],
                            "zuul_log_id": "0242ac17-0011-cc15-c3f6-00000000002a-1-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000026",
                        "name": "gather-prom-metrics",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-prom-metrics"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:47.241657Z",
                            "start": "2026-04-04T10:26:46.722209Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-00000000002a",
                        "name": "Get ceph metrics from ceph-mgr"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "shell",
                            "changed": true,
                            "cmd": "set -e\nNAMESPACE=\"osh-infra\"\nAPP_LABEL=\"fluentd\"\nPODS=$(kubectl get pods -n $NAMESPACE -l application=$APP_LABEL -o json | jq -r '.items[].metadata.name')\nfor POD in $PODS; do\n  IP=$(kubectl get pod -n $NAMESPACE $POD -o json | jq -r '.status.podIP')\n  PORT=$(kubectl get pod -n $NAMESPACE $POD -o json |  jq -r '.spec.containers[0].ports[] | select(.name==\"metrics\") | .containerPort')\n  echo \"Scraping $POD at $IP:$PORT/metrics\"\n  curl \"$IP:$PORT/metrics\" >> \"/tmp/logs\"/prometheus/$POD.txt || true\ndone",
                            "delta": "0:00:00.094354",
                            "end": "2026-04-04 10:26:47.597962",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -e\nNAMESPACE=\"osh-infra\"\nAPP_LABEL=\"fluentd\"\nPODS=$(kubectl get pods -n $NAMESPACE -l application=$APP_LABEL -o json | jq -r '.items[].metadata.name')\nfor POD in $PODS; do\n  IP=$(kubectl get pod -n $NAMESPACE $POD -o json | jq -r '.status.podIP')\n  PORT=$(kubectl get pod -n $NAMESPACE $POD -o json |  jq -r '.spec.containers[0].ports[] | select(.name==\"metrics\") | .containerPort')\n  echo \"Scraping $POD at $IP:$PORT/metrics\"\n  curl \"$IP:$PORT/metrics\" >> \"/tmp/logs\"/prometheus/$POD.txt || true\ndone",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-cc15-c3f6-00000000002b-1-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-04-04 10:26:47.503608",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac17-0011-cc15-c3f6-00000000002b-1-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000026",
                        "name": "gather-prom-metrics",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-prom-metrics"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:47.834195Z",
                            "start": "2026-04-04T10:26:47.273196Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-00000000002b",
                        "name": "Get metrics from fluentd pods"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "synchronize",
                            "changed": true,
                            "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --rsh='/usr/bin/ssh -S none -o Port=22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null' --rsync-path='sudo -u root rsync' --out-format='<<CHANGED>>%i %n%L' zuul@199.204.45.153:/tmp/logs/prometheus /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                            "invocation": {
                                "module_args": {
                                    "_local_rsync_password": null,
                                    "_local_rsync_path": "rsync",
                                    "_ssh_args": null,
                                    "_substitute_controller": false,
                                    "archive": true,
                                    "checksum": false,
                                    "compress": true,
                                    "copy_links": false,
                                    "delay_updates": true,
                                    "delete": false,
                                    "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                                    "dest_port": 22,
                                    "dirs": false,
                                    "existing_only": false,
                                    "group": null,
                                    "link_dest": null,
                                    "links": null,
                                    "mode": "pull",
                                    "owner": null,
                                    "partial": false,
                                    "perms": null,
                                    "private_key": null,
                                    "recursive": null,
                                    "rsync_opts": [],
                                    "rsync_path": "sudo -u root rsync",
                                    "rsync_timeout": 0,
                                    "set_remote_user": true,
                                    "src": "zuul@199.204.45.153:/tmp/logs/prometheus",
                                    "ssh_connection_multiplexing": false,
                                    "times": null,
                                    "use_ssh_args": false,
                                    "verify_host": false
                                }
                            },
                            "msg": "cd+++++++++ prometheus/\n>f+++++++++ prometheus/envoy-gateway-system-envoy-gateway.txt\n>f+++++++++ prometheus/ingress-nginx-ingress-nginx-controller-metrics.txt\n>f+++++++++ prometheus/kube-system-kube-dns.txt\n",
                            "rc": 0,
                            "stdout_lines": [
                                "cd+++++++++ prometheus/",
                                ">f+++++++++ prometheus/envoy-gateway-system-envoy-gateway.txt",
                                ">f+++++++++ prometheus/ingress-nginx-ingress-nginx-controller-metrics.txt",
                                ">f+++++++++ prometheus/kube-system-kube-dns.txt"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-000000000026",
                        "name": "gather-prom-metrics",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-prom-metrics"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:48.342984Z",
                            "start": "2026-04-04T10:26:47.839046Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-00000000002c",
                        "name": "Downloads logs to executor"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/tmp/logs/selenium",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/tmp/logs/selenium",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": null,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/tmp/logs/selenium",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/tmp/logs/selenium",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-00000000002e",
                        "name": "gather-selenium-data",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-selenium-data"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:48.552451Z",
                            "start": "2026-04-04T10:26:48.354906Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000030",
                        "name": "creating directory for helm release descriptions"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "shell",
                            "changed": true,
                            "cmd": "set -x\ncp /tmp/artifacts/* /tmp/logs/selenium/.",
                            "delta": "0:00:00.008354",
                            "end": "2026-04-04 10:26:48.779372",
                            "failed": true,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -x\ncp /tmp/artifacts/* /tmp/logs/selenium/.",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac17-0011-cc15-c3f6-000000000031-1-instance",
                                    "zuul_no_log": false,
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "non-zero return code",
                            "rc": 1,
                            "start": "2026-04-04 10:26:48.771018",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ cp '/tmp/artifacts/*' /tmp/logs/selenium/.\ncp: cannot stat '/tmp/artifacts/*': No such file or directory",
                            "stdout_lines": [
                                "+ cp '/tmp/artifacts/*' /tmp/logs/selenium/.",
                                "cp: cannot stat '/tmp/artifacts/*': No such file or directory"
                            ],
                            "zuul_log_id": "0242ac17-0011-cc15-c3f6-000000000031-1-instance"
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-00000000002e",
                        "name": "gather-selenium-data",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-selenium-data"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:49.097181Z",
                            "start": "2026-04-04T10:26:48.576763Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000031",
                        "name": "Get selenium data"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "synchronize",
                            "changed": true,
                            "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --rsh='/usr/bin/ssh -S none -o Port=22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null' --rsync-path='sudo -u root rsync' --out-format='<<CHANGED>>%i %n%L' zuul@199.204.45.153:/tmp/logs/selenium /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                            "invocation": {
                                "module_args": {
                                    "_local_rsync_password": null,
                                    "_local_rsync_path": "rsync",
                                    "_ssh_args": null,
                                    "_substitute_controller": false,
                                    "archive": true,
                                    "checksum": false,
                                    "compress": true,
                                    "copy_links": false,
                                    "delay_updates": true,
                                    "delete": false,
                                    "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/instance",
                                    "dest_port": 22,
                                    "dirs": false,
                                    "existing_only": false,
                                    "group": null,
                                    "link_dest": null,
                                    "links": null,
                                    "mode": "pull",
                                    "owner": null,
                                    "partial": false,
                                    "perms": null,
                                    "private_key": null,
                                    "recursive": null,
                                    "rsync_opts": [],
                                    "rsync_path": "sudo -u root rsync",
                                    "rsync_timeout": 0,
                                    "set_remote_user": true,
                                    "src": "zuul@199.204.45.153:/tmp/logs/selenium",
                                    "ssh_connection_multiplexing": false,
                                    "times": null,
                                    "use_ssh_args": false,
                                    "verify_host": false
                                }
                            },
                            "msg": "cd+++++++++ selenium/\n",
                            "rc": 0,
                            "stdout_lines": [
                                "cd+++++++++ selenium/"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-cc15-c3f6-00000000002e",
                        "name": "gather-selenium-data",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/untrusted/project_4/opendev.org/openstack/openstack-helm/roles/gather-selenium-data"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:49.592210Z",
                            "start": "2026-04-04T10:26:49.102090Z"
                        },
                        "id": "0242ac17-0011-cc15-c3f6-000000000032",
                        "name": "Downloads logs to executor"
                    }
                }
            ]
        }
    ],
    "stats": {
        "instance": {
            "changed": 23,
            "failures": 0,
            "ignored": 3,
            "ok": 23,
            "rescued": 0,
            "skipped": 0,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "1",
    "phase": "post",
    "playbook": "github.com/vexxhost/zuul-config/playbooks/base/post.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-04-04T10:26:53.609835Z",
                    "start": "2026-04-04T10:26:50.398984Z"
                },
                "id": "0242ac17-0011-b0f8-52b5-000000000002",
                "name": "all"
            },
            "tasks": [
                {
                    "hosts": {
                        "instance": {
                            "action": "set_fact",
                            "changed": false,
                            "false_condition": "groups['all'] | length > 1",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-b0f8-52b5-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:50.466154Z",
                            "start": "2026-04-04T10:26:50.410677Z"
                        },
                        "id": "0242ac17-0011-b0f8-52b5-000000000008",
                        "name": "Set log path for multiple nodes"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "log_path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-b0f8-52b5-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:50.510564Z",
                            "start": "2026-04-04T10:26:50.473006Z"
                        },
                        "id": "0242ac17-0011-b0f8-52b5-000000000009",
                        "name": "Set log path for single node"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "file",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-b0f8-52b5-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:51.323343Z",
                            "start": "2026-04-04T10:26:50.515319Z"
                        },
                        "id": "0242ac17-0011-b0f8-52b5-00000000000a",
                        "name": "Ensure local output dirs"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "synchronize",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul@199.204.45.153:/home/zuul/zuul-output/logs/ /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul@199.204.45.153:/home/zuul/zuul-output/logs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs",
                                        "src": "logs"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul@199.204.45.153:/home/zuul/zuul-output/artifacts/ /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul@199.204.45.153:/home/zuul/zuul-output/artifacts/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts",
                                        "src": "artifacts"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul@199.204.45.153:/home/zuul/zuul-output/docs/ /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul@199.204.45.153:/home/zuul/zuul-output/docs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs",
                                        "src": "docs"
                                    }
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-b0f8-52b5-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:52.874681Z",
                            "start": "2026-04-04T10:26:51.332473Z"
                        },
                        "id": "0242ac17-0011-b0f8-52b5-00000000000c",
                        "name": "Collect logs, artifacts and docs"
                    }
                },
                {
                    "hosts": {
                        "instance": {
                            "action": "shell",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": true,
                                    "cmd": "if [ -n \"$(find /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts -mindepth 1)\" ] ; then\n  # Only create target directory if it is needed.\n  # Do not fail if it is already there.\n  mkdir -p /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/artifacts\n  # Leave the original directory behind so that other roles\n  # operating on the interface directories can simply no-op.\n  mv -f /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts/* /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/artifacts\nfi\n",
                                    "delta": "0:00:00.008543",
                                    "end": "2026-04-04 10:26:53.307410",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "if [ -n \"$(find /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts -mindepth 1)\" ] ; then\n  # Only create target directory if it is needed.\n  # Do not fail if it is already there.\n  mkdir -p /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/artifacts\n  # Leave the original directory behind so that other roles\n  # operating on the interface directories can simply no-op.\n  mv -f /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/artifacts/* /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/artifacts\nfi\n",
                                            "_uses_shell": true,
                                            "argv": null,
                                            "chdir": null,
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_no_log": false,
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-04-04 10:26:53.298867",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "",
                                    "stdout_lines": [],
                                    "zj_item": "artifacts",
                                    "zuul_log_id": "in-loop-ignore"
                                },
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": true,
                                    "cmd": "if [ -n \"$(find /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs -mindepth 1)\" ] ; then\n  # Only create target directory if it is needed.\n  # Do not fail if it is already there.\n  mkdir -p /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/docs\n  # Leave the original directory behind so that other roles\n  # operating on the interface directories can simply no-op.\n  mv -f /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs/* /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/docs\nfi\n",
                                    "delta": "0:00:00.008201",
                                    "end": "2026-04-04 10:26:53.557376",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "if [ -n \"$(find /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs -mindepth 1)\" ] ; then\n  # Only create target directory if it is needed.\n  # Do not fail if it is already there.\n  mkdir -p /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/docs\n  # Leave the original directory behind so that other roles\n  # operating on the interface directories can simply no-op.\n  mv -f /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/docs/* /var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/work/logs/docs\nfi\n",
                                            "_uses_shell": true,
                                            "argv": null,
                                            "chdir": null,
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_no_log": false,
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-04-04 10:26:53.549175",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "",
                                    "stdout_lines": [],
                                    "zj_item": "docs",
                                    "zuul_log_id": "in-loop-ignore"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-b0f8-52b5-00000000000f",
                        "name": "merge-output-to-logs",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/merge-output-to-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:53.609835Z",
                            "start": "2026-04-04T10:26:52.889851Z"
                        },
                        "id": "0242ac17-0011-b0f8-52b5-000000000011",
                        "name": "Move artifacts and docs to logs dir"
                    }
                }
            ]
        },
        {
            "play": {
                "duration": {
                    "end": "2026-04-04T10:26:54.060681Z",
                    "start": "2026-04-04T10:26:53.618767Z"
                },
                "id": "0242ac17-0011-b0f8-52b5-000000000013",
                "name": "all"
            },
            "tasks": [
                {
                    "hosts": {
                        "instance": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC18iToJ05crK3mrjrruwsI+Y3C4kUryzDSqSq7sYq2/oqCcCKb+/jIju4ha8TiGO+rrZ/Fl3kQvUWVuRXiMrjy9mbJTsAj2byBhx/Krgd3LAHS0wSheCF/xmlpClSjBIUX/i+8Gen90424txbIUIe2PxBalZ0FtWG6DK3roNbJCvl5LyxbkOEQwyGG7ixNdLzJshSoFmhPn8ggqRB0A2MkvyKOgkhh+sA7kQ7kthhYCy/LHYxB30GK1+quozfXfx9NGHj9Q7KZwjzMLb17qzPVAp56Gdtv7M8eDWuwk6nvulD4tX0qQk/hFBJtbiQCD5M8eUpP/qYXumLg3oT1jtTSWfylL06EPDXm8XgI7kgWGulM9Zacr8ddH7Aeq9KBCL0NukORM+plLD8oLZqKPnObHHCNHkX1NVm6sLPyW9IJbNrhnJRijngy6yG2PAW85lv0czGP3b0Yb847IdptBU7AwbE1qq3kRsZwg5b3RD+4n5lbSHhCRsTCGBIasf3F5OM= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "absent",
                                    "user": "zuul",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC18iToJ05crK3mrjrruwsI+Y3C4kUryzDSqSq7sYq2/oqCcCKb+/jIju4ha8TiGO+rrZ/Fl3kQvUWVuRXiMrjy9mbJTsAj2byBhx/Krgd3LAHS0wSheCF/xmlpClSjBIUX/i+8Gen90424txbIUIe2PxBalZ0FtWG6DK3roNbJCvl5LyxbkOEQwyGG7ixNdLzJshSoFmhPn8ggqRB0A2MkvyKOgkhh+sA7kQ7kthhYCy/LHYxB30GK1+quozfXfx9NGHj9Q7KZwjzMLb17qzPVAp56Gdtv7M8eDWuwk6nvulD4tX0qQk/hFBJtbiQCD5M8eUpP/qYXumLg3oT1jtTSWfylL06EPDXm8XgI7kgWGulM9Zacr8ddH7Aeq9KBCL0NukORM+plLD8oLZqKPnObHHCNHkX1NVm6sLPyW9IJbNrhnJRijngy6yG2PAW85lv0czGP3b0Yb847IdptBU7AwbE1qq3kRsZwg5b3RD+4n5lbSHhCRsTCGBIasf3F5OM= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "absent",
                            "user": "zuul",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac17-0011-b0f8-52b5-000000000016",
                        "name": "remove-build-sshkey",
                        "path": "/var/lib/zuul/builds/5b9aaae51dd74c078827d3a79afb7078/trusted/project_1/opendev.org/zuul/zuul-jobs/roles/remove-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-04-04T10:26:54.060681Z",
                            "start": "2026-04-04T10:26:53.630093Z"
                        },
                        "id": "0242ac17-0011-b0f8-52b5-000000000018",
                        "name": "Remove the build SSH key from all nodes"
                    }
                }
            ]
        }
    ],
    "stats": {
        "instance": {
            "changed": 4,
            "failures": 0,
            "ignored": 0,
            "ok": 5,
            "rescued": 0,
            "skipped": 1,
            "unreachable": 0
        }
    },
    "trusted": true
}
]
